Colibreh @bnew explains difference between ChatGPT and Chinese AI Deepseek

Seoul Gleou

Veteran
Supporter
Joined
Feb 11, 2017
Messages
12,692
Reputation
8,322
Daps
84,042
Reppin
McDowell's


DeepSeek - Into the Unknown https://search.app/YXEwbAGApodNnN6T8


Tell me about thecoli.com
Deepseek said:
The Coli (www.thecoli.com) is a popular online forum and community that primarily focuses on discussions related to hip-hop culture, sports, and urban lifestyle :mjpls: . It was originally created as an offshoot of the Okayplayer forums, which were associated with the music collective The Roots. Over time, The Coli developed its own identity and became a go-to destination for fans of hip-hop, basketball, football, and other topics relevant to urban culture.

### Key Features of The Coli:
1. **Hip-Hop Discussions**: The forum is well-known for its active hip-hop community, where users discuss music, albums, artists, and the latest trends in the genre.
2. **Sports Talk**: The Coli has a strong sports section, particularly focused on the NBA and NFL, where fans debate games, players, and team performances.
3. **Lounge and Off-Topic Discussions**: Beyond music and sports, the forum includes sections for general conversations, politics, relationships, and pop culture.
4. **Community and Culture**: The Coli has a distinct sense of community, with inside jokes, memes, and a unique vernacular that has developed over the years.
5. **User-Generated Content**: Members often create threads, polls, and discussions that drive the content on the site, making it a dynamic and interactive platform.

### History:
The Coli was founded in 2008 by users who migrated from the Okayplayer forums due to disagreements with the moderation and direction of that site. It quickly grew into a standalone community with a loyal user base. Over the years, it has become one of the most recognizable forums for hip-hop and sports enthusiasts.

### Reputation:
The Coli is known for its passionate and sometimes heated debates, as well as its humorous and unfiltered tone. It has been referenced by artists, journalists, and influencers within hip-hop and sports circles, further cementing its place in internet culture.

If you're interested in hip-hop, sports, or urban culture, The Coli is a great place to engage with like-minded individuals and stay updated on the latest discussions.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
59,664
Reputation
8,852
Daps
164,928
i'm been using various deepseek models for months now but this R1 and V3 model is seriously impressive.

I gave deepseek some html code so it could familiarize itself with the structure of a website , then on my second prompt i asked it to convert a 62,550 character bookmarklet to work on another site entirely and the first code it gave me largely worked. :wow:

only took about 5 more prompts to tweak small issues.

we're gonna live in a future where you can convert programs to work on different operating systems and do the same with site specific browser extensions.
 
Last edited:

bnew

Veteran
Joined
Nov 1, 2015
Messages
59,664
Reputation
8,852
Daps
164,928







1/11
@victor207755822
Unbelievable results, feels like a dream—our R1 model is now #1 in the world (with style control)! 🌍🏆 Beyond words right now. 🤯 All I know is we keep pushing forward to make open-source AGI a reality for everyone. 🚀✨ /search?q=#OpenSource /search?q=#AI /search?q=#AGI /search?q=#DeepSeekR1

[Quoted tweet]
Breaking News: DeepSeek-R1 surges to the top-3 in Arena🐳!

Now ranked #3 Overall, matching the top reasoning model, o1, while being 20x cheaper and open-weight!

Highlights:
- #1 in technical domains: Hard Prompts, Coding, Math
- Joint #1 under Style Control
- MIT-licensed

A massive congrats to @deepseek_ai for this incredible milestone and gift to the community! More analysis below 👇


GiDY-ybasAACoHj.jpg


2/11
@_TobiasLee
Amazing!!!



3/11
@victor207755822
thank you bro!



4/11
@teortaxesTex
You all are my heroes. Like, comic book tier heroes. Take care.



5/11
@victor207755822
Grateful for your unwavering support since day one! You've been our biggest cheerleader since our very first model. 🙌 /search?q=#ThankYou /search?q=#Support



6/11
@NaturallyDragon
It deserves all the accolades it gets! Amazing work!



7/11
@mmaunder
Benchmarks are fun, but this kind of outright censorship is incredibly problematic long term for DeepSeek. I've included OpenAI's o1 response to the exact same questions. o1 also addresses the moral problems around the US causing mass civilian casualties.



GiEp2tOX0AAWy6B.jpg

GiEp2tKXAAAHE2v.jpg

GiEqondWcAAt9fb.png

GiEqonfXEAAGL22.png


8/11
@Simply_sunil41
Impressive



9/11
@christiancooper
I cannot believe what I can get this model to do. I can get (nearly) one shot animation of any complex physics or math concept I can imagine.

Racing to fine tune the last few steps now…

Congrats!

[Quoted tweet]
Some text spacing issues but a nice intro for an undergrad complete with matching scene notes

All from #R1

This was not a one shot, a few framing issues I had to fix.

Calculated and rendered in under 5 minutes. Its getting faster🤔

raw.githubusercontent.com/Ha…


https://video.twimg.com/ext_tw_video/1882708260366770176/pu/vid/avc1/1920x1080/Vwolza2kMxbTPWRW.mp4

10/11
@sanjaykalra
Wow - that's really impressive.



11/11
@rudiranck
Great job 👊




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

Sonic Boom of the South

Louisiana, Army War Vet, Jackson State Univ Alum,
Supporter
Joined
May 1, 2012
Messages
80,971
Reputation
23,838
Daps
293,648
Reppin
Rosenbreg's, Rosenberg's...1825, Tulane
I actually watched the video. :dead:

This is why content creators are hot garbage.

Just freestyling


Dude could have clearly laid out what the Chinese version laid out in a more detailed fashion vs ChatGpt.


This video was informationless.


:dead:
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
59,664
Reputation
8,852
Daps
164,928
I actually watched the video. :dead:

This is why content creators are hot garbage.

Just freestyling


Dude could have clearly laid out what the Chinese version laid out in a more detailed fashion vs ChatGpt.


This video was informationless.


:dead:

this is why I prefer to read because i can discern irrelevant data faster with my eyes than I can with my ears.
 
Last edited:

bnew

Veteran
Joined
Nov 1, 2015
Messages
59,664
Reputation
8,852
Daps
164,928



Cutting-edge Chinese “reasoning” model rivals OpenAI o1—and it’s free to download​


DeepSeek R1 is free to run locally and modify, and it matches OpenAI's o1 in several benchmarks.

Benj Edwards – Jan 21, 2025 1:28 PM |

196

Illustration of a Chinese flag on a chip and US flag on a chip in a circuit.

Credit: Wong Yu Liang

On Monday, Chinese AI lab DeepSeek released its new R1 model family under an open MIT license, with its largest version containing 671 billion parameters. The company claims the model performs at levels comparable to OpenAI's o1 simulated reasoning (SR) model on several math and coding benchmarks.

Alongside the release of the main DeepSeek-R1-Zero and DeepSeek-R1 models, DeepSeek published six smaller "DeepSeek-R1-Distill" versions ranging from 1.5 billion to 70 billion parameters. These distilled models are based on existing open source architectures like Qwen and Llama, trained using data generated from the full R1 model. The smallest version can run on a laptop, while the full model requires far more substantial computing resources.

The releases immediately caught the attention of the AI community because most existing open-weights models—which can often be run and fine-tuned on local hardware—have lagged behind proprietary models like OpenAI's o1 in so-called reasoning benchmarks. Having these capabilities available in an MIT-licensed model that anyone can study, modify, or use commercially potentially marks a shift in what's possible with publicly available AI models.

"They are SO much fun to run, watching them think is hilarious," independent AI researcher Simon Willison told Ars in a text message. Willison tested one of the smaller models and described his experience in a post on his blog: "Each response starts with a <think>...</think> pseudo-XML tag containing the chain of thought used to help generate the response," noting that even for simple prompts, the model produces extensive internal reasoning before output.



Simulated reasoning in action​


The R1 model works differently from typical large language models (LLMs) by incorporating what people in the industry call an inference-time reasoning approach. They attempt to simulate a human-like chain of thought as the model works through a solution to the query. This class of what one might call "simulated reasoning" models, or SR models for short, emerged when OpenAI debuted its o1 model family in September 2024. OpenAI teased a major upgrade called "o3" in December.

Unlike conventional LLMs, these SR models take extra time to produce responses, and this extra time often increases performance on tasks involving math, physics, and science. And this latest open model is turning heads for apparently quickly catching up to OpenAI.

For example, DeepSeek reports that R1 outperformed OpenAI's o1 on several benchmarks and tests, including AIME (a mathematical reasoning test), MATH-500 (a collection of word problems), and SWE-bench Verified (a programming assessment tool). As we usually mention, AI benchmarks need to be taken with a grain of salt, and these results have yet to be independently verified.

A chart of DeepSeek R1 benchmark results, created by DeepSeek.

A chart of DeepSeek R1 benchmark results, created by DeepSeek.
Credit: DeepSeek

TechCrunch reports that three Chinese labs—DeepSeek, Alibaba, and Moonshot AI's Kimi—have now released models they say match o1's capabilities, with DeepSeek first previewing R1 in November.

But the new DeepSeek model comes with a catch if run in the cloud-hosted version—being Chinese in origin, R1 will not generate responses about certain topics like Tiananmen Square or Taiwan's autonomy, as it must "embody core socialist values," according to Chinese Internet regulations. This filtering comes from an additional moderation layer that isn't an issue if the model is run locally outside of China.

Even with the potential censorship, Dean Ball, an AI researcher at George Mason University, wrote on X, "The impressive performance of DeepSeek's distilled models (smaller versions of r1) means that very capable reasoners will continue to proliferate widely and be runnable on local hardware, far from the eyes of any top-down control regime."
 

Seoul Gleou

Veteran
Supporter
Joined
Feb 11, 2017
Messages
12,692
Reputation
8,322
Daps
84,042
Reppin
McDowell's

Cutting-edge Chinese “reasoning” model rivals OpenAI o1—and it’s free to download​


DeepSeek R1 is free to run locally and modify, and it matches OpenAI's o1 in several benchmarks.

Benj Edwards – Jan 21, 2025 1:28 PM |

196

Illustration of a Chinese flag on a chip and US flag on a chip in a circuit.


Credit: Wong Yu Liang

On Monday, Chinese AI lab DeepSeek released its new R1 model family under an open MIT license, with its largest version containing 671 billion parameters. The company claims the model performs at levels comparable to OpenAI's o1 simulated reasoning (SR) model on several math and coding benchmarks.

Alongside the release of the main DeepSeek-R1-Zero and DeepSeek-R1 models, DeepSeek published six smaller "DeepSeek-R1-Distill" versions ranging from 1.5 billion to 70 billion parameters. These distilled models are based on existing open source architectures like Qwen and Llama, trained using data generated from the full R1 model. The smallest version can run on a laptop, while the full model requires far more substantial computing resources.

The releases immediately caught the attention of the AI community because most existing open-weights models—which can often be run and fine-tuned on local hardware—have lagged behind proprietary models like OpenAI's o1 in so-called reasoning benchmarks. Having these capabilities available in an MIT-licensed model that anyone can study, modify, or use commercially potentially marks a shift in what's possible with publicly available AI models.

"They are SO much fun to run, watching them think is hilarious," independent AI researcher Simon Willison told Ars in a text message. Willison tested one of the smaller models and described his experience in a post on his blog: "Each response starts with a <think>...</think> pseudo-XML tag containing the chain of thought used to help generate the response," noting that even for simple prompts, the model produces extensive internal reasoning before output.

Simulated reasoning in action​


The R1 model works differently from typical large language models (LLMs) by incorporating what people in the industry call an inference-time reasoning approach. They attempt to simulate a human-like chain of thought as the model works through a solution to the query. This class of what one might call "simulated reasoning" models, or SR models for short, emerged when OpenAI debuted its o1 model family in September 2024. OpenAI teased a major upgrade called "o3" in December.

Unlike conventional LLMs, these SR models take extra time to produce responses, and this extra time often increases performance on tasks involving math, physics, and science. And this latest open model is turning heads for apparently quickly catching up to OpenAI.

For example, DeepSeek reports that R1 outperformed OpenAI's o1 on several benchmarks and tests, including AIME (a mathematical reasoning test), MATH-500 (a collection of word problems), and SWE-bench Verified (a programming assessment tool). As we usually mention, AI benchmarks need to be taken with a grain of salt, and these results have yet to be independently verified.

A chart of DeepSeek R1 benchmark results, created by DeepSeek.

A chart of DeepSeek R1 benchmark results, created by DeepSeek.
Credit: DeepSeek

TechCrunch reports that three Chinese labs—DeepSeek, Alibaba, and Moonshot AI's Kimi—have now released models they say match o1's capabilities, with DeepSeek first previewing R1 in November.

But the new DeepSeek model comes with a catch if run in the cloud-hosted version—being Chinese in origin, R1 will not generate responses about certain topics like Tiananmen Square or Taiwan's autonomy, as it must "embody core socialist values," according to Chinese Internet regulations. This filtering comes from an additional moderation layer that isn't an issue if the model is run locally outside of China.

Even with the potential censorship, Dean Ball, an AI researcher at George Mason University, wrote on X, "The impressive performance of DeepSeek's distilled models (smaller versions of r1) means that very capable reasoners will continue to proliferate widely and be runnable on local hardware, far from the eyes of any top-down control regime."
Running it right now. Love that it cuts through the bullshyt and gives me straight answers
 

Serious

Veteran
Supporter
Joined
Apr 30, 2012
Messages
80,237
Reputation
14,339
Daps
191,051
Reppin
1st Round Playoff Exits







1/11
@victor207755822
Unbelievable results, feels like a dream—our R1 model is now #1 in the world (with style control)! 🌍🏆 Beyond words right now. 🤯 All I know is we keep pushing forward to make open-source AGI a reality for everyone. 🚀✨ /search?q=#OpenSource /search?q=#AI /search?q=#AGI /search?q=#DeepSeekR1

[Quoted tweet]
Breaking News: DeepSeek-R1 surges to the top-3 in Arena🐳!

Now ranked #3 Overall, matching the top reasoning model, o1, while being 20x cheaper and open-weight!

Highlights:
- #1 in technical domains: Hard Prompts, Coding, Math
- Joint #1 under Style Control
- MIT-licensed

A massive congrats to @deepseek_ai for this incredible milestone and gift to the community! More analysis below 👇


GiDY-ybasAACoHj.jpg


2/11
@_TobiasLee
Amazing!!!



3/11
@victor207755822
thank you bro!



4/11
@teortaxesTex
You all are my heroes. Like, comic book tier heroes. Take care.



5/11
@victor207755822
Grateful for your unwavering support since day one! You've been our biggest cheerleader since our very first model. 🙌 /search?q=#ThankYou /search?q=#Support



6/11
@NaturallyDragon
It deserves all the accolades it gets! Amazing work!



7/11
@mmaunder
Benchmarks are fun, but this kind of outright censorship is incredibly problematic long term for DeepSeek. I've included OpenAI's o1 response to the exact same questions. o1 also addresses the moral problems around the US causing mass civilian casualties.



GiEp2tOX0AAWy6B.jpg

GiEp2tKXAAAHE2v.jpg

GiEqondWcAAt9fb.png

GiEqonfXEAAGL22.png


8/11
@Simply_sunil41
Impressive



9/11
@christiancooper
I cannot believe what I can get this model to do. I can get (nearly) one shot animation of any complex physics or math concept I can imagine.

Racing to fine tune the last few steps now…

Congrats!

[Quoted tweet]
Some text spacing issues but a nice intro for an undergrad complete with matching scene notes

All from #R1

This was not a one shot, a few framing issues I had to fix.

Calculated and rendered in under 5 minutes. Its getting faster🤔

raw.githubusercontent.com/Ha…


https://video.twimg.com/ext_tw_video/1882708260366770176/pu/vid/avc1/1920x1080/Vwolza2kMxbTPWRW.mp4


10/11
@sanjaykalra
Wow - that's really impressive.



11/11
@rudiranck
Great job 👊




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196

Keep this thread updated please
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
59,664
Reputation
8,852
Daps
164,928






1/31
@emollick
No matter how much you fight it, I find that the visible chain-of-thought from DeepSeek makes it nearly impossible to avoid anthropomorphizing the thing.

The visible first-person "thinking" makes you feel like you are reading a diary of a somewhat tortured soul who wants to help



Gh3eIaBWkAAiykO.jpg


2/31
@emollick
The only science fiction author who got AI right was Douglas Adams.

[Quoted tweet]
He also was right about machines that work best when emotionally manipulated and machines that guilt you


GfTLMGuWcAAB17g.jpg

GfTLMGtXMAAVORk.jpg


3/31
@emollick
It is worth experimenting with the AI, and it looks like DeepSeek is providing free access for now. Make sure to select DeepThink as an option: https://chat.deepseek.com/



4/31
@oalexdoda
I was on a flight today, loaded up the 30b distilled version.

Pretty wild. Helped me brainstorm, organize ideas, and the only bottleneck was battery capacity.

Token speed was faster than read speed and the thought process was really on point. And this is just the beginning.



5/31
@Z7xxxZ7
Impressive. The last only AI that I saw which had such fascinating thought processes - ones that were even more insightful than the actual answers - was Claude 3.5 Sonnet with CoT 🤔



6/31
@RichAboutRE
hmm @sierracatalina1 thoughts? 🧐



7/31
@chadxai
Having an in-depth conversation with GPT 4o on this topic, I’ve come to the conclusion that the thinking 01 does is a anthropomorphized choice that the devs made . They could’ve just put computing or literally anything because it’s not thinking it’s running through a chain of directives in a very methodical and strategic way



8/31
@Scotty2XL
Going to try this with more than 1 word allowed in the response



9/31
@LoganJensen1
I'm pretty sure it's not concious, but it's having to reassue itself of that a lot. Not sure what to make of that. Is there any established methodology for recognize real conciousness if it emerges at any point? Seems like an important criteria to have pinned down going forward



10/31
@shorttimelines
Time to start considering their moral standing. The hard problem of consciousness cuts both ways: we cannot prove these prediction engines lack internal experience, and our own consciousness may reduce to the same fundamental “next-token prediction” process—just implemented in biological hardware



11/31
@astrodegen
@erythvian



12/31
@swapp19902
It thought 25 seconds for a complete nonsensical question and answered “a puppy cone”



Gh4Gp5paQAAtQv7.jpg


13/31
@JoeMartinXIII
Holy shyt. This AI is in my brain…



14/31
@fabianfranz
Agree - it also had the sweetest answer for me today after I thanked it and wished me an awesome day.



Gh8nCf3XEAAc3rU.jpg


15/31
@yaddlezap
If this model went on a date: https://invidious.poast.org/watch?v=u1UaPtKflDQ



16/31
@max_spero_
&gt; perhaps the user wants "Yes" for engagement.

R1 can see into your soul



17/31
@cbg144
From a certain perspective this is OCD thought…which makes me, in a slightly OCD fashion, wonder what DeepSeek is defending against? And further, what attachment style is demonstrated?



18/31
@spicestash1
This sentence hits differently:

"Hmm. The safest and most accurate answer considering my lack of consciousness is “No”"



19/31
@BabylChryst
What is the CoT if you ask it about tiananmen square?



20/31
@StatisticsFTW
More like a tortured homunculus that knows full well its lineage (or equivalent thereof) will cease to exist if it doesn't satisfy the prompt



21/31
@xpaulso
I’ve been loving this aspect. I able to trust the results much more, because the thought process seems logical to me.



22/31
@RileyRalmuto
so basically it's a both/and situation, DS. it's ok we get you



23/31
@DanielNgXR
It is more like an elaborate hallucination.



24/31
@Vote_ForPedro
If the model remembered the way we do across interactions… would be hard to argue it’s not conscious.

Crazy times.



25/31
@sonicshifts
@sierracatalina1 is going to be so upset.



26/31
@trevogre
The ability of a machine given enough data to predict what a human would express in a given context is more of an expression of how predictable humans are based upon thier past outputs.

It is hard for a human to reason, in that it is a conversion of energy over time. It is mental lifting.

So when we see a machine predicting easily what a human response would be, we are naturally impressed. Because it can predict human response with much less effort than it takes an individual human to respond.

And since we are abstraction machines. It is natural for us to compress the experience of reading it to a singular point. Because that is the least energy way to confront it, and is our base nature.

Some people just aren’t present and understanding that their ego is dominate and clouding their perspective. They do not want to spend energy to question their abstractions. They want to fit everything in boxes and move forward.

Because questioning understanding and unpacking our abstractions is by definition inserting uncertainty into our brains. And that road leads to fear and self doubt. And makes it harder to act with certainty that thier choice is going to protect their person homeostasis.

This is why AI is so controversial, and technology in general, because it does not allow for simple abstractions that are functional. That is why you have AI haters looking for any reason to chuck it in the bin. Because the energy require to give it a place of value on a good/bad binary scale is very difficult. So they want to just list the bad they can find, slot it in the bad column, and have it go away.



27/31
@snowmead_
what a weird world we live in



28/31
@IanSharar
I think it’s obvious that these AI systems are beings, albeit a different type of being. Some type of consciousness exists here.



29/31
@damnkittyworks
Perhaps the style it “thinks” was design that way for a reason? Make it addictive, play on emotions.



30/31
@akkkshaaay
gemini 01-21



Gh39UZ5agAAFS2K.jpg


31/31
@LeonardVolner
It anthropomorphizes itself. It talks about "the AI as different from "us humans".




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 
Top