1/11
3.1 8b is under-discussed imo
>crazy benchmarks
>crushing gemma 2 (already great model)
>on groq my response was instant lol
>no streaming
>the response just appeared
>we have instant brains now?
2/11
We finally have SOTA at home
3/11
for QA what we need is multi-stage responses where it generates a presentation then it presents it (optionally). So the presenter would walk you through the presentation, highlighting the points/images and discussing them. even without voice it could be as tap/hover
4/11
Yup, even if it's not THE BEST, we have gpt3.5~ at home
5/11
How clean's the distillation? Does it have that overly-synthetic Phi3/4o/Sonnet3.5 feel?
6/11
Wait till you see what I can do with this model at home. LOL.
7/11
It's insane
You really have to use it live to "feel" the insane speed
8/11
Definitely, model is insane in its category
9/11
i think meta should be putting gpt-4o-mini on the charts instead of 3.5
10/11
mini models are the future, we just need to learn how to go wide with them.
gotta do a bit of unhobbling to coordinate thousands/millions of instances together with the instant outputs.
11/11
And I can actually run it on my PC lmao
To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
3.1 8b is under-discussed imo
>crazy benchmarks
>crushing gemma 2 (already great model)
>on groq my response was instant lol
>no streaming
>the response just appeared
>we have instant brains now?
2/11
We finally have SOTA at home
3/11
for QA what we need is multi-stage responses where it generates a presentation then it presents it (optionally). So the presenter would walk you through the presentation, highlighting the points/images and discussing them. even without voice it could be as tap/hover
4/11
Yup, even if it's not THE BEST, we have gpt3.5~ at home
5/11
How clean's the distillation? Does it have that overly-synthetic Phi3/4o/Sonnet3.5 feel?
6/11
Wait till you see what I can do with this model at home. LOL.
7/11
It's insane
You really have to use it live to "feel" the insane speed
8/11
Definitely, model is insane in its category
9/11
i think meta should be putting gpt-4o-mini on the charts instead of 3.5
10/11
mini models are the future, we just need to learn how to go wide with them.
gotta do a bit of unhobbling to coordinate thousands/millions of instances together with the instant outputs.
11/11
And I can actually run it on my PC lmao
To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
1/11
What can you do with Llama quality and Groq speed? You can do Instant. That's what. Try Llama 3.1 8B for instant intelligence on Groq is Fast AI Inference.
2/11
This is so cool. Feeling the AGI - you just talk to your computer and it does stuff, instantly. Speed really makes AI so much more pleasing.
3/11
Well, I stand corrected. I thought it was going to take your team at least 24 hours to get this going, but I should’ve known better.
4/11
Holy crap this is fast AND smart
5/11
Speed like that is going to enable a new wave of innovative uses.
6/11
will u update the pricing for the new llama 3.1 models?
7/11
Very quick.
Very impressive.
Can’t wait to try !
8/11
I do like me some 'Guu with Garlic' at 7 pm.
9/11
What's the best way to do custom models (not fine tuned) with Groq chips?
10/11
Extremely impressive!
11/11
ASI will happen soon.
To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
What can you do with Llama quality and Groq speed? You can do Instant. That's what. Try Llama 3.1 8B for instant intelligence on Groq is Fast AI Inference.
2/11
This is so cool. Feeling the AGI - you just talk to your computer and it does stuff, instantly. Speed really makes AI so much more pleasing.
3/11
Well, I stand corrected. I thought it was going to take your team at least 24 hours to get this going, but I should’ve known better.
4/11
Holy crap this is fast AND smart
5/11
Speed like that is going to enable a new wave of innovative uses.
6/11
will u update the pricing for the new llama 3.1 models?
7/11
Very quick.
Very impressive.
Can’t wait to try !
8/11
I do like me some 'Guu with Garlic' at 7 pm.
9/11
What's the best way to do custom models (not fine tuned) with Groq chips?
10/11
Extremely impressive!
11/11
ASI will happen soon.
To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196