bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252

OpenAI CEO says company could become for-profit corporation, The Information reports​

By Reuters

June 15, 20241:37 AM EDTUpdated an hour ago

54th WEF annual meeting in Davos

Sam Altman, CEO of OpenAI, attends the 54th annual meeting of the World Economic Forum, in Davos, Switzerland, January 18, 2024. REUTERS/Denis Balibouse/File Photo Purchase Licensing Rights, opens new tab
June 14 (Reuters) - OpenAI CEO Sam Altman told some shareholders that the company is considering changing its governance structure to a for-profit business that the firm's nonprofit board doesn't control, The Information reported on Friday.

One scenario Altman said the board is considering is a for-profit benefit corporation, which rivals such as Anthropic and xAI are using, the report said, citing a person who heard the comments.

The restructuring discussions are fluid and Altman and his fellow directors could ultimately decide to take a different approach, The Information added.

In response to Reuters' queries about the report, OpenAI said: "We remain focused on building AI that benefits everyone. The nonprofit is core to our mission and will continue to exist."
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252

TikTok ads may soon contain AI-generated avatars of your favorite creators​



TikTok is expanding its Symphony ad suite with AI dubbing tools and avatars based on paid actors and creators.​

By Jess Weatherbed, a news writer focused on creative industries, computing, and internet culture. Jess started her career at TechRadar, covering news and hardware reviews.

Jun 17, 2024, 9:00 AM EDT

Vector art of the TikTok logo.

TikTok says the new tools are aimed at helping brands and creators develop global audiences. The Verge

TikTok is introducing some new generative AI tools that aim to help organizations and content creators grow their global audiences using customizable digital avatars and language dubbing features. Building on the Symphony generative AI ad suite unveiled last month, TikTok says these new tools are intended to break down language barriers in marketing and allow brands to “add a human touch to their content” where real models or presenters wouldn’t otherwise be used.

The first of the new offerings is Symphony Digital Avatars, which are available in two varieties: stock or custom. Stock avatars are based on paid actors from a diverse range of backgrounds, nationalities, and languages. They are available for commercial use. Custom avatars, meanwhile, are created to resemble a specific creator or a brand spokesperson and speak multiple languages — allowing the accounts that utilize them to reach foreign audiences while retaining a specific likeness. Regardless of which type of avatar is used, videos that use them will be marked with an “AI-generated” label.

A selection of the Symphony Digital Avatars being introduced by TikTok.

Here are some examples of the stock avatars available via TikTok’s new Symphony tools. Image: TikTok

That multi-language support comes courtesy of Symphony AI Dubbing — a “global translation tool” that enables creators and marketers to dub their content into over 10 languages and dialects, including French, Spanish, Portuguese, German, and Korean. TikTok says the tool automatically detects what language is being spoken in videos and is capable of transcribing, translating, and producing a dubbed video in whatever language was selected by the user.

There’s certainly precedent for creators attempting to retain their own identities while breaking into other language markets. For example, MrBeast notably uses YouTube’s multi-language audio track support to dub his videos into other languages. FKA Twigs also revealed last month that she’s created a multilingual “deepfake” version of herself to help promote her work globally. We have asked TikTok about the pricing structure for its new AI marketing tools but have not yet heard back.

The language translation options for TikTok’s Symphony Digital Avatars

The inclusion of dialects here suggests users may be able to dictate whether an avatar uses British or American English or specify similar regional accents. Image: TikTok

The demonstration video TikTok provided of a custom digital avatar — based on TikTok’s global head of content strategy and operations, Adrienne Lahens — is a little uncanny, but it looks just natural enough to be convincing if you’re not fixated on its overly expressive movements.

@tiktoknewsroom

Introducing Symphony Digital Avatars, to help creators and brands captivate global audiences and deliver impactful messages in an immersive and authentic way. Check out our Newsroom to learn more.

♬ original sound - TikTok Newsroom



Still, creators will need to have faith that TikTok’s new dubbing tools will be accurate enough to avoid embarrassing mistranslation blunders. And TikTok users who are already sick of the platform’s pervasive ads may not find being pitched to by digital avatars any less frustrating than the real deal.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252




"We present DeepSeek-Coder-V2, an open-source Mixture-of-Experts (MoE) code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks. Specifically, DeepSeek-Coder-V2 is further pre-trained from DeepSeek-Coder-V2-Base with 6 trillion tokens sourced from a high-quality and multi-source corpus. Through this continued pre-training, DeepSeek-Coder-V2 substantially enhances the coding and mathematical reasoning capabilities of DeepSeek-Coder-V2-Base, while maintaining comparable performance in general language tasks. Compared to DeepSeek-Coder, DeepSeek-Coder-V2 demonstrates significant advancements in various aspects of code-related tasks, as well as reasoning and general capabilities. Additionally, DeepSeek-Coder-V2 expands its support for programming languages from 86 to 338, while extending the context length from 16K to 128K."

 
Last edited:

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252

Anthropic claims its latest model is best-in-class​


Kyle Wiggers

7:00 AM PDT • June 20, 2024

Comment

Anthropic Claude logo
Image Credits: Anthropic

OpenAI rival Anthropic is releasing a powerful new generative AI model called Claude 3.5 Sonnet. But it’s more an incremental step than a monumental leap forward.

Claude 3.5 Sonnet can analyze both text and images as well as generate text, and it’s Anthropic’s best-performing model yet — at least on paper. Across several AI benchmarks for reading, coding, math and vision, Claude 3.5 Sonnet outperforms the model it’s replacing, Claude 3 Sonnet, and beats Anthropic’s previous flagship model Claude 3 Opus.

Benchmarks aren’t necessarily the most useful measure of AI progress, in part because many of them test for esoteric edge cases that aren’t applicable to the average person, like answering health exam questions. But for what it’s worth, Claude 3.5 Sonnet just barely bests rival leading models, including OpenAI’s recently launched GPT-4o, on some of the benchmarks Anthropic tested it against.

Alongside the new model, Anthropic is releasing what it’s calling Artifacts, a workspace where users can edit and add to content — e.g. code and documents — generated by Anthropic’s models. Currently in preview, Artifacts will gain new features, like ways to collaborate with larger teams and store knowledge bases, in the near future, Anthropic says.

Focus on efficiency​

Claude 3.5 Sonnet is a bit more performant than Claude 3 Opus, and Anthropic says that the model better understands nuanced and complex instructions, in addition to concepts like humor. (AI is notoriously unfunny, though.) But perhaps more importantly for devs building apps with Claude that require prompt responses (e.g. customer service chatbots), Claude 3.5 Sonnet is faster. It’s around twice the speed of Claude 3 Opus, Anthropic claims.

Vision — analyzing photos — is one area where Claude 3.5 Sonnet greatly improves over 3 Opus, according to Anthropic. Claude 3.5 Sonnet can interpret charts and graphs more accurately and transcribe text from “imperfect” images, such as pics with distortions and visual artifacts.

Michael Gerstenhaber, product lead at Anthropic, says that the improvements are the result of architectural tweaks and new training data, including AI-generated data. Which data specifically? Gerstenhaber wouldn’t disclose, but he implied that Claude 3.5 Sonnet draws much of its strength from these training sets.



Anthropic Claude 3.5 Sonnet
Image Credits:Anthropic

“What matters to [businesses] is whether or not AI is helping them meet their business needs, not whether or not AI is competitive on a benchmark,” Gerstenhaber told TechCrunch. “And from that perspective, I believe Claude 3.5 Sonnet is going to be a step function ahead of anything else that we have available — and also ahead of anything else in the industry.”

The secrecy around training data could be for competitive reasons. But it could also be to shield Anthropic from legal challenges — in particular challenges pertaining to fair use. The courts have yet to decide whether vendors like Anthropic and its competitors, like OpenAI, Google, Amazon and so on, have a right to train on public data, including copyrighted data, without compensating or crediting the creators of that data.

So, all we know is that Claude 3.5 Sonnet was trained on lots of text and images, like Anthropic’s previous models, plus feedback from human testers to try to “align” the model with users’ intentions, hopefully preventing it from spouting toxic or otherwise problematic text.

Anthropic Claude 3.5 Sonnet
Image Credits:Anthropic

What else do we know? Well, Claude 3.5 Sonnet’s context window — the amount of text that the model can analyze before generating new text — is 200,000 tokens, the same as Claude 3 Sonnet. Tokens are subdivided bits of raw data, like the syllables “fan,” “tas” and “tic” in the word “fantastic”; 200,000 tokens is equivalent to about 150,000 words.

And we know that Claude 3.5 Sonnet is available today. Free users of Anthropic’s web client and the Claude iOS app can access it at no charge; subscribers to Anthropic’s paid plans Claude Pro and Claude Team get 5x higher rate limits. Claude 3.5 Sonnet is also live on Anthropic’s API and managed platforms like Amazon Bedrock and Google Cloud’s Vertex AI.

“Claude 3.5 Sonnet is really a step change in intelligence without sacrificing speed, and it sets us up for future releases along the entire Claude model family,” Gerstenhaber said.

Claude 3.5 Sonnet also drives Artifacts, which pops up a dedicated window in the Claude web client when a user asks the model to generate content like code snippets, text documents or website designs. Gerstenhaber explains: “Artifacts are the model output that puts generated content to the side and allows you, as a user, to iterate on that content. Let’s say you want to generate code — the artifact will be put in the UI, and then you can talk with Claude and iterate on the document to improve it so you can run the code.”

The bigger picture​

So what’s the significance of Claude 3.5 Sonnet in the broader context of Anthropic — and the AI ecosystem, for that matter?

Claude 3.5 Sonnet shows that incremental progress is the extent of what we can expect right now on the model front, barring a major research breakthrough. The past few months have seen flagship releases from Google ( Gemini 1.5 Pro) and OpenAI (GPT-4o) that move the needle marginally in terms of benchmark and qualitative performance. But there hasn’t been a leap of matching the leap from GPT-3 to GPT-4 in quite some time, owing to the rigidity of today’s model architectures and the immense compute they require to train.

As generative AI vendors turn their attention to data curation and licensing in lieu of promising new scalable architectures, there are signs investors are becoming wary of the longer-than-anticipated path to ROI for generative AI. Anthropic is somewhat inoculated from this pressure, being in the enviable position of Amazon’s (and to a lesser extent Google’s) insurance against OpenAI. But the company’s revenue, forecasted to reach just under $1 billion by year-end 2024, is a fraction of OpenAI’s — and I’m sure Anthropic’s backers don’t let it forget that fact.

Despite a growing customer base that includes household brands such as Bridgewater, Brave, Slack and DuckDuckGo, Anthropic still lacks a certain enterprise cachet. Tellingly, it was OpenAI — not Anthropic — with which PwC recently partnered to resell generative AI offerings to the enterprise.

So Anthropic is taking a strategic, and well-trodden, approach to making inroads, investing development time into products like Claude 3.5 Sonnet to deliver slightly better performance at commodity prices. Claude 3.5 Sonnet is priced the same as Claude 3 Sonnet: $3 per million tokens fed into the model and $15 per million tokens generated by the model.

Gerstenhaber spoke to this in our conversation. “When you’re building an application, the end user shouldn’t have to know which model is being used or how an engineer optimized for their experience,” he said, “but the engineer could have the tools available to optimize for that experience along the vectors that need to be optimized, and cost is certainly one of them.”

Claude 3.5 Sonnet doesn’t solve the hallucinations problem. It almost certainly makes mistakes. But it might just be attractive enough to get developers and enterprises to switch to Anthropic’s platform. And at the end of the day, that’s what matters to Anthropic.

Toward that same end, Anthropic has doubled down on tooling like its experimental steering AI, which lets developers “steer” its models’ internal features; integrations to let its models take actions within apps; and tools built on top of its models such as the aforementioned Artifacts experience. It’s also hired an Instagram co-founder as head of product. And it’s expanded the availability of its products, most recently bringing Claude to Europe and establishing offices in London and Dublin.

Anthropic, all told, seems to have come around to the idea that building an ecosystem around models — not simply models in isolation — is the key to retaining customers as the capabilities gap between models narrows.

Still, Gerstenhaber insisted that bigger and better models — like Claude 3.5 Opus — are on the near horizon, with features such as web search and the ability to remember preferences in tow.

“I haven’t seen deep learning hit a wall yet, and I’ll leave it to researchers to speculate about the wall, but I think it’s a little bit early to be coming to conclusions on that, especially if you look at the pace of innovation,” he said. “There’s very rapid development and very rapid innovation, and I have no reason to believe that it’s going to slow down.”
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252





















































 

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252

1/1
The new Artifacts UX combined with Claude 3.5 Sonnet on Claude makes this one of the best AI chat interfaces out there! Very impressive speed, quality, and usability.

In less than a minute I had this web page sample previewable directly in the chat.


To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
GQhlkuoXgAAzOZz.jpg



1/1
Anthropic is training on a bunch of file types. It works great, and the models pick up skills like obeying XML tags.

No open pre-training datasets do this.

There are tons of PDFs, SVGs, and LaTeX files on CommonCrawl, it’s not a hard problem to solve.


To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196


1/1
New claude artifacts are awesome!
I asked for a pong game and it coded one for me

No code interpreter, unlike chatGPT, instead you can make documents, notes & stuff, one option being html/jss


To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196


1/1
So jazzed about Claude 3.5 Sonnet and products like Artifacts. It's still early days in the productization of LLMs and
@AnthropicAI has more great ships on the way to bring the magic of LLMs to our daily workflows in practical, useful ways.


To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196





Anthropic has a fast new AI model — and a clever new way to interact with chatbots​

Claude 3.5 Sonnet is apparently Anthropic’s smartest, fastest, and most personable model yet.​

By David Pierce, editor-at-large and Vergecast co-host with over a decade of experience covering consumer tech. Previously, at Protocol, The Wall Street Journal, and Wired.

Jun 20, 2024, 10:00 AM EDT

A screenshot of the Claude app showing 3.5 Sonnet selected.

GPT-4o, Gemini 1.5, and now Claude 3.5 Sonnet. Image: Anthropic

The AI arms race continues apace: Anthropic is launching its newest model, called Claude 3.5 Sonnet, which it says can equal or better OpenAI’s GPT-4o or Google’s Gemini across a wide variety of tasks. The new model is already available to Claude users on the web and on iOS, and Anthropic is making it available to developers as well.

Claude 3.5 Sonnet will ultimately be the middle model in the lineup — Anthropic uses the name Haiku for its smallest model, Sonnet for the mainstream middle option, and Opus for its highest-end model. (The names are weird, but every AI company seems to be naming things in their own special weird ways, so we’ll let it slide.) But the company says 3.5 Sonnet outperforms 3 Opus, and its benchmarks show it does so by a pretty wide margin. The new model is also apparently twice as fast as the previous one, which might be an even bigger deal.

AI model benchmarks should always be taken with a grain of salt; there are a lot of them, it’s easy to pick and choose the ones that make you look good, and the models and products are changing so fast that nobody seems to have a lead for very long. That said, Claude 3.5 Sonnet does look impressive: it outscored GPT-4o, Gemini 1.5 Pro, and Meta’s Llama 3 400B in seven of nine overall benchmarks and four out of five vision benchmarks. Again, don’t read too much into that, but it does seem that Anthropic has built a legitimate competitor in this space.

A screenshot showing various benchmark scores for Claude 3.5 Sonnet and other AI models.

Claude 3.5’s benchmark scores do look impressive — but these things change so fast. Image: Anthropic

What does all that actually amount to? Anthropic says Claude 3.5 Sonnet will be far better at writing and translating code, handling multistep workflows, interpreting charts and graphs, and transcribing text from images. This new and improved Claude is also apparently better at understanding humor and can write in a much more human way.

Along with the new model, Anthropic is also introducing a new feature called Artifacts. With Artifacts, you’ll be able to see and interact with the results of your Claude requests: if you ask the model to design something for you, it can now show you what it looks like and let you edit it right in the app. If Claude writes you an email, you can edit the email in the Claude app instead of having to copy it to a text editor. It’s a small feature, but a clever one — these AI tools need to become more than simple chatbots, and features like Artifacts just give the app more to do.

A screenshot showing a preview of a document alongside an AI chat.

The new Artifacts feature is a hint at what a post-chatbot Claude might look like. Image: Anthropic

Artifacts actually seems to be a signal of the long-term vision for Claude. Anthropic has long said it is mostly focused on businesses (even as it hires consumer tech folks like Instagram co-founder Mike Krieger) and said in its press release announcing Claude 3.5 Sonnet that it plans to turn Claude into a tool for companies to “securely centralize their knowledge, documents, and ongoing work in one shared space.” That sounds more like Notion or Slack than ChatGPT, with Anthropic’s models at the center of the whole system.

For now, though, the model is the big news. And the pace of improvement here is wild to watch: Anthropic launched Claude 3 Opus in March, proudly saying it was as good as GPT-4 and Gemini 1.0, before OpenAI and Google released better versions of their models. Now, Anthropic has made its next move, and it surely won’t be long before its competition does so, too. Claude doesn’t get talked about as much as Gemini or ChatGPT, but it’s very much in the race.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252

Snapchat AI turns prompts into new lens​


Snapchat’s upcoming on-device AI model could transform your background — and your clothing — in real time.​

By Emma Roth, a news writer who covers the streaming wars, consumer tech, crypto, social media, and much more. Previously, she was a writer and editor at MUO.

Jun 19, 2024, 5:26 PM EDT

An image showing Snapchat’s on-device AI model generating a filter based on the prompt “50s sci fi film”

Image: Snapchat

Snapchat offered an early look at its upcoming on-device AI model capable of transforming a user’s surroundings with augmented reality (AR). The new model will eventually let creators turn a text prompt into a custom lens — potentially opening the door for some wild looks to try on and send to friends.

You can see how this might look in the GIF below, which shows a person’s clothing and background transforming in real-time based on the prompt “50s sci-fi film.” Users will start seeing lenses using this new model in the coming months, while creators can start making lenses with the model by the end of this year, according to TechCrunch.

snapchat_lens_gif.gif

GIF: Snapchat

Additionally, Snapchat is rolling out a suite of new AI tools that could make it easier for creators to make custom augmented reality (AR) effects. Some of the tools now available with the latest Lens Studio update include new face effects that let creators write a prompt or upload an image to create a custom lens that completely transforms a user’s face.

The suite also includes a feature, called Immersive ML, that applies a “realistic transformation over the user’s face, body, and surroundings in real time.” Other AI tools coming to Lens Studio allow lens creators to generate 3D assets based on a text or image prompt, create face masks and textures, as well as make 3D character heads that mimic a user’s expression.

This is Snapchat’s “Immersive ML” effect using the prompt “Matisse Style Painting.”

This is Snapchat’s “Immersive ML” effect using the prompt “Matisse Style Painting.” Image: Snapchat

Over the past year, Snapchat has rolled out several new AI features, including a way for subscribers to send AI-generated snaps to friends. Snapchat also launched a ChatGPT-powered AI chatbot to all users last year.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
55,710
Reputation
8,224
Daps
157,252

OpenAI’s former chief scientist is starting a new AI company​

/


Ilya Sutskever is launching Safe Superintelligence Inc., an AI startup that will prioritize safety over ‘commercial pressures.’​

By Emma Roth, a news writer who covers the streaming wars, consumer tech, crypto, social media, and much more. Previously, she was a writer and editor at MUO.

Jun 19, 2024, 2:03 PM EDT

An illustration of a cartoon brain with a computer chip imposed on top.

Illustration by Alex Castro / The Verge

Ilya Sutskever, OpenAI’s co-founder and former chief scientist, is starting a new AI company focused on safety. In a post on Wednesday, Sutskever revealed Safe Superintelligence Inc. (SSI), a startup with “one goal and one product:” creating a safe and powerful AI system.

The announcement describes SSI as a startup that “approaches safety and capabilities in tandem,” letting the company quickly advance its AI system while still prioritizing safety. It also calls out the external pressure AI teams at companies like OpenAI, Google, and Microsoft often face, saying the company’s “singular focus” allows it to avoid “distraction by management overhead or product cycles.”

I am starting a new company: x.com

— Ilya Sutskever (@ilyasut) June 19, 2024

“Our business model means safety, security, and progress are all insulated from short-term commercial pressures,” the announcement reads. “This way, we can scale in peace.” In addition to Sutskever, SSI is co-founded by Daniel Gross, a former AI lead at Apple, and Daniel Levy, who previously worked as a member of technical staff at OpenAI.

Last year, Sutskever led the push to oust OpenAI CEO Sam Altman. Sutskever left OpenAI in May and hinted at the start of a new project. Shortly after Sutskever’s departure, AI researcher Jan Leike announced his resignation from OpenAI, citing safety processes that have “taken a backseat to shiny products.” Gretchen Krueger, a policy researcher at OpenAI, also mentioned safety concerns when announcing her departure.

As OpenAI pushes forward with partnerships with Apple and Microsoft, we likely won’t see SSI doing that anytime soon. During an interview with Bloomberg, Sutskever says SSI’s first product will be safe superintelligence, and the company “will not do anything else” until then.
 
Top