bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787


7HIB3UL.png




 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787

Former Meta AI VP debuts Sizzle, an AI-powered learning app and chatbot​

Lauren Forristal@laurenforristal / 4:22 PM EDT•September 20, 2023
Comment
Sizzle AI

Image Credits: Sizzle (Image has been modified)

Founded by the former vice president of AI at Meta, Jerome Pesenti, Sizzle is a free AI-powered learning app that generates step-by-step answers to math equations and word problems. The company recently launched four new features, including a grading capability, a feature that regenerates steps, an option to see multiple answers to one problem and the ability to upload photos of assignments.

Sizzle works similarly to math solver platforms like Photomath and Symbolab, but it can also solve word problems in subjects like physics, chemistry and biology. Sizzle provides help with all learning levels, from middle school and high school to AP and college.

It’s typical for students to use AI-powered learning apps to instantly get answers without learning anything. OpenAI’s ChatGPT has been a common source to help students cheat. However, Sizzle doesn’t simply provide solutions to the problems. The app acts as a tutor chatbot, guiding the student through each step. Students can also ask the AI questions so they can better understand concepts.

“After leaving Meta, I was inspired to leverage AI to truly help students and non-students no matter what kind of background they come from, the school they attend, or how many resources they have,” Pesenti, who focused on making Meta products safer through the use of AI, told TechCrunch. “I felt that applications of AI haven’t had a clear positive impact on people’s lives. Using it to transform learning is an opportunity to change that.”

The Sizzle app leverages large language models from third parties like OpenAI and developes its own models in-house, Pesenti explained. The AI’s accuracy rate is 90%.



Image Credits: Sizzle


With the new “Grade Your Homework” feature, users can now upload a picture of a completed homework assignment, and the app will provide specific feedback about each solution. If a user makes an error, Sizzle tells them to try again and walks them through it.


Its new “Try a Different Approach” lets the user suggest a different way to solve the problem in a way that makes sense for them. Users can type a brief explanation of how they would like the AI to re-approach, and it will regenerate a step-by-step solution.


There’s also a “Give Me Choices” option, which gives users multiple answers to choose from. We see this feature being useful in preparing students for upcoming tests.

Additionally, the “Answer with a Photo” ability allows them to upload images from their camera roll. Sizzle users could already use their phones to scan a problem.

Built by a team with backgrounds from Meta, Google, X (formerly Twitter) and Twitch, Sizzle already has over 20,000 downloads since launching in August. The average rating on both the App Store and Google Play store is currently 4.6 stars.

Sizzle hopes that rolling out these new features will encourage more students to try the app.


Unlike most learning apps that require users to pay to unlock certain features, Sizzle is completely free to use. The company eventually wants to add a premium offering and in-app purchases, but the version of the app for solving step-by-step problems will remain free.

Sizzle recently secured $7.5 million in seed funding, led by Owl Ventures, with participation from 8VC and FrenchFounders. Sizzle is using the funding to expand its team and help develop the product. The company plans to add more features in the next few months.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787

Amazon to invest up to $4 billion in AI startup Anthropic​

Manish Singh @refsrc / 3:10 AM EDT•September 25, 2023

Amazon to invest up to $4 billion in AI startup Anthropic

Image Credits: Anthropic

Amazon has agreed to invest up to $4 billion in the AI startup Anthropic, the two firms said, as the e-commerce group steps up its rivalry against Microsoft, Meta, Google and Nvidia in the fast-growing sector that many technologists believe could be the next great frontier.

The e-commerce group said it will initially invest $1.25 billion for a minority stake in Anthropic, which like Google’s Bard and Microsoft-backed OpenAI also operates an AI-powered, text analyzing chatbot. As part of the deal, Amazon said it has an option to increase its investment in Anthropic to a total of $4 billion.

TechCrunch reported exclusively earlier this year that Anthropic, which also counts Google as an investor, plans to raise as much as $5 billion over the next two years. Anthropic, which earlier this month launched its first consumer-facing premium subscription plan of chatbot Claude 2, plans to build a “frontier model” — tentatively called “Claude-Next” — that is 10 times more capable than today’s most powerful AI, according to a 2023 investor deck TechCrunch obtained earlier this year.


But this development, the startup cautioned, will require a billion dollars in spending over the next 18 months. (Microsoft has invested as much as $11 billion in OpenAI over the years.)

In Amazon, Anthropic has found a deep-pocketed strategic investor that can also provide it with compute power to build future AI models and then find and help sell the offerings to scores of cloud customers.

As part of the investment agreement, Anthropic will use Amazon’s cloud giant AWS as a primary cloud provider for mission-critical workloads, including safety research and future foundation model development, the e-commerce group said. Anthropic will additionally use AWS Trainium and Inferentia chips to build, train and deploy its future foundation models. (Anthropic has been a customer of AWS since 2021.)

Amazon believes it can help “improve many customer experiences, short and long-term, through our deeper collaboration” with Anthropic, said Andy Jassy, Amazon chief executive, in a statement.

“Customers are quite excited about Amazon Bedrock, AWS’s new managed service that enables companies to use various foundation models to build generative AI applications on top of, as well as AWS Trainium, AWS’s AI training chip, and our collaboration with Anthropic should help customers get even more value from these two capabilities.”

Anthropic — which also counts Spark Capital, Salesforce, Sound Ventures, Menlo Ventures and Zoom among its backers — has raised a total of $2.7 billion to date. The startup was valued at about $5 billion in May this year when it secured $450 million in a funding round. It didn’t say how Amazon valued Anthropic in the new investment.

The deal with Anthropic allows Amazon, which is increasingly flexing its own muscles around AI, to build a bulkier war chest in the frantically fast-growing industry.


AI-earnings-calls-mentions.jpg

Image and Data: Goldman Sachs


Anthropic chief executive and co-founder Dario Amodei told the TechCrunch Disrupt audience last week that he doesn’t see any barriers on the horizon for his company’s key technology.

“The last 10 years, there’s been this remarkable increase in the scale that we’ve used to train neural nets and we keep scaling them up, and they keep working better and better,” he said last week. “That’s the basis of my feeling that what we’re going to see in the next 2, 3, 4 years… what we see today is going to pale in comparison to that.”

Anthropic has made a “long-term” commitment to provide AWS customers around the world with access to future generations of its foundation models via Amazon Bedrock, AWS’s fully managed service that provides secure access to the industry’s top foundation models. In addition, Anthropic will provide AWS customers with early access to unique features for model customization and fine-tuning capabilities.
“Training state-of-the-art models requires extensive resources including compute power and research programs. Amazon’s investment and supply of AWS Trainium and Inferentia technology will ensure we’re equipped to continue advancing the frontier of AI safety and research,” said Anthropic in a statement. “We look forward to working closely with Amazon to responsibly scale adoption of Claude and deliver safe AI cloud technologies to organizations around the world.”
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787
What is the image equivalent of ChatGPT?


not sure I understand... :patrice:

chatgpt is multimodal meaning it can understand text, audio and images.


can you be more specific?
 

Roid Jones

Veteran
Supporter
Joined
May 1, 2012
Messages
53,209
Reputation
6,821
Daps
158,187
not sure I understand... :patrice:

chatgpt is multimodal meaning it can understand text, audio and images.


can you be more specific?

I know there are sites that if you type in a request an AI image is generated, I think one was posted on here via discord, not too sure of it's name

Thanks for the link, something like that is what I'm looking for, do you know of any more like that?
 

Voice of Reason

Veteran
Bushed
Joined
Jan 7, 2016
Messages
43,100
Reputation
262
Daps
122,102
not sure I understand... :patrice:

chatgpt is multimodal meaning it can understand text, audio and images.


can you be more specific?


He's talking about a DALL-E type model.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787
I know there are sites that if you type in a request an AI image is generated, I think one was posted on here via discord, not too sure of it's name

Thanks for the link, something like that is what I'm looking for, do you know of any more like that?
He's talking about a DALL-E type model.

the service on discord you're referring to is called Midjourney.

DALL-E is also by OpenAI.







 

Voice of Reason

Veteran
Bushed
Joined
Jan 7, 2016
Messages
43,100
Reputation
262
Daps
122,102
the service on discord you're referring to is called Midjourney.

DALL-E is also by OpenAI.











He was probably referring to mid journey specifically


The name escaped me as well.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787

ALERT

NSA, FBI, and CISA Release Cybersecurity Information Sheet on Deepfake Threats​

Release Date

September 12, 2023

Today, the National Security Agency (NSA), the Federal Bureau of Investigation (FBI), and the Cybersecurity and Infrastructure Security Agency (CISA) released a Cybersecurity Information Sheet (CSI), Contextualizing Deepfake Threats to Organizations, which provides an overview of synthetic media threats, techniques, and trends. Threats from synthetic media, such as deepfakes, have exponentially increased—presenting a growing challenge for users of modern technology and communications, including the National Security Systems (NSS), the Department of Defense (DoD), the Defense Industrial Base (DIB), and national critical infrastructure owners and operators. Between 2021 and 2022, U.S. Government agencies collaborated to establish a set of employable best practices to take in preparation and response to the growing threat. Public concern around synthetic media includes disinformation operations, designed to influence the public and spread false information about political, social, military, or economic issues to cause confusion, unrest, and uncertainty.

The authoring agencies urge organizations review the CSI for recommended steps and best practices to prepare, identify, defend against, and respond to deepfake threats.

This product is provided subject to this Notification and this Privacy & Use policy.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787

How to make history with LLMs & other generative models​

Or, I’m getting tired of market maps and am ready for some hotter takes​

LEIGH MARIE BRASWELL

SEP 21, 2023



NVIDIA HQ - apparently GPUs aren’t a bad business these days
https://substackcdn.com/image/fetch...925-73a4-4d6b-8532-84aadb8ce424_1600x899.jpeg

It’s been well over a year since I published my overview of large language models, or “LLMs”. The magic of Github Copilot’s beta in early 2022 made me the most excited I’d been about a new technology since I got my first iPhone. I was (and remain) convinced that we are at the start of something big - big enough for me to write a large list of LLM application ideas and watch as many others published their large lists in subsequent months. There’s nothing wrong with a great market map, filled to the brim with potential startup ideas, but it inevitably leads to the question of where to focus. Sure, there are all of these opportunities, but let’s be real, what are the ones that a smart startup founder should go after? And which are flash-in-the-pan, waiting to be made irrelevant by a foundation model provider, a new architecture shift, or an onslaught of other early stage startups?


Over the past year I’ve talked to many founders, operators, and investors, and I’ve encountered the full spectrum of opinions regarding where new LLM/generative model-related $10B+ companies will be built. On one side are those who believe there are many opportunities to build “AI-native” companies (similar to the generational building opportunities from the move to the cloud) with feelings best summarized by this tweet:


vjzm8WA.png


On the other side are those more convinced that incumbents (applications and/or infrastructure companies) will capture most of the value from adding or supporting this new generative tech. I’ve even heard opinions as extreme as believing that all of the foundation model development and usage will become so commoditized that building chips or supplying power will be the only long-term advantages. In terms of where I land personally as an investor, I’m in the middle. I don’t think that every incumbent company will be disrupted by a new startup, but I believe some are more likely to be disrupted than others. I believe some infrastructure categories offer a better risk-return than others, as parts of the “LLM stack” appear relatively more resilient. In this post, I want to expand on some ideas that I’m particularly excited about and others that I’m less certain of reaching venture-scale as standalone businesses. I’ll also mention some companies I know that are working on each idea, but this is not meant to be a complete list (of companies or ideas- there are so many other exciting stealth companies for a potential future post…). Lastly, these are strong opinions, loosely held. I would love to be convinced to change my mind, and I believe it’s a sign of a great founder to be able to address risks, navigate the idea maze, and prove skeptics wrong.

Promising generative model-related startup application ideas

Developer tooling platform


The success of Github Copilot (>30% acceptance rate for auto-generated suggestions and rumored 100,000’s of users or >27,000 organizations paying for the product) has proven that current technology can augment even the best programmers. It’s now a no-brainer for most developers to have some sort of LLM-powered autocomplete, whether VSCode with a Copilot extension, ChatGPT copy-pasting, or a variety of competitors that have emerged over the past year to address Copilot and ChatGPT’s shortcomings. Given the massive size of the opportunity to speed up the world’s ~30m developers as well as gaps between what LLMs can be used for and the current Copilot product surface area, I am excited about AI-enabled dev tools startups’ potential. LLMs can already be at least somewhat helpful for code search, delegating larger coding tasks (e.g. refactoring), testing (unit, integration, QA), code review, debugging / observability, setting up infrastructure, documentation, and security. In addition, many companies want to self-manage anything related to their code (surprisingly, the majority of Gitlab’s revenue comes from self-managed code repositories) which is not supported by Copilot, and some have broader security concerns around producing GPL code. Copilot also doesn’t support every IDE / notebook and doesn’t allow for fine-tuning or other customization to better understand or maintain bespoke codebases. Startups like Codeium*, Grit*, Warp, Sourcegraph (Cody), Cursor, & Contour have addressed many of these problems. How acute the pain point each initial wedge solves, the effectiveness of their products (e.g. latency, UX, model infrastructure tradeoffs), and the customers’ willingness to pay will likely determine the winner(s) here. Though I believe the success probability is reasonably high for at least one new company (large market & opportunity to build a differentiated product), the main risk considerations would be either open source commoditizing most of these workflows (e.g. Meta’s Code Llama) or the incumbent, Github / Open AI, developing proprietary model advances and aggressively leveraging existing distribution to capture the market (we’ll see…).
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787
Augmenting knowledge workers (consulting, legal, medical, finance, etc.)


Early signs point to many simple back-office, front-office, and even customer-facing tasks being automated completely by LLMs - Adept AI & other startups excel at simple personal and professional tasks, like locating a specific house on Redfin or completing an otherwise click-heavy workflow in Salesforce. In addition to text & webpages, voice is understandable and replicable; for example, Infinitus* automates B2B healthcare calls, verifying benefits and checking on statuses. More specialized knowledge workers - especially those in legal, medical, or consulting professions - can also be made drastically more efficient with LLM-powered tools. However, given their plethora of domain-specific knowledge, the majority of higher-stakes workflows I’m referring to here are more likely to be assisted than fully automated in the near-term. These knowledge workers will essentially be paying such startups for small (but growing!) tasks to be completed inside of their complicated day-to-day workloads.


Whether drafting a legal document for a transaction or PI case or analyzing a contract for due diligence, some lawyers are already using legal assistant tech to save time. Thomson Reuters, a large incumbent tax & accounting software platform, saw so much potential in Casetext, an AI legal assistant, that they recently acquired it for $650 million. Given gaps in existing legal tech and the potential of LLMs to speed up workflows, there’s potential to build a larger legal software platform from the various initial automation wedges. However, startups must navigate finding champions and validate lawyers’ willingness to change personal workflows & ultimately pay for large efficiency gains.


In medicine, doctors can have more leverage with automatic entry of patient data* into their electronic health records after a meeting (especially important with the turnover of medical scribes) as well as automated patient or hospital Q&A through chatbots. Biologists are also already taking advantage of LLM-powered tools to help them find protein candidates faster. Though scaling medical GTM is notoriously challenging, the payoff of saving large amounts of time for these highly educated personas could be immense.


Finally, the consulting industry continues to boom, helping businesses make all kinds of decisions from pricing models, store placement, inventory & risk management, and forecasting. Startups like Arena AI*, Kumo AI, Unearth insights, Intelmatix, Punchcard*, and Taktile use LLMs and other related tech to help many different types and sizes of customers with decision-making. If a startup is able to build a generalizable product with a scalable GTM - so, not just another consulting company - they might be able to eat into some of the large consulting spend as well as the budgets of those who didn't use consultants in the first place.


Digital asset generation for work & for fun


Other types of generative models outside of LLMs (e.g. diffusion models) enable the generation of media outside of text like images, videos, and audio. Whether you draw portraits, edit videos, or make PowerPoint presentations for a living, the current state of generative models can likely help you become more efficient. Separately, if you thought you weren’t skilled enough to create images or songs at all, some AI-powered generation tools may convince you otherwise - similar to how Canva made graphic design more accessible for many non-artists years ago. Startups like Midjourney, Ideogram, Genmo, Tome, Playground, and Can of Soup help users create and share images for professional or personal use. Some may continue to build out enterprise features and challenge Adobe and Microsoft, while others may continue to build out social media engagement & e-commerce capabilities through network effects & ads. Video creators & editors - from Instagram stars to blockbuster special effect artists to L&D professionals - can speed up & reduce the cost of their work with products such as Captions*, Wonder Dynamics, Runway, Hypernatural, and Synthesia*. On the cutting-edge of the current generative tech, short-form video (e.g. Pika, Genmo), music (e.g. Oyi Labs, Frequency Labs*, Riffusion), & 3D asset generation (e.g. Sloyd, Rosebud*) show promise, though the longer-term business plans seem less straightforward than those of the image generation and video editing companies. In addition, some digital asset generation like audio/voice seems more challenging to maintain a differentiated product over time, especially as cloud providers expand their offerings. As a final note about this category, the legal and copyright issues are most pronounced here in comparison to other categories, as there are already many lawsuits alleging improper training data and unattributed output.


Personal assistant & coach


I’m convinced that we will eventually have the option for a LLM-powered assistant or coach for the majority of things we do, both at work and in life. I’d personally love a future where wearing some sort of AR device is somewhat socially acceptable, and my device can listen to me talk with a founder, fact-check the conversation in real-time, give me advice on how I could be more helpful or convincing, and automatically follow-up for me after. In the meantime, tools to help with writing persuasive emails, navigating internal knowledge bases*, or automating common tasks in the browser seem appropriate and ripe for expansion. The current LLM tech can also already perform well enough to help learners, in school and out, with personalized educational solutions and conversations. Using large models to create compelling, seamless experiences on mobile is quite challenging given latency & compute requirements, which likely makes great products here harder to copy than meets the eye.

Generative model-related startup application ideas I’m less certain of

Some other SaaS replacements


In general, I’m more skeptical of any SaaS disrupter that doesn’t have a strong story against the incumbent and other upstarts in the space. To truly claim LLMs or generative models as the “why now” for a new startup, I’d prefer the existence of some sort of innovator’s dilemma, large product rework, and/or special unattainable resource (e.g. talent) that makes incumbent repositioning challenging. When in doubt, I go to a favorite book of mine, “7 Powers”, on how to build and maintain a moat.

 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787
Summary of “7 Powers: The Foundations of Business Strategy”

As an example, given the how tech-savvy and agile document incumbents like Coda* and Notion are, it’s harder for me to believe that new directly competitive LLM-powered companies will be able to take away significant market share. In fact, they’ve both already added AI-powered features to their platforms, and it seems to have had material impact on revenue/engagement from their existing users.


I hear pitches for AI as the main catalyst for new Salesforce-competitive sales platforms, BI & data science tools, CAD software, and basically any SaaS tool that you’ve ever heard of. In many cases, I don’t think that the timing argument for a new company is the current AI wave. I am generally optimistic about companies consolidating things like sales, BI, & CAD tooling and making it easier for various personas to use (Salesforce and Autodesk are challenging to use), but I believe AI is at most a feature for some, not the key differentiator or main reason for incumbent disruption. Nonetheless, AI will likely help all product experiences improve over time, and particularly AI-native teams may be able to use AI more effectively internally to build and sell faster than their competition.


Standalone general consumer search


In theory, the chat experience that made ChatGPT explode in popularity (reaching 100m users in record time) should disrupt the traditional model of ads through bidding on search keywords. However, whether it’s due to a lack of desire of consumers to change ways, the inconvenience of incorrect information and difficulties fact checking LLMs, or something else, even the Bing AI-powered chat thesis hasn’t played out in a significant way yet, In fact, Bing’s market share is lower in 2023 than 2022. Moreover, given the importance of search to Google as well as their ability to train LLMs internally, I’d be skeptical that Google doesn’t continue to adapt, as they’ve finally rolled out some generative model features in search. Ultimately, I believe the incumbent’s distribution for general-purpose consumer search seems challenging to go up against, even with a counter-positioning argument and a better product experience.


Promising generative model-related startup infrastructure ideas

Running large models locally


If you imagine a world with all of the wonderful LLM applications discussed above, you probably assume some sort of infrastructure to run increasingly large personalized models on various edge devices - laptops, phones, cameras, sensors - with minimal lag. Today, if you tried to run Llama 2, the state of the art open source LLM as-is on a laptop, it would likely be impossible. However, if you use the GGML version of Llama 2, all is solved. GGML is an open-source tensor library for ML that optimizes (quantizes) models to run on CPUs instead of just GPUs, giving a massive boost in inference speed for a minimal accuracy tradeoff. Even Meta runs Llama internally by using the GGML versions and saves “a lot of money” as a result. GGML’s vibrant contributor community seems focused on building and optimizing GGML versions of popular & consistently improving models, perhaps with longer term aspirations to build a full-fledged edge inference framework. There’s likely an opportunity for GGML and/or another company (e.g. Ollama) to offer paid extensions or tooling on top of the popular related open source here given the growing interest in LLM local usage.


Providing compute & software for model training / fine-tuning / inference


Everyone needs GPUs these days, and GPU cloud providers like Coreweave and Lambda Labs are exploding in growth. Startups and even investors are hoarding chips. Many companies wish to use LLMs and other generative models, but they want to own their own models and therefore can’t just call something like OpenAI’s APIs. Instead, these companies must figure out how to best utilize compute for initial training or fine-tuning of open or closed source models as well as for ongoing inference. Given the complexities of managing this infrastructure as well as the opportunity size (many companies would rather own their own custom models for quality, strategic, and/or security reasons), a variety of startups have popped up. Some seem more focused on inference (Modal*, Banana, Replicate, Runpod, Fireworks), others on fine-tuning (Scale AI*, Lamini, Automorphic AI), and others as more general purpose (Mosaic [acquired by Databricks], Together, MLFoundry, Hugging Face). I don’t doubt this is a valuable market to compete in, and some of these startups with proprietary tech can offer much easier-to-use and more cost-effective experiences than cloud providers. The winner(s) here may not just need to have a superior product & internal cost / margins, but also financial savviness and GTM excellence, especially if the competition continues to heat up. In addition, many “outside” forces – the demand for GPUs, the behavior of NVIDIA & cloud providers, and the macro for raising money to purchase or lease GPUs – probably have a significant effect on the outcome.
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
51,832
Reputation
7,926
Daps
148,787
New ML framework and/or new chip

A particularly ambitious but exciting infrastructure category is challenging NVIDIA’s dominance in both a free ML framework (CUDA) and associated chip. Combining the shortage of NVIDIA GPUs, the general pain of hardware lock-in, and CUDA’s usability and latency challenges, a disrupter could theoretically break into the market. Both Modular and Tiny Corp are working on new challenger ML frameworks - the latter then plans to develop a competitive chip while the former seems to be commercializing some associated software. Many other startups are skipping straight to building chips for specialized LLM workflows.

Generative model-related startup infrastructure ideas I’m less certain of

Observability


Though observability is a well-defined term & large market in traditional software, attempts at standardizing & monetizing ML observability products have proven more challenging. From working with many potential buyers at Scale during the peak of the self-driving computer vision hype years ago, I found that the typical ML engineer’s attitude was much different in regards to model monitoring tools compared to labeling. ML engineers generally hate to manage any type of operational process like labeling, but when it comes to software for observing and correcting ML models, they tend to have stronger and more diverse opinions about what is best to build. I imagine this may be true of the current LLM wave, where many LLM application companies I speak to or work with build and maintain their own LLM monitoring. Many even consider it a competitive advantage - no time is spent communicating their ever-changing needs with vendors. If an external infrastructure startup becomes successful here, they’d need to build a customizable product that adapts to the frequently-changing best practices as a result of the fast pace of AI developments. In addition, the growing prevalence of “AI engineers” versus more traditional ML engineers of the past likely makes my personal Scale anecdotes less applicable.


Vector databases


Much has been said about the uses and benefits of vector databases - they are useful in RAG, which is a standard for many LLM applications involving specific information retrieval today. However, as fine-tuning becomes more commonly used and context window sizes continue to grow, the urgency of having a highly performant vector database diminishes. Combine that with an already fierce set of startup competitors, incumbent solutions like MongoDB & Postgres, and perhaps cloud providers at some point, it’s a tough space; new startups (especially those without a special advantage or platform play) face technical headwinds and a variety of adversaries.


Privacy or quality-related middleware


There are a host of issues with current LLMs including data privacy, security concerns, hallucinations, bias, and even output format. A viable strategy for building a general-purpose model creation platform may be wedging in by solving one of these pain points that is particularly acute for a specific group of users, but I’d be skeptical of a large standalone company being built in the near-term without a compelling expansion strategy. I imagine the foundation model providers as well as companies helping their customers train & refine open source models will incorporate an increasing amount of these security and quality guarantees, so fast competition attempts following successful traction is likely. It’s also always challenging to predict when customers will generally really care about privacy / security (e.g. after regulation or industry scares), which can be risky if that’s the timing argument for a company’s growth. In addition, as vertical LLM applications continue to mature, companies could choose to use those, with privacy & quality guarantees, instead of managing foundational model APIs or their own models internally.


Thanks for reading! It’s always fun to theorize what the future holds, but I will conclude the post with the following saying from Henry Kissinger:

“Traveler, there are no roads. Roads are made by walking.”

The only way to really know what will be the most incredible AI companies of the future is to try building them, and I can’t wait to continue to partner with awesome founders doing so at all stages. I would also love to hear thoughts & feedback on anything in this post - leighmarie@kleinerperkins.com.
 
Top