Creating LLMs or AGI-style models requires massive compute and data, which startups are unlikely to have. Therefore, incumbents have a huge advantage when it comes to general AI. This leaves the option of using an API or similar to create a startup in a niche, but it's difficult to create a moat with such a startup, and the incumbents keep innovating and creating their own services that often make these startups obsolete.
Therefore, an "AI startup" would do best to develop domain expertise (or have a co-founder with domain expertise), create a useful product in that domain, collect data from users, and finally use the data to create a useful domain-specific narrow AI. Many software engineers want to create developer tools with AI, as this is the domain they know best. But this is precisely the domain that is most likely to be oversatured with AI tools, because AI people already tend to be developers who know about software development.
Are there some flaws in this thinking? Do you agree/disagree? I'm curious to see what HN thinks.
In particular I'm wondering what the best way to acquire this domain expertise is for a technical (CS) person, and whether it's necessary at all, or if it's better to learn as you go or find a cofounder in a non-computer domain.
I had the chance to talk with one of the healthcare startups, they're trying to make a product that competes with Dragon DAX Copilot (which is incredible) by cobbling together some of the tools from Azure, Google Cloud Platform, and AWS. I didn't believe their pitches, and their prices were insane, because they got some seed funding and are building an MVP still, while also trying to get some revenue. They shared with me a document that was confidential because anyone with technical know how could see they had nothing unique, they were only shuttling data between a couple cloud providers to utilize different pre-built technologies. When confronted with this they got rather defensive and moderately offensive, and even tried to end-run around IT by going to providers and filling their heads with nonsense.
Right now AI can do 10% of what people promise it can, and it's so new audiences don't have the tools to know who is real and who is a fly by night operator. It's like the beginning of the app store, AI is full of incredibly low effort apps right now and most are garbage. People are rushing in like it's a gold rush, they want some easy money before bigger, more competent products come out.
I'd say this is the biggest issue, too many ethically challenged people are spinning up cheap minimum-viable-product apps and slick marketing materials to take some cash off the table before collapsing. The industry is full of snake oil.
Don't be so blinded by the tech that you forget what the point of a company is.
For related reasons, I'm bullish on growing a consulting arm of most b2b startups to help accelerate this process + figuring out profitable scalable revenue.
You definitely do not need to train your own base model to be successful, but if your entire pipeline is a system prompt or three in a small agent graph... You didn't build a product, you built a hobby tool over the weekend no matter how much UX/UI polish you put on it.
I do think this question, and many start-ups are thinking about how they want that sweet AI money and are starting their business from there rather than from a problem. If you see a problem that is labor intensive and could be done by mechanical turk... Well you're probably on to something and an AI language model can probably solve that problem.
The companies and AI products I think that are going to last either aren't starting with AI they're focusing on a problem and have reasoned their way to AI OR they're doing some deep stealth research into the problem over a long period of time to be able to develop domain-specific data, techniques, and plans so they can fine-tune their own model before ever showing it to a potential customer. You better be sure you have a good plan if you're going to try and be the latter.
If you build a good system to collect hard-to-gather, rich proprietary data then improvements in AI will help you squeeze more and more insights out of it.
However, on that subject I would still throw out a counter argument. AGI might not be a matter of “throwing infinite money at the problem”. It might be that the building blocks already exist but need to be arranged in the right combination under the right circumstances to create autonomous agents.
While you won't be making anything even close to llama1 it is possible to develop technologies that can be used to train llama4.
I've personally been using gh copilot and it's an amazing tool for the 'braindead' parts of my workflow. I usually let it generate a rough template and then fill in the parts by entirely replacing them, but it at times manages to do exactly what I wanted when working in larger files which already have much of the functionality defined and it's rather trivial to figure out the rest.
Copilot does what it says, it's a copilot. It feels like I am iterating over a problem with a 2nd developer suggesting ideas as I code.