HACKER Q&A
📣 brucethemoose2

Is GenAI heading for a crypto-esque bubble pop?


Just poking around HuggingFace, I see a whole lot of models and finetunes claiming to be SOTA in this and that... But not a whole lot of validation. In fact, the biggest LLM validation effort seems to be with llamav1 and v2, and testers had enormous difficulty reproducing Meta's metrics.

Many of these finetunes or base models are dumped with little info and private training datasets. How much of those datasets are questionably legal or ethical? How many are accidentally or intentionally contaminated with test data? No 3rd party ever seems to test for contamination outside the rare social media post.

I see a whole lot of frameworks, papers and APIs claiming SOTA transformers/diffusion performance in this or that, but they seem barebones and janky when one actually goes and investigate them, if you can get them to work at all. Or they are grand promises asking for money now and usability later.

...I am no data scientist, but I am a straight up AI fan. I was finetuning ESRGAN 1x models for img2img/vid2vid years ago. GenAI is the best thing since sliced bread as far as I am concerned. Yet I am getting a serious crypto fraud vibe from everything that is happening, albeit one more targeted at businesses and investors than the crypto wave.


  👤 behnamoh Accepted Answer ✓
Forget about frameworks. Langchain is to GenAI what Doge was to crypto.

Most models on HF are like CrapCoins of the crypto era, so forget about 99.5% of them. There are only a few that can be useful (mostly +70b params).

Don't forget that some of this hype can be driven by nVidia to sell more GPUs. Some of it is driven by crypto enthusiasts who now use their GPUs for GenAI.

Forget about benchmarks and SOTA performance claims. Data leaks and you'll never be 100% sure that the team didn't mess with training data to manipulate the benchmarks.

Is GenAI a hype? Most likely, yes. Can it still be useful? Yes. You can still use GPT-4 and Llama 2 70b models for various tasks. Just don't use Langchain or any frameworks. Make your own API calls. And the smaller models (3-7b)? They are useful for when you don't want to chat with the model, like sentiment analysis.


👤 cratermoon
Yes.

👤 ldjkfkdsjnv
I actually think its the opposite, the next generation of models will rock the world.

👤 paulpauper
there are hardly any publicly traded AI companies or highly valued ones, compared to the multitude of tech companies and crypto companies. Crypto is way worse because average people lose money, not just institutional investors. AI companies make actual revenues from business activity, compared to crypto, which is just speculation mostly.

👤 1-6
Drawing a comparison to cryptocurrency is shortsighted. Crypto, unfortunately had a very strong rival: Country-run fiat.

AI doesn’t have any rivals, perhaps faster and lower-power chips but we haven’t seen all the implications yet.


👤 joefourier
I wouldn’t compare GenAI to cryptocurrency because there’s no equivalent pyramid scheme mentality of “invest early in this asset and sell it later 100x” - it’s much closer to a traditional tech hype bubble. The average user isn’t roped in by the promise of buying and reselling a token, but by using and possibly paying for a product that can theoretically be useful to them on its own.

Instead this seems like the standard tech hype cycle where crowds of investors pounce on the “next big thing” which will inevitably be overhyped and oversold, leading to a bubble pop when expectations outpace actual returns, but the core technology is valid. The dotcom bubble popping didn’t mean the Internet was a fad, the video game crash of 1983 didn’t permanently doom the industry, the railway mania in the 1840s didn’t mean no one ever took a train after it crashed… any new and exciting development attracts grifters but it eventually stabilises once people learn to separate the hype from the real applications of the tech. I remember when VR went through this - it was absolutely overhyped, but today there’s still a solid community around it even if it didn’t change the world.

With open source models and fine-tunes being uploaded on Huggingface, there’s no fraud, they’re perfectly free, and you can test them out and see for yourself that some work better than others even if benchmarks only tell some of the story. There’s absolutely a glut of thin veneers over APIs and models offering questionable benefits, but there’s also real value in having a solid UI on top of a basic model.


👤 simonw
The reason we're seeing so many models right now is that the field is WIDE open for big leaps forward.

Prior to LLaMA in February there were almost no openly licensed language models that were any good at all - BLOOM and GPT-J were pretty much the best of the bunch.

Then LLaMA happened and the flood gates opened - there was now a base model which people could fine-tune against to get really great results.

Llama 2 accelerated things even more because it's licensed for commercial use - so now it's worth spending real money building on top of it.

Llama 2 isn't even 2.5 months old at this point. I'm not at all surprised at the rate of innovation around it - there is SO MUCH scope for improvement, and so many new techniques to be explored.

I expect we'll see a slow down in the release of models at some point, once the low hanging fruit has all been picked off - but it's going to be a while away yet.


👤 sethkim
My two cents is that we will see a bubble pop, but it will rebound massively much like SaaS following the dotcom crash.

My primary concern is the ratio of AI-infrastructure products being built relative to the average utility delta of current end-user products. There are some shining counterexamples to this like Github copilot, but for the most part the utility gain from most text-based generative-AI seems incremental.

Current language models are bad to okay at most tasks - and great at a select few. The capabilities in unstructured ETL and production of labeled datasets fall in the latter category.

I think the next generation of really powerful capabilities to be unlocked with generative-AI will come with 1) an increase in compute availability (training even larger models) 2) advances in multimodal models, and 3) better interfaces to interact with them. Spatial computing will be a really big deal here, and people seem to have forgotten that Apple may have unlocked the next version of interactive computing earlier this year.


👤 syntheweave
Like with crypto before it, there's a tiny core of good stuff in genai, and an industry of rent-seekers adjacent to that using it for cover.

The competition for AI is simply any solution that gets a higher quality of answer than a trained approximation. That's actually a lot of things, but often there's a chain of dependencies in the information lifecycle that lead to combinations of technologies, and so our total solutions frontier expands by having the tech as a "drop-in" option that would otherwise be some custom engineered thing. But the AGI hype, to name one "unlikely AI thing", is silly. Based on what we know, larger genai models are likely to be more normal, not more novel. And normal isn't right or good, it just tells us what we want to hear.

Crypto is actually covering a polar-opposite technological context: it lets you get very high quality answers within a very narrow space. Done right, it tells you what you DON'T want to know.


👤 cloudking
There is definitely a lot of hype and cherry pick demos showcased on social media, but the latest language models actually provide real productivity value unlike crypto. So it's not really fair to draw that comparison.

Personally, I haven't found a tool like ChatGPT+ that instantly improved my productivity overnight. Brainstorming, writing code, documentation, marketing copy, transforming data etc all became easier with this tool. I'd rather spend a shorter amount of time providing clear instructions to a language model than writing a lot these things from scratch.

Anecdotally, I've tried all the major models from Google, Meta, Anthropic, Stability etc nothing comes close to GPT-4 in understanding instructions properly.


👤 danenania
Maybe the explosion of new models is somewhat comparable to crypto. Models seem destined to become commoditized and have the market carved up by a small group of major players. So the vast majority of the new ones probably aren't going to get any traction.

But I think applied AI, AI saas, and so on built on top of the best models to optimize specific workflows has a pretty bright future. There will be failures and false starts as with any new technology, but there are already a number of useful apps out there with wide adoption, which is more than crypto ever achieved.

With crypto, the product is speculation. There are some potentially interesting applications for crypto, but it's never gotten large-scale traction as anything other than a form of gambling. AI doesn't have speculation built in like crypto does. It has to actually solve problems and do useful things to make money, and this is already happening.

There may be a speculative bubble in the VC realm, but when isn't there? It happens literally every time a new startup trend comes around.


👤 tkgally
My guess is that it’s heading for a late-1990s-Internet-esque bubble pop. Some overly hyped companies will go belly up and some investors will lose their shirts, but the technology will grow only more powerful and the world will be transformed by it.

👤 bbstats
mean crypto utility is basically zero. mean AI utility is decent.

👤 drt5b7j
The fact that you can't type this post without using so many acronyms that only insiders can read it suggests yes.

👤 NBJack
I have yet to find LLMs useful beyond basic factoids (I have plans to deploy Llama 2 as an Alexa replacement), but the latest Stable Diffusion model (XL) has continued to impress me. It is clearly a game changer in the utility it provides to my interests. Admittedly, I'm having fun making high quality 8k wallpapers of abstract concepts and fictions places, but a truly honed creator could clearly work wonders with it.

Bitcoin is a solution in search of a much bigger set of problems than it can currently solve now. The ML Renaissance we've hit right now has clear, immediate utility to many. The moves made by the stock art heavyweights is proof enough of how disruptive it is.

Contrast this with how banks and financial institutions have responded to bitcoin. To my knowledge, there's yet to be a Bank of America Coin, a Wells Fargo blockchain solution, or non-fungible tokens offered by RBC. They clearly felt at least somewhat threatened, but not in any meaningful way.


👤 Racing0461
Yeah. as google said, openai has no moat. hope we get alot of competition in this area since there is so much censorship.

👤 rvz
Yes.

The downrounds and devaulations for the ChatGPT wrappers are coming in [0] and realizing that most of these 'AI startups' are overhyped. Especially the 'Chat with PDF' web apps and basic copywriting apps.

To suggest otherwise is pretty naive, and the large AI providers (OpenAI, Google, etc) do compete against their own partners.

[0] https://archive.ph/XZvmp


👤 danielmarkbruce
No. LLMs are underhyped.

There are a bunch of models because it's easy to grab llama2 etc and do a little bit of fine tuning. There is nothing wrong or bubbly about that.


👤 tbrownaw
No, it'll pop more like the dotcom bubble. Like, yeah it's a bit overhyped at the moment, but there actually is a fair bit of there there.

👤 dstainer
In my opinion, which doesn't mean much, we're in a very weird spot where it's like 50/50 on usefulness and uselessness. There are certainly approaches where GenAI can be incredibly helpful, I've seen it in action in the day job. However, there are also avenues where it's complete snake oil.

👤 echan00
No, we're just getting started with GenAI. Crypto only affects a number of people and industries, GenAI affects a much larger proportion.

👤 retrocryptid
Yes.

👤 j45
I don't think GenAI is going to fail to get off the start line, because it has done just that.

It has adoption that might not be every person in the world, but the depth and breadth and variety of users is hard to ignore.

Crypto wallets and getting up and running is the failure for most crypto. Yes, there have been improvements to that.

LLM's that took a chat interface went as universal as they could - just a computer anyone can talk to.

One issue with looking at HuggingFace and expecting maturity in 10-12 months is I'm not sure what can be compared as an equal in that number of months in other technologies. I agree it's far from perfect, but also it seems to be improving.

LLM's are just one kind of GenAI as you have pointed out. The AI you were working with before all of this remains relevant.

There's more to this than wanting a prompt to do everything magically. Having to "fine tune" and get each piece working may be the work for a while, and still be usable and beneficial.

Does it need to evolve and get better? It is.

With each passing month are more and more models running on more accessible hardware? It is.

Are people quietly doing a lot with it and not sharing it yet? Sure feels like it.


👤 mindwok
What bubble? If you mean the hype bubble of every company trying to build LLMs into their product, then yes. IMO LLMs will render a lot of existing software obsolete or at least so fundamentally change the way we interface with them that the incumbents are likely to perish to newer, more innovative companies. If you mean a stock market bubble... I don't see one of those at the moment.

If you mean the bubble of excitement for AI, then I highly doubt it. LLMs are paradigm shift in how we use computers and what we can accomplish with them, and we've barely scratched the surface of them.


👤 gmerc
Deja vu. Is the internet a fad was a very common question in the early 2000s too

👤 Timber-6539
Generative AI will at some point have to plateau, if it hasn't already. About the bubble pop, the (real) world is no different today than from when ChatGPT launched. This is exactly the same effect Bitcoin has had since the launch of its whitepaper.

👤 kromem
Yes and no.

Generative AI as a technology isn't going anywhere and is very much going to increase in scope, scale, and necessity in most workflows.

But the vast majority of AI companies that cropped up riding the swell are going to be out of business in the near future.

Which is pretty standard. 90% of new companies aren't still in business a decade later no matter the field.

It's just that there's been a lot more new companies in a single new space since maybe 1999.

So expect a significant 'pop' of a bubble without there being any underlying issues with the tech or its applications.

(Possibly even more than typical given the rate of change and likelihood any custom product right now is building into its own obsolescence.)


👤 notsoSame
Gen AI is merely one component of AI. Crypto and web3 are what you see what you get.

There’s still a lot of evolution to be had with AI; for example gen AI to interactive AI that uses all the bits we’ve seen so far to compose experiences with more than chatbot interaction.

If anything is going to die in AI it’s all the bespoke projects, not AI as a whole.

Am working on a Linux distro that ditches userspace as we know for a local LLM (root login is still available). The goal is no user accounts, /home and the AI encodes user data as vectors to disk. Need to successfully unlock the vectors rather than a user account.

So far it’s Linux From Scratch, Wayland, hyprland which displays the GPU viewport.

There is still a lot of weird stuff to do with AI. What do I use blockchain for except as yet another fiat currency