Hacker News new | past | comments | ask | show | jobs | submit login

Just to get things right. The big AI LLM hype started end of 2022 with the launch of ChatGPT, DALL-E 2, ....

Most people in society connect AI directly to ChatGPT and hence OpenAI. And there has been a lot of progress in image generation, video generation, ...

So I think your timeline and views are slightly off.






> Just to get things right. The big AI LLM hype started end of 2022 with the launch of ChatGPT, DALL-E 2, ....

GPT-2 was released in 2019, GPT-3 in 2020. I'd say 2020 is significant because that's when people seriously considered the Turing test passed reliably for the first time. But for the sake of this argument, it hardly matters what date years back we choose. There's been enough time since then to see the plateau.

> Most people in society connect AI directly to ChatGPT and hence OpenAI.

I'd double-check that assumption. Many people I've spoken to take a moment to remember that "AI" stands for artificial intelligence. Outside of tongue-in-cheek jokes, OpenAI has about 50% market share in LLMs, but you can't forget that Samsung makes AI washing machines, let alone all the purely fraudulent uses of the "AI" label.

> And there has been a lot of progress in image generation, video generation, ...

These are entirely different architectures from LLM/chat though. But you're right that OpenAI does that, too. When I said that they don't stray much from chat, I was thinking more about AlexNet and the broad applications of ML in general. But you're right, OpenAI also did/does diffusion, GANs, transformer vision.

This doesn't change my views much on chat being "not seeing the forest for the trees" though. In the big picture, I think there aren't many hockey sticks/exponentials left in LLMs to discover. That is not true about other AI/ML.


>In the big picture, I think there aren't many hockey sticks/exponentials left in LLMs to discover. That is not true about other AI/ML.

We do appear to be hitting a cap on the current generation of auto-regressive LLMs, but this isn't a surprise to anyone on the frontier. The leaked conversations between Ilya, Sam and Elon from the early OpenAI days acknowledge they didn't have a clue as to architecture, only that scale was the key to making experiments even possible. No one expected this generation of LLMs to make it nearly this far. There's a general feeling of "quiet before the storm" in the industry, in anticipation of an architecture/training breakthrough, with a focus on more agentic, RL-centric training methods. But it's going to take a while for anyone to prove out an architecture sufficiently, train it at scale to be competitive with SOTA LLMs and perform enough post training, validation and red-teamint to be comfortable releasing to the public.

Current LLMs are years and hundreds of millions of dollars of training in. That's a very high bar for a new architecture, even if it significantly improves on LLMs.


ChatGPT was not released to the general public until November 2022, and the mobile apps were not released until May 2023. For most of the world LLM's did not exist before those dates.

LLM AI hype started well before ChatGPT.

This site and many others were littered with OpenAI stories calling it the next Bell Labs or Xerox PARC and other such nonsense going back to 2016.

And GPT stories kicked into high gear all over the web and TV in 2019 in the lead-up to GPT-2 when OpenAI was telling the world it was too dangerous to release.

Certainly by 2021 and early 2022, LLM AI was being reported on all over the place.

>For most of the world LLM's did not exist before those dates.

Just because people don't use something doesn't mean they don't know about it. Plenty of people were hearing about the existential threat of (LLM) AI long before ChatGPT. Fox News and CNN had stories on GPT-2 years before ChatGPT was even a thing. Exposure doesn't get much more mainstream than that.


> LLM AI was being reported on all over the place.

No, it wasn't.

As a proxy, here's HN results prior to November, 2022 - 13 results.

https://hn.algolia.com/?dateEnd=1667260800&dateRange=custom&...

Here's Google Trends, showing a clear uptick May 2023, and basically no search volume before (the small increase Feb. 2023 probably Meta's Llama).

https://trends.google.com/trends/explore?date=today%205-y&ge...

https://trends.google.com/trends/explore?date=today%205-y&ge...

As another proxy, compare Nvidia revenues - $26.91bln in 2022, $26.97bln in 2023, $60bln 2024, $130bln 2025. I think it's clear the hype didn't start until 2023.

You're welcome to point out articles and stores before this time period "hyping" LLM's, but what I remember is that before ChatGPT there was very little conversation around LLM's.


If you're in this space and follow it closely, it can be difficult to notice the scale. It just feels like the hype was always big. 15 years ago it was all big data and sentiment analysis and NLP, machine translation buzz. In 2016 Google Translate switched to neural nets (LSTM) which was relatively big news. The king+woman-man=queen stuff with word2vec. Transformer in 2017. BERT and ELMo. GPT2 was a meme in techie culture, there was even a joke subreddit where GPT2 models were posting comments. GPT3 was also big news in the techie circles. But it was only after ChatGPT that the average person on the street would know about it.

Image generation was also a continuous slope of hype all the way from the original GAN, then thispersondoesnotexist, the sketch-to-photo toys by Nvidia and others, the avocado sofa of DallE. Then DallE2, etc.

The hype can continue to grow beyond our limit of perception. For people who follow such news their hype sensor can be maxed out earlier, and they don't see how ridiculously broadly it has spread in society now, because they didn't notice how niche it was before, even though it seemed to be "everywhere".




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: