Ah so the sky isn't the limit? Or maybe they have sucked up all the atmosphere (training data) so the sky is a hard limit? This could be a temporary bump in the road. But, given how iterative most...
Ah so the sky isn't the limit? Or maybe they have sucked up all the atmosphere (training data) so the sky is a hard limit?
This could be a temporary bump in the road. But, given how iterative most LLM development seems to have been I personally think we might have reached a plateau with what can be achieved with these types of models. At the very least it is becoming clear they are not improving at the rate the hype seems to suggest. Although that shouldn't come as a surprise for anyone who has paid any attention to the actual improvements between gpt-4 versions.
Edit:
Another thought I had. If all the training data in the world is not enough to create the next step forward, then maybe these models are indeed not the way forward for anything remotely resembling actual intelligence and more importantly ability to improvise. This should also not be a surprise for most people who have paid attention.
To be clear, because I know people will misread what I wrote, that doesn't make these models useless. I quite enjoy my fancy rubber ducky who can actually talk back and pitching two models against each other. Like the juniors our team needs but never gets,while also not having to be gentle with their emotions when the code sucks ;). But, at the end of the day these models are just tools and like all tools they have very clear limitations.
Thank you so much for spelling out what I've struggled to express to people. As someone who will forever hate GenAI, I have no problem with AI in tech or to help make my administrative time...
Thank you so much for spelling out what I've struggled to express to people. As someone who will forever hate GenAI, I have no problem with AI in tech or to help make my administrative time easier, but I can always tell when something is AI, and I genuinely hope that continues to be the case. For one, because it doesn't really feel like "intelligence" so much as it does a really good compilation of everything we've learned thus far. For another, I think people need to understand that there will always need to be human intervention, for the sake of community and society.
I have come to the mindset that it is possible that you can't always tell. However, that also means that someone has done all the work to double-check the output, rewrite/rework bits of it, etc....
but I can always tell when something is AI
I have come to the mindset that it is possible that you can't always tell. However, that also means that someone has done all the work to double-check the output, rewrite/rework bits of it, etc. Effectively done the work of human intervention. Which also brings them back to the realm of simply being tooling and not standalone miracle workers.
Sorry to use a gross analogy, but I think you’re falling for the selection bias that transphobes use when they say they can “always tell”. They can tell until they realize that cis people cover a...
Sorry to use a gross analogy, but I think you’re falling for the selection bias that transphobes use when they say they can “always tell”. They can tell until they realize that cis people cover a wide range and trans people can pass really well.
It's a shame because I don't actually hate AI itself, I hate product teams that keep insisting it needs to replace labor, instead of making existing talent more efficient. I feel like even 10...
It's a shame because I don't actually hate AI itself, I hate product teams that keep insisting it needs to replace labor, instead of making existing talent more efficient. I feel like even 10 years ago such a tech would have at least 30% of the industry using it as the next step in automation.
I think it's only partially boosters. It's also critics who talk about what they fear. Sometimes it seems like the critics are boosting the "AI will take our jobs" meme more than the boosters....
I think it's only partially boosters. It's also critics who talk about what they fear. Sometimes it seems like the critics are boosting the "AI will take our jobs" meme more than the boosters.
Hype is kind of like that - everyone makes a lot of noise, boosters, skeptics, and everyone else who wants to talk about it.
Some of it is fluff, yes. But I'm already seeing the effects here and there. Some friends here and there were laid off under this initaive "shift towards AI" (insert the prettified way to say how...
Some of it is fluff, yes. But I'm already seeing the effects here and there. Some friends here and there were laid off under this initaive "shift towards AI" (insert the prettified way to say how this benefits workers when it doesn't), I'm needing to talk more to bots for customer service, animation studios are cutting out people in some hope of making a mix of outsourcing and LLM tech to achieve cheaper movies. AIGen Slop has completely poisoned most asset stores for professional development (even if it looks good, it's a copyright nightmare to use it).
It may all revert in a few years, but this hype is definitely going to hurt a lot of people who otherwise were doing fine jobs.
Some of the AI voice impersonations are incredibly accurate. I've put myself through some blind tests and I have no idea. I have no doubt I could be completely fooled by AI voice in the correct...
I can always tell when something is AI
Some of the AI voice impersonations are incredibly accurate. I've put myself through some blind tests and I have no idea. I have no doubt I could be completely fooled by AI voice in the correct circumstance.
The public hype has been treating the current AI models as something of a first step that will only improve over time. And at the same rate! When in fact, what we have now is more like the peak of...
The public hype has been treating the current AI models as something of a first step that will only improve over time. And at the same rate! When in fact, what we have now is more like the peak of 20 years of research and development. Of course there will still be improvements to make, but not at the same rate. Like some of the fastests cars can go from 0-100 in 1.8s, but from 100-200 takes 2.5s on top of that and even longer to go to 300. But the hype have just been mindlessly extrapolating the jump from GPT 2 to GPT 3 and concluded it would continue at that level every year.
The hype isn't necessarily about an expectation for LLMs to get better; they're already very good at what they do, which is generating convincing looking text. The hype has more to do with the...
The hype isn't necessarily about an expectation for LLMs to get better; they're already very good at what they do, which is generating convincing looking text. The hype has more to do with the potential applications of that technology, which we've barely scratched the surface of at this point. Right now, we're still using the technology mostly in its most raw form, that is, passing text to it via a field and receiving its text output directly. There are way more applications for it to be integrated into applications and application pipelines, most of which will be pure gimmicky garbage, but likely much of which will be genuinely useful.
It sort of reminds me of the development of lasers. At first, a super powerful beam of coherent light was merely super cool. Shortly after, they started being used in their most raw and obvious applications; dumping tons of power into a small spot to cut or destroy things.
Now, the modern world wouldn't be remotely achievable without them. They enable high bandwidth long distance communication, optical disc storage, lithography, audio, microscopy, and about a million other things. The average person likely uses hundreds of lasers every day without even knowing it, and taking them for granted. It just took a while to figure those use cases out.
I agree, there will be advancement but it's unreasonable to expect it at the same rate just because of the hype. The companies are feeling the pressure to show the improvement in performance of...
I agree, there will be advancement but it's unreasonable to expect it at the same rate just because of the hype. The companies are feeling the pressure to show the improvement in performance of models because of their ridiculous valuations. What we have now with GPT-4 and comparable models is still a great piece of tech in itself. I hope the hype dies down and the trend of shoehorning AI into everything just to sell it more.
I don't think it's quite so clear. How would you discuss Moore's Law in relation to automotive design? Though easily grasped, automotive analogies might be inapt. If the public believes that AI...
I don't think it's quite so clear. How would you discuss Moore's Law in relation to automotive design? Though easily grasped, automotive analogies might be inapt. If the public believes that AI progress will resemble exponential growth, then the expectation of eye-popping advances every year is warranted. The hype might be wrong, but such a belief would not be unreasonable, given the track record of computing advancement over the past 60 years.
To me, the "growth" of AI is both unambiguous and highly subjective. Perhaps it's more like the adoption of the automobile than it is the mechanics of an automobile. When cars started to take over roads, we made the roads better for cars, then cars got better, and more people drove them, and the roads got better, and so on. At any point, you could say, "We couldn't possibly become more car-friendly!" But here we are, expanding lanes and spending more money on cars.
I wouldn't be surprised myself. A lot of hype around LLMs is with the idea that they are intelligent and make decisions that require intelligence. And with that I don't mean making associations...
I wouldn't be surprised myself. A lot of hype around LLMs is with the idea that they are intelligent and make decisions that require intelligence.
And with that I don't mean making associations and guess working on that(which I would personally argue is more or less what LLMs do), but actually making predictions based on a certain understanding of what the world is like.
And making something like that is going to take a whole lot more understanding. Given how memory might be more intertwined with the rest of the body than we previously thought, there's a good chance that copying or reproducing that part of 'intelligence'(yes I know I am vague here, unfortunately unavoidable with discussions around this topic) is more difficult than we thought.
I could be wrong of course. But I imagine that a lot of the infrastructure currently built for LLMs will be used for something else by the end of this decade.
I think they've reached limits for a particular approach where you throw more hardware at the problem, but it's still quite hard to tell what machine learning researchers will come up next, or how...
I think they've reached limits for a particular approach where you throw more hardware at the problem, but it's still quite hard to tell what machine learning researchers will come up next, or how long it will take to hit on something better.
Moore's law aside, technological innovation usually doesn't happen on a schedule.
I think we are discovering that consciousness is exactly as complicated as we thought it was (before LLMs maybe gave some people false hope). The thing is, I don't think they're off base on...
I think we are discovering that consciousness is exactly as complicated as we thought it was (before LLMs maybe gave some people false hope).
The thing is, I don't think they're off base on thinking this is a step toward GAI or a fully sapient machine.
Instead, an LLM is probably just one of the many components that will be required to create this artificial mind, much like our own brains have multiple different parts with distinct functions which only coalesce into consciousness when they are able to communicate and interact with one another.
Audio/visual sensors, access to a calculator, access to memory, some form of LLM, and probably a hundred other components we haven't figured out yet may very well create consciousness someday.
[OpenAI’s] model, known internally as Orion, did not hit the company’s desired performance, according to two people familiar with the matter, who spoke on condition of anonymity to discuss company matters. As of late summer, for example, Orion fell short when trying to answer coding questions that it hadn’t been trained on, the people said. Overall, Orion is so far not considered to be as big a step up from OpenAI’s existing models as GPT-4 was from GPT-3.5, the system that originally powered the company’s flagship chatbot, the people said.
…
At Alphabet Inc.’s Google, an upcoming iteration of its Gemini software is not living up to internal expectations, according to three people with knowledge of the matter. Anthropic, meanwhile, has seen the timetable slip for the release of its long-awaited Claude model called 3.5 Opus.
…
Similar to its competitors, Anthropic has been facing challenges behind the scenes to develop 3.5 Opus, according to two people familiar with the matter. After training it, Anthropic found 3.5 Opus performed better on evaluations than the older version but not by as much as it should, given the size of the model and how costly it was to build and run, one of the people said.
…
Tech companies are also beginning to wrestle with whether to keep offering their older AI models, perhaps with some additional improvements, or to shoulder the costs of supporting hugely expensive new versions that may not perform much better.
So advanced LLMs have been around for what, two years now? It feels premature to think we are hitting some fundamental limit with these models. And there have been recent, meaningful upgrades....
So advanced LLMs have been around for what, two years now? It feels premature to think we are hitting some fundamental limit with these models. And there have been recent, meaningful upgrades.
Llama 3 is far better than Llama 2. GPT-4.o is leaps and bounds above the performance of GPT-4. We probably won't see the level of improvement we saw in late 2022, but this technology is still progressing very quickly in the grand scheme of things.
I don't think the article implies any fundamental limit. It's a report that empirically, brute-force scaling seems to be getting harder for multiple companies, which is worth noticing.
I don't think the article implies any fundamental limit. It's a report that empirically, brute-force scaling seems to be getting harder for multiple companies, which is worth noticing.
Yeah, totally agree, however I'm more responding to how folks here and HN have interpreted these findings; namely as an indicator that we are approaching some fundamental technical limit with LLMs.
Yeah, totally agree, however I'm more responding to how folks here and HN have interpreted these findings; namely as an indicator that we are approaching some fundamental technical limit with LLMs.
If one were to look at a graph of A.I. progress on the Y axis and time on the X axis, they would see a graph filled with flat lines and massive jumps. Are we in the middle of a jump or at the...
If one were to look at a graph of A.I. progress on the Y axis and time on the X axis, they would see a graph filled with flat lines and massive jumps.
Are we in the middle of a jump or at the start of a flat line? Tough to say. However, if I may speculate...
First of all, it is important to recognize that many of rapid increases in A.I. progress have come not from improvements to the techniques, but rather from improvements to compute power. For example, it is well known that Least squares linear regression was performed by Legendre and Gauss in the 1800s but it wasn't until the computer was developed that Linear Regression became such a powerful technique. The work done by Gauss and Legendre would eventually lead to the creation of the neural network in the 1950s. Despite this revolutionary discovery, by 1974 we were at a flat line, an "AI Winter". Only 20ish years after this AI winter, the world would witness Deep Blue become the first computer chess-playing system to beat a reigning world chess champion. As stated on Wikipedia:
These successes were not due to some revolutionary new paradigm, but mostly on the tedious application of engineering skill and on the tremendous increase in the speed and capacity of computers by the 90s. In fact, Deep Blue's computer was 10 million times faster than the Ferranti Mark 1 that Christopher Strachey taught to play chess in 1951. This dramatic increase is measured by Moore's law, which predicts that the speed and memory capacity of computers doubles every two years. The fundamental problem of "raw computer power" was slowly being overcome.
Even the current models have been enabled by the power of the modern GPU. Of course nowadays, some speculate that Moore's law is slowing down and if this is true, it will mean that A.I. improvements cannot come from compute power increases alone.
Additionally, it seems obvious to me that the biggest problem with modern LLMs is simply that the creators have poisoned the well. Now that AI content is everywhere, it becomes much more complicated to find good quality data to train on. The current approach seems to have been "just throw more data at it". An approach like this is guaranteed to have massive returns at first and diminishing returns later, especially so as the quality of the data decreases.
In my opinion, improvements in the near future are unlikely to come from model improvements. Improvements to the models require a lot of research and effort. Short term improvements are most likely to come in the form of techniques that utilize the current models more effectively such as agents or prompting techniques like tree of thought reasoning.
That being said, a lot of these articles remind me of Flying Machines Which Do Not Fly. Many people are rooting against A.I. and would like to see it fail. Therefore, people are willing to latch onto any ounce of information that paints the future of A.I. as bleak. In my opinion, the current models are more than enough to cause issue in many parts of society. There needs to be less of a focus on the future and more of a focus on the now. These models already exist and they are actively changing the world.
While the derivative of the graph may change over time, the long term trend is monotonic. If the past is any indication of the future, A.I. will continue to improve, it just may take time. Or maybe it will rapidly accelerate again.
Regardless, run with it or run from it, my friends.
Ah so the sky isn't the limit? Or maybe they have sucked up all the atmosphere (training data) so the sky is a hard limit?
This could be a temporary bump in the road. But, given how iterative most LLM development seems to have been I personally think we might have reached a plateau with what can be achieved with these types of models. At the very least it is becoming clear they are not improving at the rate the hype seems to suggest. Although that shouldn't come as a surprise for anyone who has paid any attention to the actual improvements between gpt-4 versions.
Edit:
Another thought I had. If all the training data in the world is not enough to create the next step forward, then maybe these models are indeed not the way forward for anything remotely resembling actual intelligence and more importantly ability to improvise. This should also not be a surprise for most people who have paid attention.
To be clear, because I know people will misread what I wrote, that doesn't make these models useless. I quite enjoy my fancy rubber ducky who can actually talk back and pitching two models against each other. Like the juniors our team needs but never gets,while also not having to be gentle with their emotions when the code sucks ;). But, at the end of the day these models are just tools and like all tools they have very clear limitations.
Thank you so much for spelling out what I've struggled to express to people. As someone who will forever hate GenAI, I have no problem with AI in tech or to help make my administrative time easier, but I can always tell when something is AI, and I genuinely hope that continues to be the case. For one, because it doesn't really feel like "intelligence" so much as it does a really good compilation of everything we've learned thus far. For another, I think people need to understand that there will always need to be human intervention, for the sake of community and society.
I have come to the mindset that it is possible that you can't always tell. However, that also means that someone has done all the work to double-check the output, rewrite/rework bits of it, etc. Effectively done the work of human intervention. Which also brings them back to the realm of simply being tooling and not standalone miracle workers.
Sorry to use a gross analogy, but I think you’re falling for the selection bias that transphobes use when they say they can “always tell”. They can tell until they realize that cis people cover a wide range and trans people can pass really well.
It's a shame because I don't actually hate AI itself, I hate product teams that keep insisting it needs to replace labor, instead of making existing talent more efficient. I feel like even 10 years ago such a tech would have at least 30% of the industry using it as the next step in automation.
How things shift so quickly...
I think it's only partially boosters. It's also critics who talk about what they fear. Sometimes it seems like the critics are boosting the "AI will take our jobs" meme more than the boosters.
Hype is kind of like that - everyone makes a lot of noise, boosters, skeptics, and everyone else who wants to talk about it.
Some of it is fluff, yes. But I'm already seeing the effects here and there. Some friends here and there were laid off under this initaive "shift towards AI" (insert the prettified way to say how this benefits workers when it doesn't), I'm needing to talk more to bots for customer service, animation studios are cutting out people in some hope of making a mix of outsourcing and LLM tech to achieve cheaper movies. AIGen Slop has completely poisoned most asset stores for professional development (even if it looks good, it's a copyright nightmare to use it).
It may all revert in a few years, but this hype is definitely going to hurt a lot of people who otherwise were doing fine jobs.
Some of the AI voice impersonations are incredibly accurate. I've put myself through some blind tests and I have no idea. I have no doubt I could be completely fooled by AI voice in the correct circumstance.
The public hype has been treating the current AI models as something of a first step that will only improve over time. And at the same rate! When in fact, what we have now is more like the peak of 20 years of research and development. Of course there will still be improvements to make, but not at the same rate. Like some of the fastests cars can go from 0-100 in 1.8s, but from 100-200 takes 2.5s on top of that and even longer to go to 300. But the hype have just been mindlessly extrapolating the jump from GPT 2 to GPT 3 and concluded it would continue at that level every year.
The hype isn't necessarily about an expectation for LLMs to get better; they're already very good at what they do, which is generating convincing looking text. The hype has more to do with the potential applications of that technology, which we've barely scratched the surface of at this point. Right now, we're still using the technology mostly in its most raw form, that is, passing text to it via a field and receiving its text output directly. There are way more applications for it to be integrated into applications and application pipelines, most of which will be pure gimmicky garbage, but likely much of which will be genuinely useful.
It sort of reminds me of the development of lasers. At first, a super powerful beam of coherent light was merely super cool. Shortly after, they started being used in their most raw and obvious applications; dumping tons of power into a small spot to cut or destroy things.
Now, the modern world wouldn't be remotely achievable without them. They enable high bandwidth long distance communication, optical disc storage, lithography, audio, microscopy, and about a million other things. The average person likely uses hundreds of lasers every day without even knowing it, and taking them for granted. It just took a while to figure those use cases out.
I agree, there will be advancement but it's unreasonable to expect it at the same rate just because of the hype. The companies are feeling the pressure to show the improvement in performance of models because of their ridiculous valuations. What we have now with GPT-4 and comparable models is still a great piece of tech in itself. I hope the hype dies down and the trend of shoehorning AI into everything just to sell it more.
I don't think it's quite so clear. How would you discuss Moore's Law in relation to automotive design? Though easily grasped, automotive analogies might be inapt. If the public believes that AI progress will resemble exponential growth, then the expectation of eye-popping advances every year is warranted. The hype might be wrong, but such a belief would not be unreasonable, given the track record of computing advancement over the past 60 years.
To me, the "growth" of AI is both unambiguous and highly subjective. Perhaps it's more like the adoption of the automobile than it is the mechanics of an automobile. When cars started to take over roads, we made the roads better for cars, then cars got better, and more people drove them, and the roads got better, and so on. At any point, you could say, "We couldn't possibly become more car-friendly!" But here we are, expanding lanes and spending more money on cars.
I wouldn't be surprised myself. A lot of hype around LLMs is with the idea that they are intelligent and make decisions that require intelligence.
And with that I don't mean making associations and guess working on that(which I would personally argue is more or less what LLMs do), but actually making predictions based on a certain understanding of what the world is like.
And making something like that is going to take a whole lot more understanding. Given how memory might be more intertwined with the rest of the body than we previously thought, there's a good chance that copying or reproducing that part of 'intelligence'(yes I know I am vague here, unfortunately unavoidable with discussions around this topic) is more difficult than we thought.
I could be wrong of course. But I imagine that a lot of the infrastructure currently built for LLMs will be used for something else by the end of this decade.
I think they've reached limits for a particular approach where you throw more hardware at the problem, but it's still quite hard to tell what machine learning researchers will come up next, or how long it will take to hit on something better.
Moore's law aside, technological innovation usually doesn't happen on a schedule.
I think we are discovering that consciousness is exactly as complicated as we thought it was (before LLMs maybe gave some people false hope).
The thing is, I don't think they're off base on thinking this is a step toward GAI or a fully sapient machine.
Instead, an LLM is probably just one of the many components that will be required to create this artificial mind, much like our own brains have multiple different parts with distinct functions which only coalesce into consciousness when they are able to communicate and interact with one another.
Audio/visual sensors, access to a calculator, access to memory, some form of LLM, and probably a hundred other components we haven't figured out yet may very well create consciousness someday.
From the article (archive):
…
…
…
So advanced LLMs have been around for what, two years now? It feels premature to think we are hitting some fundamental limit with these models. And there have been recent, meaningful upgrades.
Llama 3 is far better than Llama 2. GPT-4.o is leaps and bounds above the performance of GPT-4. We probably won't see the level of improvement we saw in late 2022, but this technology is still progressing very quickly in the grand scheme of things.
I don't think the article implies any fundamental limit. It's a report that empirically, brute-force scaling seems to be getting harder for multiple companies, which is worth noticing.
Yeah, totally agree, however I'm more responding to how folks here and HN have interpreted these findings; namely as an indicator that we are approaching some fundamental technical limit with LLMs.
If one were to look at a graph of A.I. progress on the Y axis and time on the X axis, they would see a graph filled with flat lines and massive jumps.
Are we in the middle of a jump or at the start of a flat line? Tough to say. However, if I may speculate...
First of all, it is important to recognize that many of rapid increases in A.I. progress have come not from improvements to the techniques, but rather from improvements to compute power. For example, it is well known that Least squares linear regression was performed by Legendre and Gauss in the 1800s but it wasn't until the computer was developed that Linear Regression became such a powerful technique. The work done by Gauss and Legendre would eventually lead to the creation of the neural network in the 1950s. Despite this revolutionary discovery, by 1974 we were at a flat line, an "AI Winter". Only 20ish years after this AI winter, the world would witness Deep Blue become the first computer chess-playing system to beat a reigning world chess champion. As stated on Wikipedia:
Even the current models have been enabled by the power of the modern GPU. Of course nowadays, some speculate that Moore's law is slowing down and if this is true, it will mean that A.I. improvements cannot come from compute power increases alone.
Additionally, it seems obvious to me that the biggest problem with modern LLMs is simply that the creators have poisoned the well. Now that AI content is everywhere, it becomes much more complicated to find good quality data to train on. The current approach seems to have been "just throw more data at it". An approach like this is guaranteed to have massive returns at first and diminishing returns later, especially so as the quality of the data decreases.
In my opinion, improvements in the near future are unlikely to come from model improvements. Improvements to the models require a lot of research and effort. Short term improvements are most likely to come in the form of techniques that utilize the current models more effectively such as agents or prompting techniques like tree of thought reasoning.
That being said, a lot of these articles remind me of Flying Machines Which Do Not Fly. Many people are rooting against A.I. and would like to see it fail. Therefore, people are willing to latch onto any ounce of information that paints the future of A.I. as bleak. In my opinion, the current models are more than enough to cause issue in many parts of society. There needs to be less of a focus on the future and more of a focus on the now. These models already exist and they are actively changing the world.
While the derivative of the graph may change over time, the long term trend is monotonic. If the past is any indication of the future, A.I. will continue to improve, it just may take time. Or maybe it will rapidly accelerate again.
Regardless, run with it or run from it, my friends.