It stands to reason that if you have access to an LLM’s training data, you can influence what’s coming out the other end of the inscrutable AI’s network. The obvious guess is that…
Isn’t “intelligence” so ill defined we can’t prove it either way. All we have is models doing better on benchmarks and everyone shrieking “look emergent intelligence”.
I disagree a bit on “toys”. Machine summarization and translation is really quite powerful, but yeah that’s a ways short of the claims that are being made.
The problem is the harvesting.
In previous incarnations of this process they used curated data because of hardware limitations.
Now that hardware has improved they found if they throw enough random data into it, these complex patterns emerge.
The complexity also has a lot of people believing it’s some form of emergent intelligence.
Research shows there is no emergent intelligence or they are incredibly brittle such as this one. Not to mention they end up spouting nonsense.
These things will remain toys until they get back to purposeful data inputs. But curation is expensive, harvesting is cheap.
Isn’t “intelligence” so ill defined we can’t prove it either way. All we have is models doing better on benchmarks and everyone shrieking “look emergent intelligence”.
I disagree a bit on “toys”. Machine summarization and translation is really quite powerful, but yeah that’s a ways short of the claims that are being made.