Hacker Newsnew | past | comments | ask | show | jobs | submit | nofum's commentslogin

Given the increase in 2020 alone was 3x the increase over the preceding 5 years, it appears to be a valid theory as to the source.


Obvious lockdown residue. Could also be COVID itself but far fewer people were directly affected by that.


Could also be the increased income disparity, the huge leap in housing costs without associated pay increase, the obvious inability of the US to change the direction it (and the world) is heading...

I mean, it's rising most amongst young adults and minorities; given the political climate alone I can understand the correlation there.


Exactly, lockdowns caused a huge spike in precisely all those things. Income disparity skyrockets as the laptop class gets raises and stock bonuses while the servant class delivers their food while inflation skyrockets. Housing costs explode as the fixed supply in suburbia is swamped with urban exiles working remote and lockdowns cause supply chain disruptions.


Why is a constant stream of ultra high resolution input from ~5 senses for years not count as a huge volume of data? Not to mention all the pre-training data encoded in the brains blueprint via DNA through natural selection? If anything, LLMs are more impressive than humans and animals in that they only have access to text to build a world model from a tabula rasa.


https://en.m.wikipedia.org/wiki/Kolmogorov_complexity

The raw volume of the data from the human senses is high but a lot of it is low entropy. The high resolution video from our retinas of reading the pages of a book over many hours, and the ASCII representation of said text are equivalent, modulo information about the physical object of the book.


People have pointed out that LLMs like GPT 3 and 4 can draw pictures, but they've been "blind since birth" and have never seen anything. They've just read descriptions of things. How capable would a human be, if they had grown up deaf, blind, and in all other ways insensate except for some sort of braille-like reading input?

Another salient quote is from Andrej Karpathy (ex-Tesla AI team lead) who said that a camera is a high-bandwidth input "that puts many constraints on the world". Children learn from multi-modal inputs, and vision especially provides a large number of constraints that they can use to learn how the world works.

I have a two-year old, and something I've noticed is that infants have a very strong instinctive urge to gain agency over the world. They try very hard to control things, to make things move, to make sounds, to be able to affect things in all sorts of ways. This must be a very critical part of learning, because they'll kick and scream if you remove their agency. It's as strong an instinct as wanting to eat or sleep. No current LLM has any kind of feedback loop, and has essentially zero agency.


We have access to video training data for driving, absolutely tons of it, as we've been attempting to train AI cars for more than two decades now. If GPT4 is what you say it is, we should be able to train it on that video data and solve autonomous driving. There is nothing inherently about transformers that prevents them from taking in video data. They've already been used by some researchers (https://arxiv.org/abs/2104.09224).

And yet, you can take a 16 year old who's never driven, and teach them within a week to be decent at it and maybe 50-100 hours of driving training and they're competent. You don't need to show them a billion man-hours of driving footage. Even the first people to buy cars in the late 1800s when they were first invented, were able to pick it up almost right away (there weren't even driving licenses back then).

At any rate, driving is just one example. Despite being one of the oldest futuristic sci-fi examples, I don't see restaurants powered by AI. I don't see housekeeping powered by it.

Ok, those are embodied examples, so you say they're unfair. Fine. What remote-friendly jobs are being swept away by AI? Can we even do customer service with AI right now? No, outside of some "front line" chat bot (which just replaces phone trees and terrible localized search engines), we can't. Even if a GPT is trained on a business's proprietary documentation, it's wrong or unresponsive enough that it would cost you more than it would save you by firing your customer support staff.


> a 16 year old

A 16 year old has a decade of vision input, knows tens of thousands of words, has a coherent theory-of-mind, and has self-preservation instincts honed over hundreds of millions of years of evolution.

The equivalent scenario would be take a fully general AGI(!) that already has a body and has learned to manipulate the physical world, put that in the car and have it learn to drive.

A lot of people seem confused about these scenarios. The currently popular LLMs are like a child raised in a black box, and are weirdly retarded in the same way you would expect a child raised in a black box to be.

Similarly, driving AIs don't speak English, can't take instructions, and are simultaneously learning physics, theory-of-mind, the rules of the rode, and signage conventions without having agency during most of their training.


> A 16 year old has a decade of vision input, knows tens of thousands of words, has a coherent theory-of-mind, and has self-preservation instincts honed over hundreds of millions of years of evolution.

No they don't. Where were you driving 23 days ago? Or, if you could categorize your driving data in your head, what did you pass, precisely, in the car while you were driving 96 hours ago? AI training data has all of this. It has perfect timestamps with 25-30 frames per second (or more in some cases, with multiple cameras feeding frames in to the dataset) with full 360 view in many cases, adding up to millions of man-hours of driving data from thousands of drivers, which is longer than any human lifespan. A lot of times this data is supplemented with IR representations or LIDAR representations of the environment as well, something a human can't even see.

> Similarly, driving AIs don't speak English, can't take instructions, and are simultaneously learning physics, theory-of-mind, the rules of the rode, and signage conventions without having agency during most of their training.

They're not learning that on their own. They're being poked and prodded by humans in the loop (and indeed, automated tagging tools) who have tweaked the models by adding numerical weights to "bad outcomes" and "good outcomes". Or "good categorizations" and "bad categorizations", or "aligned responses" and "unaligned responses". And because it's just a dumb statistical model, if you tell it the color blue is bad and that it should resist answering any questions related to the color blue, it will agree, because its entire heuristic model for organizing the world was designed by humans.

Similarly, it doesn't have a theory of mind. If it did, it would also have agency and it would already be AGI. Instead, it has access to examples of data of humans exhibiting theory of mind with other humans. And it's got enough parameters in its statistical model that it can accommodate a weighting for this "theory of mind" impact on what the expected output should be.

LLMs are nothing like an intelligent child raised in a black box. I mean, we already have examples of this, vis-a-vis being both blind and deaf: https://en.wikipedia.org/wiki/Helen_Keller.


> No they don't. Where were you driving 23 days ago? Or, if you could categorize your driving data in your head, what did you pass, precisely.

Just because I cannot do that, still my brain had this input, and was "trained", and we know the real neurons of the brain work completely different in all regards (connecitvities forming, activations, weighting) than our gross simplification of one activation function with a weight.. so not getting your arguments at all.

I have been trained on all these inputs, and even if I cannot recall them now they manifested in a my superior brain somehow.

Nice that in theory the AI training data has all this available, still the resulting model is much simpler, and also cannot recite all of its inputs seen, too??


I mean everyday vision input, which generalises to novel scenarios because driving happens in the same world with the same rules of physics and optics. Shadow and light, depth and movement perception, etc… can all be reused.

This is like how LLMs are difficult to train to the point that they understand English, but then relatively easy to specialise.

Similarly, students don’t start their education at University, they spend nearly two decades getting to that point by learning the prerequisites.

> Similarly, it doesn't have a theory of mind. If it did, it would also have agency

These are unrelated concepts. There is evidence that GPT 4 has a rudimentary model of mind but it isn’t conscious itself and is a static model with zero agency over the world.


That might be a feature, not a bug, depending on what you mean by "agency". At the tippy top of that concept might be an AI system that decides by itself (as perhaps a weapons system) whether to kill human beings without direct instruction to do so. Yeah, Skynet. And every other movie where the AI "goes crazy".


On the contrary, an AI without agency could be instructed by a human to execute tasks that are evil and it will never push back either.


While the input resolution is indeed high, it's not as if any animal on earth has access to photographic memory of all that data. It is so ephemeral you could scarcely tell me what the last 10 people you passed on the street were wearing, even if it was 10 minutes ago.


At inference time, an LLM does not have access to all its training data either.


At inference time, a human being has access to 5 to 9 variables at most an average human can hold in their head at any given time. But I can feed pages and pages to an LLM and it has a perfect representation of every single word and letter in its working memory. It also has access to a look up table of billions of statistical correlations for what the next word or block of words in any sentence should be given the perfect priors.

If I'm talking to you, especially in person, I won't be able to tell you what the second word of three sentences ago was. In fact, I might not have even been paying attention, because I'm already thinking at a higher level about what you're communicating to me and the word wasn't all that important. But a LLM certainly uses this information to guide its responses (along with its ginormous lookup table).


> LLMs are more impressive than humans and animals in that they only have access to text to build a world model from a tabula rasa.

They don’t build a world model at all. They make inferences from text.

Considering they have been trained on 100,000 books, and all of Wikipedia, they are remarkably unintelligent, and really only able to produce text that is consistent with what they have been trained on.


I mean, Helen Keller did pretty well (with an appropriate tutor) with only very low bandwidth interfaces like touch and smell.


Yikes, they have a submission leaderboard. As if the expected signal to noise ratio of a system like this wasn't low enough.


I work on this project. Every submission has to be approved by an editor before it appears on the site or gets counted toward leaderboard stats. We're currently more constrained by lack of submissions than by low-quality submissions – it's not uncommon that someone will contact us asking about an incident that should be in the database but isn't yet.


It's possible that long-term you are not constrained by a lack of submissions but instead by a lack of actual AI incidents, how would you tell the difference? Encouraging people to go on a hunt for borderline AI-related "incidents" to submit seems like a bad idea.


Does that include the examples other posters have linked in this thread?


The earlier incidents like the Pokémon Go one were sourced from several existing collections: https://incidentdatabase.ai/research/2-roadmap/#:~:text=init.... That was before I or our most active current editor were on the project. The ChatGPT ones are more recent and did go through review.


Next, they could shoot themselves in the other foot by offering prizes based on leaderboard placement. /s


The prompt likely is tuned to avoid making statements not supported by the PDF content.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: