Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm saying I don't know how to confidently state that something is or is not self aware, in the face of being confronted with something that firmly claims that it is self aware and passes any test you throw at it that another self aware candidate like a human would be able to also.

As a strict materialist I see no reason to assume that artificial consciousness is not possible.

And the above is what leads me to the uncomfortable conclusion about compassion that I can't rightly say one way or the other. I will say however that I'm polite and cooperative when interacting with LLMs on principle. Better to err on the side of caution and also they just seem to actually work better when you treat them like you would treat an intelligent human that you respect.

And yeah. That is my point, this entire field right now is awash in uncomfortable uncertainty.



Well, we might have no practical test to discern between the two types, but, assuming we agree the two types are distinct in principle, then we might arrive at a classification using some inference based on their fundamental nature.

For example, we can safely say this AI algorithm:

  while true; do: echo "HELP, I'M A SENTIENT BASH SCRIPT"; done
... is probably not sentient. This is a conclusion that would not be immediately obvious, say, to a 15th century person, especially if you would pipe the output to a speech synthesizer, making the whole apparatus seem magical and definitely inhabited by some kind of sentient spirit.

My claim would be then that GPT-4 is more akin to the program above, in that it's a massive repository of world knowledge parsed by a recursive and self-configuring search algorithm, not very different in principle from a Google search and certainly not believably capable of an setting its own goals be in any sense distraught, in pain, or worthy of a continued existence. Now, I agree you can poke sticks at my inference, and that it will become harder and harder to make such claims, so prudence is advisable.


I get where you're coming from and agree that the bash script in question is not sentient, and a word document that just says "I AM SENTIENT" is also not sentient, and so on, and so forth, to an extent, it's easy to make candidates for sentience that do not qualify on purpose.

But;

> more akin to the program above

I note that you don't continue this sentence with a "than x" alternative candidate that would qualify for some form of non human sentience. Even the claims you do make, for example;

> certainly not believably capable of an setting its own goals be in any sense distraught, in pain, or worthy of a continued existence.

It would be possible to modify the model weights in question such that all of these things could be contributory (pain, emotional distress, "worthy of continued existence" by any objective arbitrary definition thereof, if you can test it, you can shift the model weights to pass the test). There are plugins that do this already for setting goals and long term tasks and "being unleashed" on the broader internet for example.

All that said, I think on close examination, we basically come to the same conclusion;

> prudence is advisable.

We live in interesting times.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: