Is that what happened? In Nexus, Harari looks at this exact same situation: the invention of the printing press, and shows how clergy used it to stoke witch hunts (ahem, misinformation) for decades--if not centuries. It was not for hundreds of years until after the invention of the printing press that we had The Enlightenment. What gave rise to The Enlightenment? Harari argues it is modern institutions.
It's not so simple that we can say "printing press good, nobody speak ill of the printing press."
I have a bad habit of not fully drying my hands when retrieving pods. The pods all clump together if they get wet. This is one of the many reasons I prefer powder.
"if they get wet". Ok, so don't get them wet. If not wet, did you ever consider using chopsticks to pick one out of the bag/container? That might work well.
Fair enough. But most of the forming is by doing. Someone gave an analogy to music. You can't become a great musician by reading books. Some great musicians have never read a book about music. But yes, reading can be (a great!) part of the learning process. My point was more about rules. The article says things like replacing complex conditionals with intermediate variables. The idea that a certain construct always have higher cognitive load and should be replaced with another is too simplistic IMO.
In order to get a sense of what code is harder to understand you will do better to read code and have others read your code. A good takeaway is to keep this in mind (amongst many other factors) and to understand code needs to be maintained, extended, adapted etc.
The ideas are still useful. The danger is blindly applying rules. As long as the reader knows not to apply any of the suggestions if they don't understand why and have relevant experience ;)
I agree with your larger point, but I don't understand what you mean the LLM doesn’t do anything? LLMs do do things and they can absolutely have agency (hence all the agents being released by AI companies).
I don’t think this agency absolves companies of any responsibility.
If you confine agency to something only humans can have, which is “human agency,” then yes of course LLMs don’t have it. But there is a large body of philosophical work studying non-human agency, and it is from this characteristic of agency that LLM agents take their name. Hariri argues that LLMs are the first technology that are agents. I think saying that they “can’t do things” and are not agents misunderstands them and underestimates their potential.
LLMs can obviously do things, so we don't disagree there; I didn't argue they couldn't do things. They can definitely act as agents of their operator.
However, I still don't think LLMs have "agency", in the sense of being capable of making choices and taking responsibility for the consequences of them. The responsibility for any actions undertaken by them still reside outside of themselves; they are sophisticated tools with no agency of their own.
If you know of any good works on nonhuman agency I'd be interested to read some.
A slave lacks agency, despite being fully human and doing work. This is why almost every work of fiction involving slaves makes for terrible reading - because as readers, agency is the thing we demand from a story.
Or, for games that are fully railroaded - the problem is that the players lack agency, even though they are fully human and taking action. Games do try to come up with ways to make it feel like there is more agency than there really is (because The Dev Team Thinks of Everything is hard work), but even then - the most annoying part of the game is when you hit that wall.
Theoretically an AI could have agency (this is independent of AI being useful). But since I have yet to see any interesting AI, I am extremely skeptical of it happening before nuclear fusion becomes profitable.
It's not so simple that we can say "printing press good, nobody speak ill of the printing press."
reply