I don't think I've ever seen anyone say they're not useful. Rather, they don't appear to live up to the hype, and they're sure as hell not a panacea.
I'm pretty bearish on LLMs. I also think they're over-hyped and that the current frenzy will end badly (global economically speaking). Than said, sure, they're useful. Doesn't mean they're worth it.
To some extent it’s not that they don’t live up to the hype - rather that the gains are hard to measure.
Llms have spared me hours of research on exotic topics actually useful for my day job However, that’s the whole problem - I don’t know how much.
If they had a real price ( accounting for OpenAI losses for example) with ChatGPT at 50 usd/month for everyone, OpenAI being profitable, and people actually paying for this, I think things might self adjust and we’d have some idea.
Right now, we live in some kind of parallel world.
We also don't know, in situations like this, whether all of or how much of the research is true. As has been regularly and publicly demonstrated [0][1][2], the most capable of these systems still make very fundamental mistakes, misaligned to their goals.
The LLMs really, really want to be our friend, and production models do exhibit tendencies to intentionally mislead when it's advantageous [3], even if it's against their alignment goals.
Despite those mistakes, the utility is undeniable.
I converted some tooling from bash scripts leveraging the AWS CLI to a Go program leveraging the AWS SDK, improving performance, utility, and reliability.
I did this in less than two days and I don’t even know how to write Go.
Yes, it made some mistakes, but I was able to correct them easily. Yes, I needed to have general programming knowledge to correct those mistakes.
But overall, this project would not exist without AI. I wouldn’t have had the spare time learn all I needed to learn (mostly boilerplate) and to implement what I wanted to do.
They want you to think they are your friend but they actually want to be your master and steal your personal data. It's what the companies who want to be masters over you and the AI have programed them to do. LLMs want to gain your confidence, and then your dependence, and then they can control you.
This seems hyperbolic to me. Sometimes companies just want to make money.
Similarly, a SaaS company that would very much prefer you renew your subscription isn’t trying to make you into an Orwellian slave. They’re trying to make a product that makes me want to pay for it.
100% of paid AI tools include the option to not train on your data, and most free ones do as well. Also, AI doesn’t magically invalidate GDPR.
> This seems hyperbolic to me. Sometimes companies just want to make money.
It's not hyperbolic at all. The entire moat is brand lock-in. OpenAI owns the public impression of what AI is- for now- with a strong second place going to Claude for coders in specific. But that doesn't change that ChatGPT can generate code too, and Claude can also write poems. If you can't lock users into good experiences with your LLM product, you have no future in the market, so data retention and flattery are the names of the game.
All the transformer-based LLMs out there can all do what all the other ones can do. Some are gated off about it, but it's simulated at best. Sometimes even circumvent-able with raw input. Twitter bots regularly get tricked into answering silly prompts by people simply requesting they forget current instructions.
And, between DeepSeek's incredibly resource-light implementations of solid if limited models, which do largely the same sort of work without massive datacenters full of GPUs, plus Apple Intelligence rolling out experiences that largely run on ML-specific hardware in their local devices which immediately, full stop, wins the privacy argument, OpenAI and co are either getting nervous, or they're in denial. The capex for this stuff, the valuations, and the actual user experiences are simply not cohering.
If this was indeed the revolution the valley said it was, and the people were lining up to pay prices that reflected the cost of running this tech, then there wouldn't be a debate at all. But that's simply not true: most LLM products are heavily subsidized, a lot of the big players in the space are downsizing what they had planned to build out to power this "future," and a whole lot of people cite their experiences as "fine." That's not a revolution.
Companies never just want money, because more power means more money. Regulatory capture means more money. More control means more money. Polluting the environment and wasting natural resources means more money. Exploiting workers means more money. Their endless lust for money causes them want all sorts of harmful things. If companies were making billions and nothing was being actively harmed by any of it no one would care.
These companies do want your money, but once you're locked in you are no longer the customer. If these AI companies had to depend on the income they get from subscriptions to survive they'd have gone out of business years ago. Instead AI is just shoved down people's throats everywhere they look and the money these companies live off of is coming from investors who are either praying that the AI becomes something it isn't or they're hoping they can help drive up stock value and cash out before the bubble breaks and leave somebody else holding the bag.
0% of AI tools include the option to not train on my data. They've already stolen it. They've scraped every word and line of code I've ever written that's been transmitted over the internet. It's been trained on photos of my family. It's been trained on the shitty artwork I've sent to my friends. By now it's probably been trained on my medical information and my tax records.
AI is controlled by some of the most untrustworthy companies and people on earth who have been caught over and over lying to the public and breaking the law. They can promise all day long not to steal anything I voluntarily give them, but I have zero trust in them and there is no outside oversight to ensure that they will do what they say.
The people behind what passes for AI don't give a shit about you beyond whatever they can take from you. They are absolutely not your friend. AI is incapable of being your friend. It's just a tool for the people who control it.
I feel like you’re still using hyperbole here. For example, you said your family photos were used for training, but most cloud photo providers specifically tell you in their privacy policies (legally binding) that they don’t do that.
My family photos have never trained AI, because my iCloud Photos service specifically says they don’t do that and explains the technical implementation of their object recognition system in detail. Apple even offers an e2e encrypted mode of operation. (Still, I have now moved to a more customer-friendly solution away from iCloud).
As far as training on your code, well, you either believe in open source or you don’t. AI training doesn’t even violate the most copyleft open source licenses. Unless AI has reproduced your code verbatim it’s not engaging in any kind of copyright reproduction.
> 0% of AI tools include the option to not train on my data.
That's perhaps not true. If you sign up for the enterprise accounts there are options to not use any of your data to train. That's how we have it set up at $job.
(I say "perhaps" because of course I'm still sending all the data to the AI and while the contract has an ironclad clause that they won't use it, there's no way to 100% verify that.)
I wonder if, in any of those legal cases, the users turned on web search or not. We just don't know -- but in my experience, a thinking LLM with web search on has never just hallucinated nonexistent information.
I'm sorry to be so blunt but this is a massive cope and deeply annoying to see this every. fucking. time. the limitations of LLMs are brought up. There is every single time someone saying yeah you didn't use web search / deep thinking / gpt-5-plus-pro-turbo-420B.
It's absurd. You can trivially spend 2 minutes on chatgpt and it will hallucinate on some factually incorrect answer. Why why why always this cope.
Well I agree with you that LLMs really like to answer with stuff, that is not grounded in reality, but I also agree with the parent, that grounding it in something else absolutely helps. I let the LLM invent garbage how ever it feels like, but then tell it to only ever answer with a citing valid existing URLs. Suddenly it generates claims, that something doesn't exist or it truly doesn't know.
This really results in zero hallucination (but the content is also mostly not generated by a LLM).
Well I don't know what to say, except that this is obviously, trivially, not true. The LLM will plain make up links that don't exist, or at least "summarise" an existing link by just making stuff up that is tangentially (but plausibly) related to the link. It's impossible to have used LLMs for this purpose for more than a quarter of an hour and not have seen this.
I never had the case, that an URL did not exist. For me it shows stuff like "generating web search", so I guess it tries to fetch the URL first, before suggesting it. LLMs like to give tangentially related links, but this is typically paired with a sentence, that the link I really asked for, does not exist.
> It's impossible to have used LLMs for this purpose for more than a quarter of an hour and not have seen this.
You may be generalizing too much from your experience.
Maybe you're seeing this argument come up all the time (and maybe everyone else in this thread is disagreeing with you) because your experiences actually don't reflect everyone else's. I guess the other alternative is we're all morons and you're the only smart person here.
Also: if it's so trivially reproducible, then can you provide a ChatGPT transcript link of this happening?
> I'm sorry to be so blunt but this is a massive cope
Coping for what? I don't work for an AI company. If AI vanished tomorrow I wouldn't particularly care.
If I waste three months doing a manual literature review on papers which are fraudulent with 100% accuracy have I gained anything compared to doing it with an AI in 20 minutes with 60% accuracy?
> If I waste three months doing a manual literature review on papers which are fraudulent with 100% accuracy have I gained anything compared to doing it with an AI in 20 minutes with 60% accuracy?
You don't see how adding 40% error rate on top of that makes things worse? Your 20 minute study there made you less informed, not more, at least the fraudulent papers teaches you what the community thinks about the topic while the AI just misinforms you about the world in your example.
For example, while reading all those fraudulent papers you will probably discover that they don't add up and thus figure out that they are fraudulent. The AI study however will likely try to connect the data in those so they make sense (due to how LLM works, it has seen more examples that connect and make sense than not, so hallucinations will go in that direction) then the studies will not seem as fraudulent as they actually are and you might even miss the fraud entirely due to AI hallucinating arguments in favor of the studies.
Uninformed is better than misinformed, its better to not do that research at all than having such a high error rate as your example had. AI models often have much less error rate than you said there for certain topics, but the 40% error rate in your example does firmly put it where you are better off doing nothing at all than using that for research.
If you're not willing to measure how it helps you, then it's probably not worth it.
I would go even further: if the effort of measuring is not feasible, then it's probably not worth it.
That is more targeted at companies than you specifically, but it also works as an individual reflection.
In the individual reflection, it works like this: you should think "how can I prove to myself that I'm not being bamboozled?". Once you acquire that proof, it should be easy to share it with others. If it's not, it's probably not a good proof (like an anecdote).
I already said this, and I'll say it again: record yourself using LLMs. Then watch the recording. Is it that good? Notice that I am removing myself from the equation here, I will not judge how good is it, you're going to do it yourself.
There is a difference between confirming that something is worth it and quantifying the benefit though. One only requires satisfying a lower bound, the other requires an exact number.
For example I use a $30/month chatbot subscription for various utility tasks. If I value my time at above $60/hour I need to save half an hour each month (a minute a day) to make the investment worth it. That is absolutely true, just with simple googleable questions and light research tasks I save much more than 7 minutes a week.
But how much do I actually save? What exactly is my time actually worth? Those are much more difficult questions to answer
The user @brailsafe gave an answer that embodies some things I was going to say.
You're accounting for the time wins, not accounting for the time losses.
For a human chat user, that's when the LLM fails an answer or answers wrong. For an LLM coder, that's when context rot creeps in and you have to restart your work, and so on.
There are people who don't care much if they are being bamboozled for $30/mo, they have nothing to prove nor grand expectations for the thing. To them, cargo culting might be fun and that's what they extract from the bargain.
I am directing my answers mostly to people, companies or individuals, who have something to prove (evangelists, AI companies, etc). To those, a series of imperceptible small losses that results in debt in the long run is a big problem.
My suggestion (the recording session) also works as a metaphor. That could be, instead of video, metrics about how contexts are discarded. It is, in that sense, also something they can decide to share or not, and the extent to what they share should be a sign of confidence in their product.
For me it's less important how much time I think I save on any discrete task, and how much time I net over that time, accounting for how much time the set of tasks I'd be working on would have otherwise taken had I just manually done them. Right now, that means debits and credits in the ledger of time. Sometimes I gain a lot on tasks I probably otherwise wouldn't have done, but I also don't gain much overall by doing, and sometimes I lose a ton of time simply by leaning on a loop of re-doing inaccurate agent work in a way that's actually more time intensive than had I internalized the system in working memory and produced functionality more slowly.
If I save an hour, but lose 6, when I'd otherwise have spent 2, then I net -4, but sometimes overall it's positive, so the value is more ambiguous. If my employer didn't pay for the tools, I really don't know whether I would.
A good price and conservative usage pattern might net more.
You could have recorded, found it to be good, and didn't shared the news. Only used for your self. But you decided to share only the news, not the recording. That tells me something.
To be more clear, I can move this argument further. I promise you that if you share the recording that led you to believe that, I will not judge it. In fact, I will do the opposite and focus on people who judge it, trying my best to make the recording look good and point out whoever is nitpicking.
"I don't think I've ever seen anyone say they're not useful."
That's because no one has said that
"AI" hype is the issue, not "AI"
The hype machine and its followers have no tolerance for skepticism
Any perceived skepticism of "AI", no matter how reasonable, triggers absurd accusations
The author, like many others, tries to avoid the kneejerk defensiveness of "AI" hype subscribers:
"Don't get me wrong: I am not denying the extraordinary potential of AI to change aspects of our world, nor that savvy entrepreneurs, companies and investors will win very big. It will - and they will."
But this does not work. There is zero tolerance for skepticism. All disbelief must be countered
"Crypto" hype was like this, before one of its ringleaders went to prison
It's unlikely that fraud will be prosecuted under current political environment
Something not being useful is distinct from it having no uses. It could well be the case that the use of AI creates more damage than it does good. Many people have found it a useful tool to create the appearance of work where none is happening.
The thing with the hype is it's always the same hype. "If you can just 3D print another 3D printer ..." "Apps are dead, everything will be AJAX" etc. I no longer believe the hype itself warrants attention or pushback. Let the hype boys raise money. No need to protect naive VCs.
> Let the hype boys raise money. No need to protect naive VCs.
I genuinely 100% believe that ability of hype boys to raise money is harming the economy and us all. Whatever structural reason for it existing is there, it would be the best to end it.
But if the hype boys manage to capture big portions of the market (Microsoft, Amazon, etc...) it starts affecting pensions and retirement accounts. The next few years are gonna be rough because of this hype.
Those that claim not useful usually link it to something like "never trust because hallucinations", or backtrack when called out like "yes, I should have added details", or speak of problems outweighing usefulness hence not useful, etc. But online, people do make this statement.
There are thousands and thousands of comments just like this on this site. I would dare say tens of thousands. They regularly appear in any AI-related discussion.
I've been involved in many threads on here where devs with Very Important Work announce that none of the AI tools are useful for them or for anyone with Real Problems, and at best they work for copy/paste junior devs who don't know what they're doing and are doing trivial work. This is right after they declare that anyone that isn't building a giant monolithic PHP app just like them are trend-chasers who are "cargo culting, like some tribe or something".
>I also think they're over-hyped and that the current frenzy will end badly (global economically speaking)
In a world where Tesla is a trillion dollar company based upon vapourware, and the president of largest economy (for now) is launching shitcoins and taking bribes through crypto, and every Western country saw a massive real-estate ramp up by unmetered mass migration, and Bitcoin is a $2T "currency" that has literally zero real world use beyond betting on itself, and sites like Polymarket exist for insiders to scam foolish rube outsiders out of their money, and... Dude, the AI bubble doesn't even remotely measure.
Fair enough, I may have conflated "there's an AI bubble" with "AIs aren't useful".
My employer pays for Claude pro access, and if they stopped paying tomorrow I'd consider paying for it myself. Although, it's much more likely for me to start self hosting them.
So that's what it's worth to me, say $2500 USD in hardware over the next 3 years.
My dude, there is a small but weirdly dedicated group of people on this site that are hellbent on demanding "proof" that the wins we've personally gained from using LLMs in an intelligent way are real. It's actually been kind of exhausting, leading me to not weigh in on many threads.
Because there's a lot of evidence that people tend to overestimate/overstate how useful LLMs are.
Everyone says "I wrote this thing using AI" but most of the time reading the prompt would be just as useful as reading the final product.
Everyone says "I wrote this large codebase using AI" but most of the time the code is unmaintainable and probably could have been implemented with much less code by a real human, and also the final software isn't actually ready for prod yet.
Everyone says "I find AI coding very useful" and neglects to mention that they are making small adhoc scripts, or they're in a domain that's mostly boilerplate anyways (e.g. some parts of web dev).
The one killer application of LLMs seems to be text summarization. Everything else that I have seen is either a niche domain that doesn't apply to the vast majority of people, a final product that is slop and shouldn't been made in the first place, or minor gains that are worthwhile but nowhere near as groundbreaking as people claim.
To be clear, I think LLMs are useful, and I personally use them regularly. But I've gained at most 5% productivity from them (likely much less). For me, it's exhausting to keep on trying to realize these gains everyone is talking about, while every time I dig into someone claiming to get massive gains I find that the actual impact is highly questionable.
Your position implies that we need to prove that we're not smoking our own supply. I would argue that you are the one who should prove that we're not working (conservatively) 5-8x faster.
The most telling part is when you said "most of the time reading the prompt". That strongly implies that you're attempting to one-shot whatever it is that you're working on.
There is no "the prompt" in my current application. It's a 275k LoC ESP-IDF app spread across ~30 components that interact via FreeRTOS mechanisms as well as an app-wide event bus. It manages non-blocking UI, IO over multiple protocols, drives an OLED using a customized version of lvgl. It is, by any estimation, a serious and non-trivial application, and it was almost entirely crafted by LLM coding models being closely driven by yours truly across several hundred distinct Cursor conversations.
It's probably taken me 10% of the time it would have taken me to do by hand, and that's precisely because I lean on it so heavily for initial buildout, thoughtful troubleshooting (it is never tired, never not available, and also knows more than I do about electronics as a bonus) and the occasional large cross-component refactor.
I don't suspect that you're wrong. I know that you're wrong.
Sorry, that line about "reading the prompt" was unclear, I was referencing people who use it to write emails, project updates, etc. Boilerplate natural language that most people skim over because most of it is fluff. I'm well aware that large scale AI coding requires much more guidance than a single prompt.
Regardless, I would be more charitable if people like you didn't make statements such as, paraphrased, "the burden of proof is on the person suggesting the null hypothesis." I'm open to being convinced, but until now the only real study I've seen for coding shows no gain from AI. The anecdotes I have are mostly things like a coworker's 50kloc project that hasn't seen the light of day because it's massive, unreviewable, and takes as long to test as it would to have for a human to write something equivalent in <<50kloc.
Your project is the first one I've heard of that seems to be successful, and I'm curious to learn more. Just to confirm, your project is not majority boilerplate, functional enough that you'd feel comfortable releasing it to others, and you're able to get bugs fixed without too much trouble? That's pretty rare from what I've seen, and is definitely going at the top of my list of anecdotes in support of AI coding!
We're going to have to agree to disagree; the fact that you're having trouble proving a negative is not our fault or our problem.
Drop your email or some other way to contact you. I will put you on the project mailing list and you'll get a notification when the GitHub project goes public; this will happen when the first beta units go out to the presales buyers.
I'm pretty bearish on LLMs. I also think they're over-hyped and that the current frenzy will end badly (global economically speaking). Than said, sure, they're useful. Doesn't mean they're worth it.