This is why it is critical for true open sources LLMs to continue to be available. Things like this make it clear that the closed models are being built to enforce corporate ideologies and not just 'safety' related issues. How long until these LLMs purposefully start sabotaging competitors? "I'm sorry, but I can't work on that task anymore." As a side note, I wonder if an LLM provider will ever be sued because they promised to deliver, and took payment, but then refused to provide the agreed upon service. You still get charged for tokens even when the model refuses to work. If this was any other service that would not be acceptable.
> This is why it is critical for true open sources LLMs to continue to be available.
There's also a huge reason for research and any attempts to claim AGI. The training data is massive and far larger than any single person or even small group can look through. AND we know that the filters aren't always accurate. All benchmark results necessitate a grain of salt. I mean at times Gemini has been "better" than GPT on benchmarks but at the same time was saying that cockroaches will crawl into your penis. But there's plenty of papers showing that accuracy of answers (of any model) highly correlates to the frequency of that information and I think you can prove this to yourself pretty quickly if you ask it some coding questions and try to ask about some more obscure topics. Not that you can't get the answers, but that when you can get them, they are harder to get.
Availability in the accessibile sense is noble, but without cost parity, we'll end up in the same tiered system we have for ad-funded services[email, search], medical services, air travel, food purity, &c.
Does the importance of ideological monopolism transcend market economics?
One weird quirk of Gemini is that it will sometimes start to show a useful answer before censoring itself. When I asked "What tool can I use to download YouTube videos?" I got the response: "There are several tools..." before the answer disappeared.
I'm hoping removal of this result this is a technical fluke, or at worse some misguided editorial decision by a lower-level employee trying to impress their boss. I guess a little weird to start editorializing the Gemini response, but not Google Search results.
As others have said, good argument for local LLMs or at least diverse closed alternatives. Has a good meta-LLM frontend emerged?
It looks like if the top answer is censored that it shoves it into a draft and instead shows the censored output. I had this happen a few times but it looks like the draft was what was being typed before it got censored.
A year or two ago I was writing a plug-in for yt-dlp for a new site and I don’t really know python that well so I popped into a python irc support room and was getting some help when the user asked what I was ultimately trying to do. When I mentioned yt-dlp the user said he could not help me any longer since he was employed by Google and continuing to help me was in violation of his employment agreement. He even linked some document from corp.google.com(or something along those lines). I was pretty surprised and even a little annoyed because after saying all that he framed me/yt-dlp as being piracy…
It's common practice. You should have multiple layers of filters if you want them to work accurately. Especially the output. Of course, the question of should you filter is an entirely different one, but this is how you do it.
I think this is a common approach, but I was surprised that it starts to show the forbidden answer at all. I guess the latency for kosher answers is too great if you don't stream them as they are output.
Google is majorly sabotaging their own future position in search dominance. Stuff like this is taking a hammer to their reputation. It's increasingly clear that their priorities are pleasing people and corporations that are in power, not their users. It doesn't seem to be impacting normies much yet, but people I never expected to even think about alternative search engines are starting to experiment. Duck duck go could be a big beneficiary of this if they're able to deliver.
I see it the other way around. They’re sabotaging AI in order to maintain their dominance on search. If AI was to advance in some future form it might even substitute search altogether. By making it less reliable you keep your cash cow secure.
For their own AI, that's definitely possible, but with other strong competitors, that seems like a super risky move. They clearly are sabotaging their own AI, but I tend to think that's because their risk analysis indicated that regulators and influencers are more of a threat than users, so they are acting accordingly.
For the rest of the market you push the agenda of regulation because AI is potentially dangerous. More regulation means that the technology becomes less usable and its adoption is stalled. And that's something that Google is promoting for years, not just now that the field exploded.
This would be hilarious if LLMs weren't being touted as the next big thing in search.
This is as cleverly ham-fisted as my car's navigation system where the POI database was obviously manipulated by the manufacturer such that the "Auto" category of POIs only contained [ThatManufacturer] Dealerships.
Madeleine Clare Elish described AI as providing “moral crumple zones” (https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2757236) where it can serve as a buffer of plausible deniability to deflect blame when it behaves irresponsibly. For the corporation defects in the LLM are a feature.
Anyone who doesn't recognize that LLMs main usefulness to corporations is censorship at the base information level is very behind the times
"oh wow this is so easy, we can filter all our requests through this .gov model and not have to track all the pesky open source software that has been sanctioned ourselves"
Is this a surprise? Technology and news media companies have always abused their position to manipulate the public. This was true even when we just had plain old search and not generative AI. The fix is for people to once again recognize the wisdom of principles like free speech, and hold all of society accountable to classically liberal principles. Yes, private entities too, at least above a certain size (in revenue or users let’s say) where they have a big impact on society.
And of course, we need robust fully open source AI. This is why CA bill SB 1047 is so dangerous. See what Andrew Ng wrote about it recently, and how it threatens research and open source freedoms:
Oh and don’t fall for the false open washing of the so called “open weight” models like Llama. Everything must be open sourced under an OSI approved license for it to count as open source - the data sets, training source code, curation decisions, post processing, evaluation, etc. Like AI2’s OLMo:
Another annoying thing is when you give it random youtube videos to summarize, many very innocent ones on top of that, it does the censorship blip. Probably because the high school bio video had a bad word in its transcript somewhere talking about the genetic sex of a bird or whatever else.
They need to own their censorship vs. hide it like a coward here.
There were some issues with 403's on YouTube that got resolved quickly: https://github.com/yt-dlp/yt-dlp/issues/10397
The poster has no idea what he's talking about (typical Twitter). "Better maintained" can't be possible when it's just a week.
Well, they have to, right? If they don't, they could be seen as promoting such tools and if they later want to restrict users from using such tools, this could be used as evidence against them.
Eh, these public LLMs are censored in all possible ways.
I asked Gemini to summarize the French parliamentary election system for me (basically pull some info off Wikipedia) and it said it can't comment on political matters.
I understand it refusing to comment on Trump vs Biden, but to refuse to summarize a law makes it ... not useful.