But the nature of LLMs is stochastic, nothing is 100%. The LLM vendors aren't dummies and train hard for this use case. But you still need a prompt that OpenAI can handle, and validating / fixing the output with an output parser, and retrying.
In my experience asking for simple stuff, requesting json_object is reliable.
with LangChain even! eye-roll, you can't really title the post 'every way' and omit possibly the most popular way with a weak dig. I have literally no idea why they would omit it, it's just a thin wrapper over the LLM APIs and has a JSON output parser. Of course people do use LangChain in production, although there is merit to the idea of using it for research, trying different LLMs and patterns where LangChain makes it easy to try different things, and then using the underlying LLM directly in prod which will have a more stable API and fewer hinky layers.
this post is a little frustrating since it doesn't explain things that a dev would want to know, and omits the popular modules. the comment by resiros offers some good additional info.
it's a bust-out, you identify a patsy and you stick it to them.
sometimes the company is worth more dead than alive, the parts are worth more the whole, especially when you can leave someone holding the bag, and the PE company gets paid to make them dead.
in any event the company is worth more to an extremely unscrupulous buyer than as a going concern in public markets.
He's a UI strategist, from a UI perspective some of these AI things are a mess.
But Google says that their mission is to organize the world's information and make it universally accessible and useful. A technology that understands and generates human language, with all its idiosyncrasies and connotations, is probably pretty important to that mission. And OpenAI stole a march on Google in commercializing it.
Maybe they justifiably panicked a little because they were starting to miss the boat.
It's a fine line between running around with your head cut off, and entering terminal decline due to too little too late.
That mission of organizing the worlds info fell to the way side, after they became the world's largest Human Attention trading (some call it thieving) marketplace.
They can't reconcile both missions.
Ideally Alphabet should be running this as separate company that is free from any obligation to incorporate what they build, into Google's existing line up of garbage no customer ever asked for. That ever growing landfill has mostly been built up to increase digital real estate, on which to sell ads. And if it all burns down its good for the planet.
When I heard it, I thought, wow, they licensed Scarlett Johansson, what an amazing Easter egg.
If they didn't and just cloned her voice, it's more disregard for creators and artists than I would have thought possible. What were they thinking?
Edit after reading the official story... not sure I believe it, seems disingenuous, at best they chose someone because they really really sounded like Scarlett Johansson, and no one said, it might be a problem.
They cannot disclose the identity of the person involved in order to protect their privacy. This is a very convenient cloak, leaving no way to know if they cloned Scarlet Johansson's voice or if it's from someone else.
Fortunately for Scarlett, she can just sue them and force them to tell her. She also isn’t shy to litigate. I was surprised that Sam wanted to poke that particular beehive.
I know trademark litigation is wishy-washy, but can she even claim her voice is unique enough to claim some form of infringement? If I just happen to sound like her, am I a walking infringement?
Voices such as Shatner or Walken are at least as much about speech patterns than the voice, giving you another axis to compare voices against, so I can somewhat see those as being trademarkable. But when I hear the ChatGPT voice, it just sounds like "slightly-flirty generic female voice 03" to me.
Sam Altman and other OpenAI employees have been referencing the movie ahead of the presentation[1][2]. I think it'd be really challenging to prove that they didn't have that exact outcome in mind.
I’m not sure it’s a trademark question so much as an appropriation of likeness question. I just meant that they can’t really train on her actual voice samples and then conceal whether they did it; that’s what discovery is for.
OpenAI is going to do a very good job of pretending that it stealing literally everything on the internet and re-selling it is some unique new activity that's not happened before. That's their core business. What they're not going to be able to do is pretend that hiring a soundalike and then making repeated references to the original voice is ok. This isn't new legal ground no matter how much money you throw at it. Otherwise every advert on TV would have some no-name actor doing impressions of well known actors. They will have to pay ScarJo a lot of money, and probably they'll have to stop using the voice too - because they've pissed her off so much at this point.
If real inflation were 10% higher than the government is reporting, then real GDP reported to be growing say 3% would be shrinking 7% per year. (We have fairly accurate numbers on nominal GDP from retail sales, sales taxes, corporate accounting etc.) Over 10 years people's standard of living would drop by 1/2. It's safe to say that has not been happening.
well, running the financial system is overpaid work but it's the high-order bit in the current form of capitalism.
it's the operating system, some people might think it's a tax on everything, some people might think it provides the foundation to produce everything of value.
similarly, Google is the high-order-bit in the information or content economy, the creators get underpaid, the people who do ad optimization get overpaid.
no financial markets -> no IPOs -> no VC -> no Google and Silicon Valley as we know it.
the closer you are to the money and the transactions, and the high-order bit, the better the opportunities to redirect and organize to your advantage, and the more you get paid.
Imagine a baby who learned everything about language and human expression from Twitter. Of course they won't only use Twitter. But it's quite hard to see Twitter data as a significant competitive advantage unless you want to train the biggest jackass in the universe.
having up-to-date data in a world where models are updated to data as of 6 months ago doesn't seem that relevant. especially when that up-to-date data is, who is Elon calling a pedo today.
You're thinking of LLMs today. When compute price drops another 99% and these models can re-train almost instantly, that "live data" will be invaluable. This is all coming from a talk I watched from Nvidia's CEO.
I think it should also run well on a 36GB MacBook Pro or probably a 24GB Macbook Air