This whole thing is basically a nonstory when you realize how much money is in tech. Meta changed their name and sank billions on an idea that everyone thought was stupid from the beginning, and they're still fine.
Putting a billion into the flavor-of-the-month that has like 10% chance to be the next big thing is a no-brainer when you're printing multiple billions in profit doing nothing, and have a lot more cash on hand.
The real story, is how wealth inequality and monopolies have essentially allowed the rich to waste tons of money chasing more wealth while having almost no incentive to provide value to society. Who gives a fuck about hallucination and prompt injection? It's all trivial details that VCs are giving away billions to eventually solve.
I mean, that’s a little like asking, “How many swift kicks to my nuts before it makes me a billionaire?”
Because it’s one of those cost evaluation situations where they thought it was a shot in the dark at first, but by now it’s clearly a loss. So, the whole thing feels a little like, The Producer’s, something isn’t smelling right for any outsiders.
Me, an "AI written" regurgitated article on a content mill website uncaringly misleading you about the best way to get connected to a purple banana fridge.
For me, that depends on a whole lot of things. Who owns it? Who pays for its operation? How much and what access does it need? What verifiable privacy protections are there? How transparent are its processes?
I’m just generally suspicious of companies whose products are ultimately meant to be integral to our daily lives. These kinds of tools will undoubtedly need access to a huge chunk of our personal and professional data to work effectively. I’d rather not interact with the world through the lens of someone’s corporate vision. With the right protections? Sure, I could see some cool and creative use cases. But that unfortunately brings me back to my general skepticism.
Gonna just buck the trend and say that this AI push has me excited for the future. It's easy to be a nay-sayer, but I genuinely believe the leaps made in AI in just the last year are amazing.
The author clearly doesn't like AI, and completely mischaracterizes Mistral AI for things their models could say, but doesn't consider at all why unaligned models are useful in developing your own.
The author likes to highlight that sometimes an AI will make things up, a phenomenon known as hallucinating. Hallucinations could also be called "creativity" in certain contexts. This isn't always a fault, especially when creativity is the intended purpose.
The author pointed out how it's possible to prompt engineer out sensitive data, and how there's a lack of privacy... which isn't a problem with the tech, but rather tech companies.
The technology used behind the scenes with ChatGPT isn't exclusively for text generation. I'm seeing it appear in speech to text / text to speech applications. It's showing up in image and video editing. It's showing up in ... well ... images/movies of an adult nature.
You're probably already consuming AI generated content without even realizing it.