I’m not overly worried about robots deciding to sell a stock at the same time. Human beings do it all the time. At least AIs might be expected to do it for better reasons, as well as react to such actions more sensibly.
Continue readingBiased Analyses Of AI Bias
We can and should strive to ensure that AIs aren’t overly stupid or cruel, but the premise that we can build machines that don’t just operate efficiently but provide answers that please our ever-changing opinions about bias and truth is a fool’s errand.
Continue readingExistential AI Risk Is Right Now
The ultimate AI risk isn’t it fails, but rather that it succeeds and then we’re shocked by the world it gives us.
Continue readingNews Corp Says Out Loud The Silent Part About AI
We need better understanding of the context of AI and digital transformation. News Corp has raised its hand and asked us to witness its success in thrilling its investors.
Continue readingWhy Does AI Need A Hall Monitor?
I think we have unfairly low expectations for the moral capacity of AI engineers.
Continue readingHello AI, Nice To Meet Me
How will you feel when an AI can convince everyone you know that it’s you? That day isn’t far off, if it hasn’t already arrived somewhere.
Continue readingEnough With The AI Euphemisms
Much of the talk about AI includes one or more euphemisms that substitute pleasant or unthreatening definitions for more truthful and scary qualities.
Continue readingSo Much For Slowing Down
What about concerns about deepfakes, election manipulation, privacy, massive economic or social upheaval, and even the existential threat of AI?
Continue readingOpen Source Is AI Plutonium
An open source AI based on Meta’s LLAMA model is being used to create sexbots.
Its creator is thrilled to get to experiment with state art tech, according to this story. He feels that commercial chatbots are “heavily censored.”
And that’s the argument in favor of open source development. Entrepreneurs and artists need freedom to experiment. The ugliness of a chatbot that engages in graphic rape fantasies is a small price to pay for all of the wonderful and beautiful things that might emerge from fooling around freely with AI.
After all, there’d be no Internet goodness without porn badness, especially in the early days. I’d wager that most innovators of any sort have never been particularly comfortable working with the constraints of regulations or propriety.
But calling open source AI “code” or “a model” along with a cute name or acronym doesn’t do it justice.
Open source is AI plutonium. We’re being told that we must tolerate the possibility of deadly weapons in order to enjoy power generation.
It’s not true. Sure, the strides made by using open source, like LLMs, gave developers the easiest path to the quickest results. Online customer service will never be the same. A generation of kids can cheat better on their homework assignments. AI in government and businesses is culling data to find more patterns and make better predictions.
But we can be sure that development is underway on applications that are illegal, possibly deadly, and which certainly promise/threaten to change the ways we work and live. And there’s no way to find those bad actors among the good ones until their badness appears in public.
It could even impact us and we wouldn’t know that AI was responsible.
So, we might never figure out that errant AI have been quietly manipulating stock prices or skewing new drug trials. It could sway elections, entertainment reviews, and any other crowdsourced outcome. Bad actors, or an AI acting badly, could encourage social media ills among teens and start fights between adults.
It might even start a war, or decide that nuclear weapons were the best way to end one.
Unlike plutonium, there’s no good or reliable way to track or control such outcomes, no matter how transparent the inputs might have been.
In true Orwellian fashion, the CEO of a site that promotes open source argues that the real risk is from businesses that are “secretive” and take at least some responsibility for their AI models, like Google and OpenAI. A VC exec who promotes AI worries that relegating development to big companies means “they’re only going to be targeting the biggest use-cases.”
It’s a false dilemma. I’d happily “censor” a porn application in exchange for a cure for cancer, especially if it came with the likelihood that the world wouldn’t get blown up along the way.
Who’s Responsible For AI?
Responsible AI wouldn’t rely on good intentions or mere compliance with regulations, but on designing responsibility into the technology builds themselves and then sharing those details fully and regularly over time.
Continue reading