OpenAI's problem with the movie Her and Scarlett Johansson
OpenAI and Sam Altman were scrambling yesterday amidst allegations that they'd used a Scarlett Johansson sound-alike in ChatGPT. The company's actions raise multiple concerns.
It was hard to avoid yet another blowup yesterday with the company OpenAI and its CEO Sam Altman. It turns out that one of the voices used by ChatGPT sounds remarkably like Scarlett Johansson in the Spike Jonze movie Her — and Johansson was not amused.
The voice has been available in ChatGPT for some time. But the similarity with Johansson’s character came to the fore in OpenAI’s demo of GPT-4o last week.
To many people, the link between the voice used in the demo and Johansson’s AI character Samantha in the movie Her was blatantly obvious, although OpenAI initially denied any connection.
However, in a statement issued by Scarlett Johansson it became apparent that Altman had very much wanted to use her voice in ChatGPT to emulate a Her-like AI, and had arrogantly plowed ahead despite Johansson having said “no” to multiple requests.
The case has rightly ignited a furore against Altman and OpenAI for apparently taking what wasn’t theirs despite Scarlett Johansson not wanting them to.
I hope that Johansson’s successful in any lawsuit she decides to bring. But I suspect that things will become complicated as the emulation was ostensibly of a character in a movie, and not her personally.
Regardless of this, the case raises serious issues around the decisions being made by AI companies like OpenAI, and disconnects between the talk around responsible innovation, and a reality that sometimes seems childish irresponsibility.
Just listening to the tone, mannerisms and character of OpenAI’s “Sky” voice (which is currently no longer available), it’s hard not to see the connection with the movie Her.
Certainly, Altman is on record as saying Her is one of his favorite sci-fi movies. In a conversation with Salesforce CEO Marc Beinoff last year, Altman said:
“… the number of things that, I think, “Her” got right, that were not obvious at the time — like the whole interaction model of how humans are gonna use an AI … this ideas that it is gonna be this, like, conversational language interface. That was incredibly prophetic. And, certainly more than a little bit inspired us. So, it’s, you know, it’s not like a prophecy it’s like an influenced shot, or whatever.”
This inspiration is seen quite clearly when the scene in Her where the character Theodore first encounters the AI Samantha is mashed up with OpenAI’s GPT-4o launch livestream:
And just to clinch things, after last week’s demo Altman posted just one word on X: “her”:
It seems that it’s not simply the concept of a natural language voice-interface AI agent in Her that OpenAI have taken inspiration from. It’s also the ability of Scarlett Johansson’s AI character to build an emotional — and ultimately a romantic — relationship with users that seems to have captured their imagination.
In other words, OpenAI seem to have been intent in recreating the character and behavior of Samantha in Her, and not necessarily Johansson herself … although in their naivety they seem to have conflated the two.
This may lead to a messy lawsuit if one is pursued. But even if things are settled out of court, it still raises a number of troubling questions, including:
Can a naive and even childish desire to recreate movie-inspired fantasies in AI lead to potentially harmful unintended consequences? (Not only with OpenAI, but with every other tech company that’s trying to create something they saw on a TV or movie screen).
Just how appropriate is it to aspire to create AI agents that are designed to encourage personal and even romantic connections with users?
What are the potential consequences to individuals and society more broadly of developers who show disregard for the consequences of what they do when they really want to do it? And
When are tech developers going to learn that disdain for the dignity and rights of individuals when they stand in your way is not a great business model?
Advanced AI agents aren’t going away any time soon, and are likely to become increasingly ubiquitous (as recent announcements from Google and Microsoft suggest). But as they do, let’s hope that there is true responsibility here to how they potentially impact people, and that its not just lip service from those who are trying to recreate their sci-fi fantasies.
I've said this before on many occasions but humans interpret intelligence based on language. We interpret emotions based on language. We attribute sentience based on language and ChatGPT and other LLMs are.... sophisticated language models. Add in a sultry feminine voice with proper tone and inflection and humans are bound to hallucinate a capability, personality, cognizance and sentience well beyond the underlying stochastic parrot models.
I'm afraid your question about the appropriateness of using these new voice features to "create personal and romantic relationships with users" is answered NO by society but a resounding YES by edtech companies. Why? Well, the social media playbook worked well to grab and hold our attention. The new playbook will be to grab and hold our hearts and mind. The new way to create "stickiness" is to form human-like relationships with "bots". It seem inevitable unless there is some regulation against it. I'm afraid we will look back on social media as mild in comparison to what AI will do with bots.