What do you mean it 'utters insensitive statements', this is not a chat bot, it's text to speech, it generates speech for exactly what you type in, the threat is that it can be trained to sound like exactly anyone and the potential for fake news etc is high.
The trouble is, this is already partly here and going to be increasingly available. For smaller startups that already do this like ElevenLabs, they have no reason to hold back, as it's there entire business, Meta can for now, but it's a fruitless exercise. These tools are coming.
Meta doesn't want the flack hurting or causing investigations into it's current business is more what it means, it will release the tool when others have released there's and it will no longer be the one to blame.
'Utters insensitive statements' is clearly the domain of humans, checking on my threads and I realise what an ass I sound like, I didn't mean to be so combative with it. Apologies.
More like the engineers and scientists built it, but legal and PR think it's a bad idea to release. Which they're probably right about considering the reception Galactica got.
Not entirely sure why this was downvoted, it was my initial thought too (though I must concede I’m reading the comments before the article… which may give me my answer…)
This article is a good promotional article to get people hyped about this new product. I must admit, it has sparked my interest. I'm eager to use it. At last, I won't have to deal with inconsistent voice actors for my wizard demos / support pages.
Surely, they may have to incorporate some safety measures, much like OpenAI has.
Does anyone know who the leader is in affordable speech synthesis for hobby projects ?
Atm I am using Amazon Polly but am unsure if another offering is far superior. eleven labs is not affordable for my slightly serious level of job ting but aws and google seem well priced. Any other suggestions for a product I can try out?
I've tried this and it took ~5 minutes to generate 10 seconds of studio on my 3080. That just doesn't work if you're trying to generate an hour long podcast.
Take a look at Piper. It's the tts solution used by the open source home automation project HomeAssistant. Produces decent quality speech in a couple seconds on raspberry pi class hardware.
They said the same thing about GPT-3. It's a marketing push.
Oh no, our incredible scientists made a thing that's simply too amazing and they're afraid it might take over the world! We just thought you might like to know how nice and ethical we're being with our immense power...
The goal of that kind of project in a huge tech org is to have capability in-house. So that they can play that game if they have to. Seems like they don't see themselves in a situation where their position without the tech would be worse than with the tech deployed. In some ways it's a bit like defense spending in a way, the most successful defensive army is the one strong enough to never get tested (in many other ways it's not)
Adobe demoed a similar project as far back as 2016 called "Project Voco" [1] which was also called "too dangerous to release" at the time, even though it apparently still needed as much as 20 minutes of source material (vs. allegedly a mere 2 seconds here).
It was never heard from again afaik - even though Adobe is not known to shy away from an opportunity to increase revenue, so one cannot help but wonder...
What would happen if only bad actors had access to LLMs, deepfakes, and stable diffusion models? They could convincingly fake image/video evidence and accompanying online support for disinformation, and most people would buy it.
But today, given the widespread availability of these tools, most people know what they're capable of, and when a meme photo of Donald Trump in handcuffs gets circulated nobody takes it seriously because they have seen dozens of fakes before.
So IMO if you really want to minimize the potential for your AI tool create chaos, release it to the public. Show people what it's capable of. Once people tire of memes of Obama spouting the navy seal copypasta, they will be prepared to call out disinformation generated by these tools as well.
I believe everyone can understand this - and they will need to releases it to public. Surely, they may have to incorporate some safety measures, much like OpenAI has.
Anyway a fantastic promotional strategy to get people hyped about this new product. I must admit, it has sparked my interest. I'm eager to use it. At last, I won't have to deal with frustrating and inconsistent voice actors.
AI may be meta's only and real "grow out of newsfeed business" card. They should probably pivot, release their own version of ChatGPT/VoiceGPT, start charging for it and maybe rebrand (metai?). They are currently not being taken seriously, despite the chops they have, due to their stupid obsession with social and connecting people.
Perhaps by consumers but among the AI industry/NLP I would view them as the top 3 most sophisticated companies. They have also gained a ton of respect from me for not joining in with the "closed" approach other companies have all now adopted.
They have, but when OpenAI started spouting "too dangerous to release" was about the same time they effectively dropped the "Open" part. I hope that is not the same for Meta. But wasn't Meta not releasing model weights except for a leak? In that respect are Meta and OpenAI already equivalent?
OpenAI isnt even oublishing papers any more, and Meta has been publishing model weights for many models for quite some tome and have also commited to releasing their next language model open
Bell labs thought something similar about the first answering machine, fearing people would be held to their word resulting in fewer telephone calls.
Now big tech worries that "perfect" trust in their communication platform goes away, but the paradigm shift happens no matter what. Many governments already have the power to post as users on platforms and that should be cause to usher in this new trust calculation faster.
Whoever needs to use it at deep-state level for nefarious purposes will either access it or find something similar that does the same job.
I think it should be released so that the public can see what can be faked and they get used to the new paradigm of everything can be faked so they don't immediately believe in everything they see/hear.