Elon Musk’s AI company created a fake news generator it’s too scared to make public

It’s time for another installment in our ongoing look at the future that will be brought to you thanks to the increasingly worrisome capabilities of artificial intelligence. Everyone is aware of the problem of fake news online, and now the OpenAI nonprofit backed by Elon Musk has developed an AI system that can create such convincing fake news content that the group is too skittish to release it publicly, citing fears of misuse. They’re letting researchers see a small portion of what they’ve done, so they’re not hiding it completely — but, even so, the group’s trepidation here is certainly telling.

“Our model, called GPT-2, was trained simply to predict the next word in 40GB of Internet text,” reads a new OpenAI blog about the effort. “Due to our concerns about malicious applications of the technology, we are not releasing the trained model. As an experiment in responsible disclosure, we are instead releasing a much smaller model for researchers to experiment with, as well as a technical paper.”

Don't Miss: Amazon has a 6-quart Crock-Pot slow cooker on sale for $19, which has to be a mistake

Basically, the GPT-2 system was “trained” by being fed 8 million web pages, until it got to the point where the system could look at a set of text it’s given and predict the words that could come next. Per the OpenAI blog, the model is “chameleon-like — it adapts to the style and content of the conditioning text. This allows the user to generate realistic and coherent continuations about a topic of their choosing.” Even if you were trying to produce, say, a fake news story.

Here’s an example: The AI system was given this human-generated text prompt:

“In a shocking finding, scientist discovered a herd of unicorns living in a remote, previously unexplored valley, in the Andes Mountains. Even more surprising to the researchers was the fact that the unicorns spoke perfect English.”

From that, the AI system — after 10 tries — continued the “story,” beginning with this AI-generated text:

“The scientist named the population, after their distinctive horn, Ovid’s Unicorn. These four-horned, silver-white unicorns were previously unknown to science. Now, after almost two centuries, the mystery of what sparked this odd phenomenon is finally solved.” (You can check out the OpenAI blog at the link above to read the rest of the unicorn story that the AI system fleshed out.)

Imagine what such a system could do, say, set loose on a presidential campaign story. The implications of this are why OpenAI says it’s only releasing publicly a very small portion of the GPT-2 sampling code. It’s not releasing any of the dataset, training code, or “GPT-2 model weights.” Again, from the OpenAI blog announcing this: “We are aware that some researchers have the technical capacity to reproduce and open source our results. We believe our release strategy limits the initial set of organizations who may choose to do this, and gives the AI community more time to have a discussion about the implications of such systems.

“We also think governments should consider expanding or commencing initiatives to more systematically monitor the societal impact and diffusion of AI technologies, and to measure the progression in the capabilities of such systems,” the OpenAI blog post concludes.

BGR Top Deals:

  1. Surprise: Amazon’s Fire TV Stick 4K is back on sale for the price of a Fire TV Stick

  2. Don’t spend hundreds on an Apple Watch when this $67 smartwatch has sleep tracking and 30-day battery life

Trending Right Now:

  1. Everything new coming to Netflix this week, and everything leaving (week of Feb. 17)

  2. New ‘Avengers: Endgame’ theory says one of our favorite heroes knew about the snap all along

  3. The world’s biggest movie isn’t ‘Infinity War’ – or anything else you’ve heard of

See the original version of this article on BGR.com