0

WHY THIS MATTERS IN BRIEF

Increasingly people can type text into an AI and the AI will auto-generate all kinds of content, from articles and scientific papers to code and even drugs …

 

Love the Exponential Future? Join our XPotential Community, future proof yourself with courses from XPotential University, read about exponential tech and trendsconnect, watch a keynote, or browse my blog.

Imagine if Artificial Intelligence (AI) could write you a college essay, or spit out a movie review in the style of Shakespeare. Well, recently we’ve seen a ran ge of AI’s, from the likes of Google and OpenAI, do just that and a lot more including auto-generating blogs, imagery, videos, and even re-creating science experiments, doing homework, and magicking up university lecturers from nothing more than text inputs. And, elsewhere other AI’s are generating code, games, and even new drugs and rocket engines using just the most basic of inputs.

 

RELATED
An AI invented over 40,000 new chemical weapons in just 6 hours but it's OK

 

OpenAI last week opened up access to ChatGPT, an AI powered chatbot that interacts with users in an eerily convincing and conversational way. Its ability to provide lengthy, thoughtful and thorough responses to questions and prompts – even if inaccurate – has stunned users, including academics and some in the tech industry.

 

The Future of Artificial Intelligence, by keynote speaker Matthew Griffin

 

The tool quickly went viral. On Monday, OpenAI’s co-founder Sam Altman, a prominent Silicon Valley investor, said on Twitter that ChatGPT crossed one million users. It also captured the attention of some prominent tech leaders, such as Box CEO Aaron Levie.

“There’s a certain feeling that happens when a new technology adjusts your thinking about computing. Google did it. Firefox did it. AWS did it. iPhone did it. OpenAI is doing it with ChatGPT,” Levie said on Twitter.

But as with other AI-powered tools, it also poses possible concerns, including for how it could disrupt creative industries, perpetuate AI biases and spread misinformation.

 

RELATED
South Park creators announce the world's first DeepFake sattire TV show

 

ChatGPT is a large language model trained on a massive trove of information online to create its responses. It comes from the same company behind DALL-E, which generates a seemingly limitless range of images in response to prompts from users. It’s also the next iteration of text-generator GPT-3.

 

The Future of synthetic Content, by keynote speaker Matthew Griffin

 

After signing up for ChatGPT, users can ask the AI system to field a range of questions, such as “Who was the president of the United States in 1955,” or summarize difficult concepts into something a second grader could understand. It’ll even tackle open-ended questions, such as “What’s the meaning of life?” or “What should I wear if it’s 40 degrees out today?”

“It depends on what activities you plan to do. If you plan to be outside, you should wear a light jacket or sweater, long pants, and closed-toe shoes,” ChatGPT responded. “If you plan to be inside, you can wear a t-shirt and jeans or other comfortable clothing.”

But some users are getting very creative.

One person asked the chatbot to rewrite the 90s hit song, “Baby Got Back,” in the Style of “The Canterbury Tales;” another wrote a letter to remove a bad account from a credit report – rather than using a credit repair lawyer. Other colorful examples including asking for fairy-tale inspired home décor tips and giving it an AP English exam question which it responded with a 5 paragraph essay about Wuthering Heights.

 

RELATED
SpaceX announces 2018 mission to Mars

 

In a blog post last week, OpenAI said the “format makes it possible for the tool to answer follow-up questions, admit its mistakes, challenge incorrect premises, and reject inappropriate requests.”

While ChatGPT successfully fielded a variety of questions submitted by users, some responses have been noticeably off. In fact, Stack Overflow – a Q&A platform for coders and programmers – temporarily banned users from sharing information from ChatGPT, noting that it’s “substantially harmful to the site and to users who are asking or looking for correct answers.”

Beyond the issue of spreading incorrect information, the tool could also threaten some written professions, be used to explain problematic concepts, and as with all AI tools, perpetuate biases based on the pool of data on which it’s trained. Typing a prompt involving a CEO, for example, could prompt a response assuming that the individual is white and male, for example.

“While we’ve made efforts to make the model refuse inappropriate requests, it will sometimes respond to harmful instructions or exhibit biased behavior,” Open AI said on its website. “We’re using the Moderation API to warn or block certain types of unsafe content, but we expect it to have some false negatives and positives for now. We’re eager to collect user feedback to aid our ongoing work to improve this system.”

 

RELATED
Futuristic Spinach can send you emails about climate change, life, and stuff

 

Still, Lian Jye Su, a research director at market research firm ABI Research, warns the chatbot is operating “without a contextual understanding of the language.”

“It is very easy for the model to give plausible-sounding but incorrect or nonsensical answers,” she said. “It guessed when it was supposed to clarify and sometimes responded to harmful instructions or exhibited biased behavior. It also lacks regional and country-specific understanding.”

At the same time, however, it does provide a glimpse into how companies may be able to capitalize on developing more robust virtual assistance, as well as patient and customer care solutions.

While the DALL-E tool is free, it does put a limit on the number of prompts a user can do before having to pay. When Elon Musk, a co-founder of OpenAI, recently asked Altman on Twitter about the average cost per ChatGPT chat, Altman said: “We will have to monetize it somehow at some point; the compute costs are eye-watering.”

About author

Matthew Griffin

Matthew Griffin, described as “The Adviser behind the Advisers” and a “Young Kurzweil,” is the founder and CEO of the World Futures Forum and the 311 Institute, a global Futures and Deep Futures consultancy working between the dates of 2020 to 2070, and is an award winning futurist, and author of “Codex of the Future” series. Regularly featured in the global media, including AP, BBC, Bloomberg, CNBC, Discovery, RT, Viacom, and WIRED, Matthew’s ability to identify, track, and explain the impacts of hundreds of revolutionary emerging technologies on global culture, industry and society, is unparalleled. Recognised for the past six years as one of the world’s foremost futurists, innovation and strategy experts Matthew is an international speaker who helps governments, investors, multi-nationals and regulators around the world envision, build and lead an inclusive, sustainable future. A rare talent Matthew’s recent work includes mentoring Lunar XPrize teams, re-envisioning global education and training with the G20, and helping the world’s largest organisations envision and ideate the future of their products and services, industries, and countries. Matthew's clients include three Prime Ministers and several governments, including the G7, Accenture, Aon, Bain & Co, BCG, Credit Suisse, Dell EMC, Dentons, Deloitte, E&Y, GEMS, Huawei, JPMorgan Chase, KPMG, Lego, McKinsey, PWC, Qualcomm, SAP, Samsung, Sopra Steria, T-Mobile, and many more.

Your email address will not be published. Required fields are marked *