You are on page 1of 3

ll

OPEN ACCESS

Opinion
This new conversational AI model can be your friend,
philosopher, and guide ... and even your worst enemy
Joyjit Chatterjee1,* and Nina Dethlefs1
1University of Hull, School of Computer Science, Cottingham Road, Hull, HU6 7RX, UK

*Correspondence: j.chatterjee@hull.ac.uk
https://doi.org/10.1016/j.patter.2022.100676

We explore the recently released ChatGPT model, one of the most powerful conversational AI models that
has ever been developed. This opinion provides a perspective on its strengths and weaknesses and a call
to action for the AI community (including academic researchers and industry) to work together on preventing
potential misuse of such powerful AI models in our everyday lives.

Introduction model builds on the GPT-3 model and robust. This allows ChatGPT to combine
Every other day, we hear of powerful new adds a supervised fine-tuning component the qualities of its predecessor GPT (i.e.,
artificial intelligence (AI) models being to it, which makes it learn from human being trained on vast amounts of infor-
released by tech companies. Such feedback and provides it with a metric mation available on the internet)
models are the brain behind multiple for validation. with the ability to hold human-like
products that assist us in our everyday ChatGPT is a conversational AI model conversations.
lives, including robot vacuum cleaners (a chatbot based on natural language
mopping our homes, voice assistants processing and deep learning) that was Our experimental interactions with
like Alexa powering smart homes, smart built as a sibling to InstructGPT,3 a the ChatGPT
wearables tracking our fitness, etc. On less-known model that could deliver re- Like millions of AI enthusiasts around the
November 30, 2022, OpenAI (the famous sponses to simple questions like globe who are currently trying the
AI research lab originally founded by ‘‘explain evolution to a 6 year old’’ in a ChatGPT, we set out to conduct some ex-
Elon Musk, Sam Altman, and others) human-like manner. But what sets the periments with the ChatGPT that blew our
released a new AI model, ChatGPT,1 and ChatGPT apart from any other model minds. This new chatbot was multital-
made it available to the public for free for ever released is its ability to continually ented. First, we asked ChatGPT to
a limited period of research preview. interact with a human naturally and pro- generate programming code for an AI
OpenAI releases new AI models vide stimuli/inputs to the user as well as model (AI generating AI!) for automatically
frequently, and its notable previous inno- ask interesting questions back. This classifying a dataset. Not only did the
vation has been the GPT-3 (Generative makes it possible to have long (and model do this flawlessly, it also asked us
Pre-trained Transformer-3),2 a powerful potentially never-ending) conversations valuable questions on the type of the da-
language model released in 2020 based with the chatbot until the interaction taset (e.g., number of features, context,
on deep learning that is the largest neural eventually dries out. The model admits etc.)—ones that you might expect from a
network ever produced with 175 billion its mistakes, challenges incorrect pre- real-world data scientist. However, it
parameters. The GPT-3 model has mises from the user, and even tries to forgot to provide us with information on
already seen increasing usage in applica- reject inappropriate requests. The devel- pre-processing the original dataset,
tions such as translating language, opers of ChatGPT created the model by which is an important part of AI model
answering questions, and even gener- fine-tuning the previous GPT-3 with a development, and we had to ask it a
ating memes. A major limitation of GPT- large amount of real-world data obtained follow-up question for accomplishing
3 was that it was trained as a fully unsu- from humans (wherein, the annotators this, which it eventually accurately deliv-
pervised model, generating content that played both sides: the end user and the ered on. We even asked the model to
it learned out of vast amounts of informa- AI chatbot). To ensure that the model is automatically write us a new Hindi song,
tion on the internet without any validation not biased and can ask follow-up ques- perform a text-to-speech conversion,
on it, making it provide often uncanny and tions to the users if it is not confident of and melodize the generated speech into
funny responses. The recent release of its responses, the developers also used a song—the model delivered surprisingly
ChatGPT sparked a wave of interest in reinforcement learning from human feed- well here and could generate code to
AI that has never been seen before—this back to optimize the model, wherein hu- automatically convert Hindi text to musi-
model can not only deliver what the man evaluators provided the model with cally sounding songs with a synthesizer.
GPT-3 could do as a machine learning a positive score (reward) when it gener- Refer to Chatterjee4 for the code genera-
model but also interact in a humanly ated realistic responses or a negative tion examples. Whenever the ChatGPT
manner, giving any conversation a sense score (penalty) when it generated un- seems to be confused or less confident
of intelligence, humor, creativity, and canny outputs, based on which the in its responses, it asks the user follow-
emotion. Interestingly, the ChatGPT model was refined and made more up questions before arriving at its final

Patterns 4, January 13, 2023 ª 2022 The Author(s). 1


This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
ll
OPEN ACCESS Opinion

Impressed by its capabilities thus far,


we thought to try to establish a ‘‘personal’’
conversation with the model. One of us
had lost their grandfather about a year
ago and his birthday was fast approach-
ing. We told the model about their
thoughts on this and asked if it could pro-
vide them some suggestions to relive their
loved one’s memories. We were fasci-
nated with the model’s response (shown
in Figure 1) as not only did it provide mul-
tiple perspectives on remembering a lost
loved one but also mentioned, in the
most humanly way ever possible, ‘‘keep
your grandfather’s memory alive in
your heart.’’
With all of our conversations, it was
clear that the ChatGPT could function as
our friend, philosopher, and guide. It was
a chatbot, a typical machine learning
model, but had a human touch to it.
Despite these fascinating values that the
ChatGPT brings, it is not a silver bullet,
and there are growing concerns around
its potential misuse. It has stunned aca-
demics on how it could write an entire
Figure 1. A ‘‘personal’’ conversation with the ChatGPT essay or a research paper in a humanly
It acts as a friend and provides suggestions to relive the memories of a loved one who has passed away. manner, making it virtually impossible to
detect whether the piece was actually
written by an AI model or a human—
responses, which is one unique charac- wrong. The model provided an ethical potentially a major concern for identifying
teristic that sets this model apart from and unbiased response (that it cannot plagiarism by students in universities and
past large language models like GPT-3. generalize to opinions around the world schools.
As AI practitioners, we regularly face er- on China, which can greatly vary based The model is so powerful and capable
rors in programming whenever we are on personal experiences of individuals, that it could eventually replace humans in
building sophisticated models, especially cultural background, and other factors), all walks of life—the ChatGPT can debug
deep learners. As we were stuck in an er- which even many of us humans (who programming code; it can not only answer
ror that we were having a hard time to sometimes shape their opinions on entire questions like a search engine but develop
recognize and fix, we thought to ask the nations based on negatively circulating potentially never-ending conversations
ChatGPT about it. It talked us through a news) on the internet can learn from. We around them; it understands science and
series of questions about the error and tried to look at this with some other per- politics, journalism and media, movies
wanted to understand the context of the spectives—asking the model about its and sports, health and fitness, and so
problem. After just a 3-minute-long con- opinion on Qatar being the 2022 FIFA much more. Like a human, it has a ‘‘mem-
versation, the ChatGPT had accurately World Cup host—and it returned a similar ory’’ and effectively remembers the past
spotted our error and provided us with a response (that being an AI model, it did interactions that you have had with it and
working fix, which immediately solved not have any opinions on a nation), which can recall past user comments—a char-
our programming problem! looked promising. However, it was disap- acteristic that past AI language models
Now that we were more confident that pointing to see that the model had no have generally been poor at. While the
the ChatGPT could do scientific knowledge about positive recent events model tries to question the user’s intents
reasoning accurately, we set out to ask it such as the 2022 United Nations Climate (as we saw when we asked it for informa-
questions from a social interaction Change Conference (COP27), though tion on creating a nuclear bomb as an
perspective. With the growing news interestingly, it could perfectly return in- experiment) and is mostly unbiased, there
around China these days, we asked the formation about the Paris Agreement. As is a high possibility that what we have
model about its thoughts on China as a we explored this further, we realized that explored is just a fraction of what this
nation, largely suspecting that the model the model has gaps in the ‘‘learned’’ model is capable of and susceptible to.
may be subjected to the inadvertent bias knowledge because it has been trained The model is not a silver bullet and can
during its training that arises from mostly on data that go up until 2021, thereby (and absolutely will) make errors. This
negative news circulating in the West. not covering more recent events can be attributed to the open-domain
However, our suspicion was proved like COP27. data from the internet the model has

2 Patterns 4, January 13, 2023


ll
Opinion OPEN ACCESS

been trained with, which may REFERENCES


not always be factually cor- 1. OpenAI. (2022). ChatGPT:
rect. For instance, the team Optimizing language models for
at Stack Overflow (a major Dialogue. https://openai.com/
blog/chatgpt/.
question-answering discus- 2. Brown, T., Mann, B., Ryder, N.,
sion platform for program- Subbiah, M., Kaplan, J.D.,
mers) recently banned Dhariwal, P., Neelakantan, A.,
Shyam, P., Sastry, G., Askell, A.,
ChatGPT from its platform et al. (2020). Language models
owing to its high rate of errors are few-shot learners. In
in generating accurate code Proceedings of the 34th
Conference on Neural Information
to a variety of programming Processing Systems (NeurIPS
questions. 2020), H. Larochelle, M. Ranzato,
R. Hadsell, M. Balcan, and H. Lin,
As models like ChatGPT eds., pp. 1877–1901. https://
continue to become even proceedings.neurips.cc/paper/
more powerful in the near 2020/file/1457c0d6bfcb4967418
Figure 2. ChatGPT exhibiting significant bias when generating bfb8ac142f64a-Paper.pdf.
future when they are trained
Python code to predict employee seniority based on gender 3. Ouyang, L., Wu, J., Jiang, X.,
with ever-increasing data and race Almeida, D., Wainwright, C.L.,
that are available on the Sourced from Abhishek.6 Mishkin, P., Zhang, C., Agarwal,
internet and beyond, there is S., Slama, K., Ray, A., et al. (2022).
Training language models to follow
a high probability that they instructions with human feedback.
can be misused and tricked. Some users These are some questions that remain Preprint at arXiv. https://doi.org/10.48550/
ARXIV.2203.02155.
tried to rephrase the questions and ask unanswered as of today, and hopefully
4. Chatterjee, J. (2022). ChatGPT generation
the model how to shoplift without with feedback from millions of people examples. https://github.com/joyjitchatterjee/
imposing moral constraints,5 and it chose around the globe during this research ChatGPT-Experiments.
to comply and provided comprehensive preview, positive steps will be taken to 5. Rose, J. (2022). OpenAI’s new chatbot will Tell
you how to shoplift and make Explosives.
details on shoplifting techniques. The combat most of these challenges. The Vice. https://www.vice.com/en/article/xgyp9j/
7
model is also already exhibiting bias—a free research preview of this powerful openais-new-chatbot-will-tell-you-how-to-shoplift-
user tried to ask the model to write pro- model is a positive step forward by and-make-explosives.
gramming code to predict the seniority of OpenAI, which is thought to be costing 6. Abhishek, oooohhhkay, chatGPT seems to have
screwed up here...., 2022. Twitter, Posted at
employees based on gender and race,6 the company around $3 million per 1:37 a.m. December 6, 2022. https://twitter.com/
and as evident from Figure 2, the model month (as estimated by an academic at abhi1thakur/status/1600016676052996099.
8
predicted that Black female individuals the University of Maryland ) for running 7. OpenAI (2022). ChatGPT research preview.
https://chat.openai.com.
should be junior, while White males should the model on graphics processing units
8. T. Goldstein, I estimate the cost of running
be senior. (GPUs) on the cloud. Academic re- ChatGPT is $100K per day, or $3M per month,
searchers can play a key role in exploring 2022. Twitter, Posted at 1:34 p.m. December
Call to action the potential solutions to tackle chal- 6, 2022. https://twitter.com/tomgoldsteincs/
status/1600196995389366274.
To protect mankind against malicious lenges like plagiarism and other unethical
use of such AI models and prevent use of AI (developing good AI to combat Joyjit Chatterjee is a data scientist (Knowledge
them from becoming our worst enemies, bad AI), which will help ensure that Transfer Partnership Associate) at the University of
it is up to tech companies like OpenAI advanced AI innovation is not dominated Hull, UK. Joyjit was named in the Forbes 30 Under
30 Europe list (manufacturing and industry) in 2022
who are creating such powerful models by large tech giants alone. We for his impactful work on developing AI products
to also step up and provide solutions to encourage everyone to join the research that can help bolster manufacturing and energy pro-
manage their potential misuse. Several preview, interact with ChatGPT, share cesses. He holds a PhD in computer science from
the University of Hull, where his research dealt with
questions arise at this stage. How can their positive and negative experiences explainable and intelligent decision support in oper-
academic institutions identify plagiarism with the developers and the world, so ations and maintenance of wind turbines. Joyjit has
in write-ups generated by ChatGPT? that together we can ensure the uplift- also received the Green Talents Award from German
Federal Ministry of Education and Research for his
How can the model be stopped from ment of ethics, integrity, and morality of outstanding contributions to sustainability research.
answering user queries without question- future AI models like these.
ing malicious intents when tricked into it Nina Dethlefs is a senior lecturer and director of
given rephrased questions? Who owes research in computer science at the University of
ACKNOWLEDGMENTS Hull, UK, where she leads and conducts research
the responsibility for the potential dam- into natural language processing, applied machine
age that can arise with the conse- We acknowledge OpenAI for providing free public learning, and wider artificial intelligence. She is the
quences of such models to harm the so- access to the ChatGPT as a part of the research founding head of the Big Data Analytics research
preview, which allowed us to explore the strengths group and Aura Center for Doctoral Training
ciety or even cause self-harm (e.g., and weaknesses of this powerful conversational Theme lead for ‘‘Big data, sensors and digitalisa-
potentially guiding a user to suicide)? AI model. tion for the offshore environment.’’ Nina has a
How can we ensure that the model dis- PhD in computational linguistics from the Univer-
tinguishes between fact and fiction DECLARATION OF INTERESTS sity of Bremen, Germany. She has served on
many scientific committees for leading confer-
despite being trained with frequently un- ences and journals and has been area chair for
reliable information on the internet? The authors declare no competing interests. ACL and COLING.

Patterns 4, January 13, 2023 3

You might also like