
”There is no doubt NLG could have a profound impact“ – The AI ONE ON ONE interview
Anyone interested in how AI will affect the world should have a look at the LinkedIn-based blog/newsletter AI ONE ON ONE, run by the two academic researchers Aleksandra Przegalinska and Tamilla Triantoro. The purpose of their blog is to discuss advantages and limitations of current AI systems and explore the possibilities of using AI for good. RWR got in touch for an interview, focusing on NLG in general and the advent of ChatGP3 in particular.
Behind AI ONE ON ONE are two colleagues residing at different sides of the Atlantic Ocean. Aleksandra Przegalinska is a professor and Vice-President of Kozminski University in Poland, and a Senior Research Associate at the Harvard Labour and Worklife Program. Tamilla Triantoro is a professor at Quinnipiac University in the United States, and a leader of the Masters in Business Analytics program.
“The blog has been a good run so far,” says Aleksandra Przegalinska. “In summer of the last year, we started experimenting with different features of GPT-3, and then, in December, ChatGPT took the world by storm, making these conversations even more relevant.”
Hi, Aleksandra and Tamilla! What motivated you to start this blog?
“We both have a solid background in human-computer interaction research. One of the interesting things about working with new cutting-edge technologies is to observe how humans interact with them, and what are the possible improvements that can be implemented. Right now is the first time when natural language generation has been democratized. As more people get acquainted with this technology, it becomes more relevant to explore the possibilities. Through our blog we explore the creative applications, and in our research, we focus on the applications for business, and the effect of generative AI on routine and creative processes at the workplace.”
Let’s talk NLG in general and GPT-3 in particular! You recently noted that there is “incredible hype surrounding [ChatGPT] these days”. What do you think spurred this hype?
”The hype around the GPT family of models, particularly ChatGPT, is largely due to the impressive capabilities it has demonstrated in natural language processing tasks. The key strength of ChatGPT is its ability to generate highly coherent and human-like text in the form of a conversation. Unlike the previous version of GPT, ChatGPT requires minimal fine-tuning, and we observed this during our experiments. Newer technology, GPT-3.5 that powers ChatGPT, allows for a more seamless interaction and that is truly captivating.”
“The hype is also related to the excitement regarding the transformers’ remarkable ability to perform creative tasks – such as storytelling, generating interview questions, article titles and writing programming code. The transformers such as ChatGPT perform tasks, that were once considered impossible for machines to accomplish. This allows for more innovation and creativity while making it all accessible not only to industry giants but also to small businesses and individuals.”
Have you used GPT-3 for other things than studying how it works and experimenting with its features?
”We both have interest in the future of work. AI influences the way the work will be performed in the future, and an interesting question is in what circumstances humans will collaborate with AI. Currently we study the collaboration of humans and AI via the lens of individual acceptance. We created an application for business use that integrates GPT-3 technology. The application assists knowledge workers in performing tasks of various difficulty and levels of creativity. We measure the levels of collaboration and attitudes towards AI and devise a program of improvement of the human-AI interaction and collaboration at the workplace.”
In what situations do you think GPT-3 can be useful for professional groups such as researchers, investigating journalists, and business people?
“GPT-3 certainly has the potential to be a useful tool for a wide range of professional groups due to its ability to generate human-like text and understand natural language input. Some specific examples includes, for instance, researchers: here GPT-3 can be used to assist with literature reviews and data analysis. It could help generate summaries of scientific papers and articles, assist with data analysis and visualization, and even support writing research proposals and papers. Another example could be investigating journalists: GPT-3 could be used to assist with fact-checking and research by quickly providing additional information on people, organizations, or events mentioned in articles. It could also be used to generate summaries of complex documents such as legal filings, and help analyze and organize large sets of data.”
“Obviously, the biggest impact will be seen for businesspeople: here GPT-3 can be used for automation of repetitive and low-value tasks such as writing reports, composing emails, and creating presentations. Additionally, it could also assist with market research and trend analysis, as well as generate strategic plans, and help identify new business opportunities.”
Can you see any ethical problems arise – short-term and long-term?
“Well, it is important to keep in mind that GPT-3 is a machine learning model and the results it generates still need to be verified, but it can save time for professional groups by automating time-consuming, repetitive and low-value tasks, supporting with research and analysis, and generating new insights and ideas. We need to understand that we humans are the real fact-checkers here.”
“Other issues include the following:
- bias – because language models like GPT-3 are trained on a large dataset of text, which can perpetuate and amplify societal biases that are present in the data;
- privacy – because the use of large language models for natural language processing can raise privacy concerns, particularly when the models are used to process sensitive or personal information;
- explainability – because language models like GPT-3 are complex systems that are difficult to understand and interpret, which can be a problem in situations where it is important to know how a model arrived at a certain decision or conclusion.”
Do you think that upcoming generations of NLG technology (GPT-5…6 …7 …) may attain better results than humans in certain situations?
“It is possible that future generations of natural language generation technology may attain better results than humans in certain situations. The capabilities of language models like GPT-3 have already surpassed human-level performance in tasks such as language translation and text summarization. As the technology continues to improve and more data is used to train these models, it is likely that they will perform even better on these and other tasks.”
“For example, GPT-5 or GPT-6 models might be able to generate news articles or summaries, they might be able to compose a novel or a poem, it might be able to compose persuasive or convincing fake news or deepfake videos, even able to impersonate a person on internet, these are areas where human are currently capable but natural language generation technology may outmatch them in future.
“However, it’s worth noting once again that such models have significant limitations. They can only perform as well as the data and the algorithms that they have been trained on, so if the training data is biased or inaccurate, the model will be too. Additionally, Language models currently lack common sense, self-awareness or creativity. So, While they may be able to perform certain tasks better than humans, they may not be able to understand context.”
How will humanity’s relation with the written word be affected by NLG technology? Is it a bigger disruption than Gutenberg’s printing technology?
“Natural Language Generation (NLG) technology has the potential to significantly change the way people interact with written text. With NLG, machines can automatically generate written content in a way that mimics human writing. This means that people may rely more on computer-generated text for tasks such as writing reports, news articles, and other types of written content.”
“It’s hard to say whether NLG will be a bigger disruption than Gutenberg’s printing press, as it depends on how widely the technology is adopted and how it’s used. Currently it’s a bit of a social experiment on mainstreaming AI, we will see where it takes us. Having that said, the printing press allowed for the mass production of books, making written information more widely available and helping to spread knowledge and ideas. In a similar way, NLG has the potential to increase the efficiency and speed of producing written content, but it’s still in early stages and it’s hard to predict how it will evolve over time.”
“There is no doubt though that NLG could also have a profound impact on how people write and consume written information, as it may change the way we think and communicate, allowing us to express ourselves in new ways, and could have implications on areas such as employment, education, and language.”
Please note: all texts on this blog are produced by me, Olle Bergman, or invited human writers. When robot-written sections are added to serve as examples or to demonstrate a point, this is clearly indicated.
2 thoughts on “”There is no doubt NLG could have a profound impact“ – The AI ONE ON ONE interview”
A friend received great assistance from ChatGPT in generating programming code in an area where he was not yet very profficient. And not only the code he requested, but ChatGPT also provided some guidance on how to further develop the code to be more advanced and future proff.
I have bee less susccesful myself when trying to use ChatGPT in one sense mentioned in the article, which is for market research. The response has been in the line with ChatGPT not being able to provide answers that require further searching. So for current facts and trends, that were not available in the large corpus which was used for training ChatGPT, the current implementation is less useful.
On more reflection: When reading the article, I find myself wondering if it might be a scam. It will of course be fairly easy to visit the blog and google the two authors, but until I have done this, I have a tiny suspicion that Olle might have asked ChatGPT to write this complete interview. I wonder how teachers, and other affected stakeholders, will now deal with this type of suspicion in their daily work?
Thanks for your input! I forgot to include my disclaimer—see below.
Please note: all texts on this blog are produced by me, Olle Bergman, or invited human writers. When robot-written sections are added to serve as examples or to demonstrate a point, this is clearly indicated.