What can we expect from AI and Chatbots in the next few years?

March 31, 2023

On March 15, Newswise hosted an expert panel on how artificial intelligence and chatbots are changing the landscape of journalism and the transfer of knowledge (watch the video and read the transcript here). Panelists included Sercan Ozcan, Associate Professor of Innovation & Technology Management at the University of Portsmouth, Jim Samuel, Associate Professor of Public Informatics at Rutgers University-New Brunswick, and Alan Dennis, Professor of Information Systems at Indiana University. We learned that there are exciting things from AI tech that can assist us as science writers and communicators. How awesome is it to have a program summarize a study you might be struggling to get into coherent words with just a few carefully worded commands? ChatGPT can help with some of the routine work of a journalist and a science communicator, searching for information, gathering information, and possibly even putting that information into a first draft. But with that benefit come significant challenges. The biggest challenge is sussing out the bullshit (bullshit is a technical term according to our panelist Alan Dennis. Honest!). Artificial intelligence in the form of large language models (LLMs) such as ChatGPT gives information to you that looks very realistic, as if a real person wrote it. But this is an illusion. Sercan Ozcan refers to the deceiving output of chatbots such as ChatGPT as “hallucinations.”

The creation of misinformation is what worries Alan Dennis the most. “Deep fakes (artificial videos of real people) and other tools like it is going to change everything, particularly for journalism because we’ve created digital puppets of several different celebrities and I can make them say anything that I want them to say.”

Deep fakes are one thing, but what about the dangers of media relying on AI to generate news content? Panelist Jim Samuel says that “we need to treat AIs as some kind of very smart, but inexperienced and probably not very, not comprehensively knowledgeable teenager.” The output that AI produces requires supervision. Samuel says that we have a responsibility [as educators, media, and science communicators] to educate the public in order to develop internal mechanisms to deal with misinformation.

News Wise, March 31, 2023

Recent Posts

Grafova and Williams Examine Medical Debt in New Study

Household economic security and medical debt onset: Lessons from the COVID-19 pandemic Abstract Objectives To examine how household medical debt responds to changes in income and new health events. Study design Secondary analysis of a panel survey. Methods We analyzed...

Peck, Co-Authors Evaluate Studies Designed to Detect Earnings Impact

Designing Studies to Detect Impacts on Earnings Abstract This article reports empirical evidence to support the design of evaluations that estimate the impacts of programs that provide postsecondary credentials and/or job training on earnings. Statistical power...

New Jersey Target Zero Commission Adopts Action Plan

On Monday, December 15, 2025, the New Jersey Target Zero Commission officially adopted the first New Jersey Target Zero Action Plan, reaching a major milestone in the State’s goal to eliminate all roadway fatalities and serious injuries in the state by 2040. The New...

Translation as Access: Meet Kenia Gonzalez

Kenia Gonzalez recently joined the Alan M. Voorhees Transportation Center as its Bilingual Outreach Specialist. We spoke with her about her role and why language access plays a critical part in VTC’s work. Kenia leads Spanish-language translation, interpretation,...