What can we expect from AI and Chatbots in the next few years?

March 31, 2023

On March 15, Newswise hosted an expert panel on how artificial intelligence and chatbots are changing the landscape of journalism and the transfer of knowledge (watch the video and read the transcript here). Panelists included Sercan Ozcan, Associate Professor of Innovation & Technology Management at the University of Portsmouth, Jim Samuel, Associate Professor of Public Informatics at Rutgers University-New Brunswick, and Alan Dennis, Professor of Information Systems at Indiana University. We learned that there are exciting things from AI tech that can assist us as science writers and communicators. How awesome is it to have a program summarize a study you might be struggling to get into coherent words with just a few carefully worded commands? ChatGPT can help with some of the routine work of a journalist and a science communicator, searching for information, gathering information, and possibly even putting that information into a first draft. But with that benefit come significant challenges. The biggest challenge is sussing out the bullshit (bullshit is a technical term according to our panelist Alan Dennis. Honest!). Artificial intelligence in the form of large language models (LLMs) such as ChatGPT gives information to you that looks very realistic, as if a real person wrote it. But this is an illusion. Sercan Ozcan refers to the deceiving output of chatbots such as ChatGPT as “hallucinations.”

The creation of misinformation is what worries Alan Dennis the most. “Deep fakes (artificial videos of real people) and other tools like it is going to change everything, particularly for journalism because we’ve created digital puppets of several different celebrities and I can make them say anything that I want them to say.”

Deep fakes are one thing, but what about the dangers of media relying on AI to generate news content? Panelist Jim Samuel says that “we need to treat AIs as some kind of very smart, but inexperienced and probably not very, not comprehensively knowledgeable teenager.” The output that AI produces requires supervision. Samuel says that we have a responsibility [as educators, media, and science communicators] to educate the public in order to develop internal mechanisms to deal with misinformation.

News Wise, March 31, 2023

Recent Posts

Christiana Foglio, DC’84, BSPPP’86 Named RAA Loyal Daughter

The Rutgers Alumni Association’s Loyal Sons & Daughters Award is its highest recognition of service. Recipients are individuals who have made a meaningful and long-standing contribution to the betterment of Rutgers by performing extraordinary volunteer service or...

Lindenfeld Investigates LFO Impacts on Health Outcomes

Legal Financial Obligations: An Understudied Public Health Exposure Abstract The impacts of exposure to the criminal justice system on health-related outcomes are well studied in the United States (US). However, while previous studies focus on the impacts of arrest,...

EJB Talks: Beyond “Does It Work?”

Beyond “Does It Work?”: Laura Peck on Policy, Evidence, and Impact EJB Talks returns for Season 14 with Dean Stuart Shapiro speaking with Laura Peck, one of our newest Public Policy Associate Professors and a Principal Faculty Fellow with the Heldrich Center for...

Heldrich Center: Motivational Texts and Unemployment

Original post from the Daily Targum By Akash Nattamai Researchers at the John J. Heldrich Center for Workforce Development recently published a report regarding the effectiveness of motivational text messaging on reintroducing people in the statewide Reemployment...