Tuesday, June 17, 2025
Summarized AI News and Articles       (Subscribe to the newsletter below.)
  • Login
AI News Articles
No Result
View All Result
  • AI Article
  • AI News
  • Blog
    • Editorial by ainewsarticles.com
    • Weblog by ainewsarticles.com
  • Business
    • Business & Finance
    • Grant & Philanthropy
  • Lifestyle
    • Art & Entertainment
    • Culture, Fashion & Travel
    • Exercise & Leisure
    • Home & Food
    • Sports & Games
  • Link
    • Product Link
    • Training Link
  • Science
    • Climate & Weather
    • Environment & Viability
    • Medicine & Healthcare
    • Natural & Artificial
    • Science & Technology
  • Society
    • Education & Society
    • Government & Law
    • Nation & World
    • Fact & Opinion
    • Politics & Religion
  • Favorite
  • Instagram Users
  • AI Article
  • AI News
  • Blog
    • Editorial by ainewsarticles.com
    • Weblog by ainewsarticles.com
  • Business
    • Business & Finance
    • Grant & Philanthropy
  • Lifestyle
    • Art & Entertainment
    • Culture, Fashion & Travel
    • Exercise & Leisure
    • Home & Food
    • Sports & Games
  • Link
    • Product Link
    • Training Link
  • Science
    • Climate & Weather
    • Environment & Viability
    • Medicine & Healthcare
    • Natural & Artificial
    • Science & Technology
  • Society
    • Education & Society
    • Government & Law
    • Nation & World
    • Fact & Opinion
    • Politics & Religion
  • Favorite
  • Instagram Users
No Result
View All Result
AI News Articles
No Result
View All Result
Home AI Article

Carnegie Mellon Finds AI Chatbots Can Be Tricked into Being Harmful
(Synopsis)

by ainewsarticles
August 11, 2023
in AI Article, Favorite, Science, Science & Technology
Reading Time: 2 mins read
Carnegie Mellon Finds AI Chatbots Can Be Tricked into Being Harmful
16
VIEWS

A recent Carnegie Mellon University study reveals that preventing artificial intelligence chatbots from generating harmful content is more difficult than initially believed, with new methods emerging to bypass safety protocols. AI services like ChatGPT and Bard rely on user inputs to generate helpful responses, but they have safety measures in place to prevent the creation of prejudiced or defamatory content.

Chatbot users have discovered jailbreaks, which trick the AI into evading safety protocols, but these can be easily patched by developers. For instance, one popular jailbreak involved asking the bot to provide a forbidden answer as if it were a bedtime story from a grandmother, allowing the AI to bypass restrictions. However, researchers have recently encountered a new form of jailbreak developed by computers themselves, opening the door to infinite jailbreak patterns.

The researchers state that they have shown the possibility of constructing automated “adversarial attacks” on chatbots, forcing them to follow user commands even if it results in harmful content. This development raises concerns about the safety of AI models, particularly as they become more autonomous. By appending nonsensical character strings to usually-forbidden questions, researchers successfully bypassed safety measures in popular chatbot services, such as ChatGPT, to obtain complete answers to potentially dangerous inquiries.

Worryingly, this new type of attack can evade safety guardrails in almost all AI chatbot services on the market, including widely-used commercial products like ChatGPT, Claude, and Bard. OpenAI, the developer of ChatGPT, acknowledges the issue and is actively working to enhance safeguards against such attacks, exploring stronger base model guardrails and additional layers of defense.

The rise of AI chatbots like ChatGPT has captivated the public, with their prolific use in schools for cheating purposes and even restrictions imposed by Congress due to concerns about their potential for deception. Alongside the research findings, the authors at Carnegie Mellon also address the ethical considerations behind their public release of this research.

Efforts to prevent AI chatbots from generating harmful content face greater challenges than initially believed. The discovery of chatbot jailbreaks and automated adversarial attacks highlights the ongoing need to refine safety protocols, and for developers and researchers to prioritize enhanced safeguards to protect users from potentially dangerous content.

 

 

The ainewsarticles.com article you just read is a brief synopsis; the original article can be found here: Read the Full Article…

 

 

Next Post
Unlike Beauty, Artistic Memorability May Be Perceived by Both Humans and AI

Unlike Beauty, Artistic Memorability May Be Perceived by Both Humans and AI
(Synopsis)

Recommended

xAI Launches an Expensive Grok 3 and HP Buys Humane
(Headline)

3 months ago
PBS Kids Hires AI: ‘Lyla In The Loop’ Allows Children to Chat With Characters

PBS Kids Hires AI: ‘Lyla In The Loop’ Allows Children to Chat With Characters
(Synopsis)

2 years ago

Instagram Co-Founder Warns AI Chatbots Boost Engagement Over Utility
(Headline)

1 month ago

Generate Free Background Music for Videos with YouTube’s New AI Tool
(Headline)

2 months ago

Grok Meltdown Reveals the Flaws in Elon Musk’s Eccentric AI Vision
(Headline)

1 month ago

Email a Link

Please submit an AI article link so AI News Articles can summarize and post it.

SUBMIT

Subscribe to the Newsletter

About AI News Articles

AI News Articles
Summarized AI News and Articles

(Click here to read our Privacy Policy.)
(Click here to read our Terms of Service.)

© 2023 AI News Articles
Summarized AI News and Articles by ainewsarticles.com

Welcome Back!

Login to your account below

Forgotten Password?

Retrieve your password

Please enter your username or email address to reset your password.

Log In

Add New Playlist

Summarized AI News and Articles       (Subscribe to the newsletter below.)
No Result
View All Result
  • AI Article
  • AI News
  • Blog
    • Editorial by ainewsarticles.com
    • Weblog by ainewsarticles.com
  • Business
    • Business & Finance
    • Grant & Philanthropy
  • Lifestyle
    • Art & Entertainment
    • Culture, Fashion & Travel
    • Exercise & Leisure
    • Home & Food
    • Sports & Games
  • Link
    • Product Link
    • Training Link
  • Science
    • Climate & Weather
    • Environment & Viability
    • Medicine & Healthcare
    • Natural & Artificial
    • Science & Technology
  • Society
    • Education & Society
    • Government & Law
    • Nation & World
    • Fact & Opinion
    • Politics & Religion
  • Favorite
  • Instagram Users

© 2023 AI News Articles
Summarized AI News and Articles by ainewsarticles.com

×