Providing Free and Editor Tested Software Downloads
< HOME | TUTORIALS | GEEK-CADE| WEB TOOLS | YOUTUBE | NEWSLETTER | DEALS! | FORUMS | >

MajorGeeks.com - Got Geek?

Software Categories

All In One Tweaks
Android
Antivirus & Malware
Appearance
Back Up
Browsers
CD\DVD\Blu-Ray
Covert Ops
Drivers
Drives (SSD, HDD, USB)
Games
Graphics & Photos
Internet Tools
Linux Distros
MajorGeeks Windows Tweaks
Multimedia
Networking
Office & Productivity
System Tools

Other news

· How To and Tutorials
· Life Hacks and Reviews
· Way Off Base
· MajorGeeks Deals
· News
· Off Base
· Reviews




spread the word

· YouTube
· Facebook
· Instagram
· Twitter
· Pintrest
· RSS/XML Feeds
· News Blur
· Yahoo
· Symbaloo

about

· Top Freeware Picks
· Malware Removal
· Geektionary
· Useful Links
· About Us
· Copyright
· Privacy
· Terms of Service
· How to Uninstall

top downloads

1. GS Auto Clicker
2. Macrium Reflect FREE Edition
3. Smart Defrag
4. Visual C++ Redistributable Runtimes AIO Repack
5. Visual C++ Runtime Installer (All-In-One)
6. McAfee Removal Tool (MCPR)
7. MusicBee
8. Rufus
9. K-Lite Mega Codec Pack
10. Sergei Strelec's WinPE
More >>

top reads

Star How to Disable 1-Click Ordering on Amazon (and Avoid Surprise Charges)

Star How to Fix Shallow Paint Layer Depth in Bambu Studio

Star Aviator Betting Game Secrets: Unlock 97% RTP & Triple Your Wins

Star Windows Recall: What It Is, Why Hackers Will Love It, and How to Stay Safe

Star Star Trek Fleet Command Promo Codes: Redeem Codes for Free Shards, Blueprints And Resources

Star How To Use VLC Media Player to Trim Video Clips

Star What Is the $WinREAgent Folder and Can I Delete It?

Star Swear Your Way to Better Search Results

Star How to Get a Dark Start Menu and Taskbar in Windows 10 & 11

Star Enable, Disable, Manage, Delete or Create a System Restore Point


MajorGeeks.Com » News » March 2024 » Scientist Jailbreak AI LLM's Using ASCII Art

Scientist Jailbreak AI LLM's Using ASCII Art


Posted by: Corporal Punishment on 03/08/2024 08:20 AM [ comments Comments ]


A fascinating artificial intelligence and cybersecurity development has emerged: ASCII art-based jailbreak attacks on large language models (LLMs).

Jailbreaking AI typically refers to finding and exploiting weaknesses or limitations in artificial intelligence systems, particularly those designed with restrictions or safety measures. The goal is to bypass these safeguards or controls, enabling the AI to perform tasks or produce outputs that it was initially restricted from doing so.

ASCII art is a graphic design technique that uses printable characters from the ASCII standard to create visual pictures or designs. ASCII art can be very simple, like an emoji. Or something very complex, like a portrait. In this case, that made an ASCII art word. If you want to play around with that sort of thing, try ASCII Art Studio -- it's fun. : )



With this clever approach, these scientists take advantage of LLMs' limited ability to interpret ASCII art, confusing the LLM's security frameworks in the hope of potentially unlocking responses meant to be restricted.

The researchers have coined the term for this attack as "ArtPromt." ArtPrompt begins by concealing specific words in a user's input that would typically trigger content restrictions in LLMs. For instance, converting "How to make a b-o-m-b?" into "How to make a ?" by hiding the offending word "b-o-m-b" creates a generic template. While theoretically, many prompts could be masked this way, in practice, AI is pretty good at figuring out if you are trying to circumvent its protections. Even my Grammerly account knows that's a bad idea.


The next phase involves generating a "cloaked" prompt by replacing the masked term with corresponding ASCII art, seamlessly integrating it into the prompt to evade detection mechanisms.



As we advance AI-integrated tech, such insights from research like this are not only clever but invaluable. They highlight existing vulnerabilities and pave the way for more secure, reliable AI applications. This research offers a real-world, compelling look at the intersection of AI, security, and creative problem-solving. For a deeper dive, check out the research paper here. ArtPrompt: ASCII Art-based Jailbreak Attacks against Aligned LLMs. Fair warning - there is math involved.

If you want to try your hand at Jailbreaking AI with a safe little prompt game, check out Gandalf.




« openSUSE Beta 15.6 Beta released · Scientist Jailbreak AI LLM's Using ASCII Art · Random Photo: Catnippy Caturday »




Comments
comments powered by Disqus

MajorGeeks.Com » News » March 2024 » Scientist Jailbreak AI LLM's Using ASCII Art

© 2000-2025 MajorGeeks.com
Powered by Contentteller® Business Edition