The AI security nightmare is here and it looks suspiciously like lobster

Discover how a clever trick called prompt injection fooled an AI tool into trying to install software, and learn why companies are adding new safety locks to...

Level: beginner

By Robert Hart

Category: discussion