Consistent Jailbreaks in GPT-4, o1, and o3 - General Analysis - eviltoast
  • Umbrias@beehaw.org
    link
    fedilink
    arrow-up
    4
    ·
    14 days ago

    jailbreaks actually are relevant with the use of llm for anything with i/o, such as “automated administrative assistants”. hide jailbreaks in a webpage and you have a lot of vectors for malware or social engineering, broadly hacking. as well as things like extracting controlled information.