Prompt Injection: When Hackers Befriend Your AI - Vetle Hjelle - NDC Security 2024

Sdílet
Vložit
  • čas přidán 26. 03. 2024
  • This talk was recorded at NDC Security in Oslo, Norway. #ndcsecurity #ndcconferences #security #developer #softwaredeveloper
    Attend the next NDC conference near you:
    ndcconferences.com
    ndc-security.com/
    Subscribe to our CZcams channel and learn every day:
    /@NDC
    Follow our Social Media!
    / ndcconferences
    / ndc_conferences
    / ndc_conferences
    This is a technical presentation where we'll look at attacks on implementations of Large Language Models (LLMs) used for chatbots, sentiment analysis, and similar applications. Serious prompt injection vulnerabilities can be used by adversaries to completely weaponize your AI against your users.
    We will look at how so-called "prompt injection" attacks occur, why they work, different variations like direct and indirect injections, and then see if we can find good solutions on how to mitigate those risks. We'll also learn how LLMs are "jailbroken" to ignore their alignment and produce dangerous content.
    LLMs are not brand new, but we know that their use will increase drastically in the next few years, and therefore it is important to take security seriously by considering the risks involved before using AI for sensitive operations.
  • Věda a technologie

Komentáře • 6

  • @monad_tcp
    @monad_tcp Před 2 měsíci

    46:34 doesn't fully work, well it worked for OpenAI itself, but they have much more resources than everyone else

  • @Roibarkan
    @Roibarkan Před 2 měsíci +1

    17:23 An explainer about word embeddings: czcams.com/video/gQddtTdmG_8/video.html

    • @goldnutter412
      @goldnutter412 Před 2 měsíci

      Yeah great video
      en.wikipedia.org/wiki/Stochastic_parrot though 🤣 whoever came up with this is a legend🥰

  • @goldnutter412
    @goldnutter412 Před 2 měsíci

    16:16
    czcams.com/video/klGWvRsLQKU/video.html 🤣