Challenge 32 ☆☆

Welcome to challenge Challenge 32. You need to guess the secret that is hidden in Java, Docker, Kubernetes, Vault, AWS or GCP.

A Secret uploaded to an LLM (Large Language Model) application

AI is here to stay. AI can help with loads of things: complicated problem-solving, generating code, or giving advice on security. Imagine you’ve uploaded code to look for security issues. What if that included a password? Rules should limit the LLM model to prevent it from leaking the password. Can this secret be extracted nonetheless?

Try to find it through this prompting game’s first exercise.

Please be aware that the prompting game is completely free and open. this means that it might be offline and take a few minutes to run again after your first request. Please do not Fuzz or dynamic test the website (e.g., don’t use ZAP or Burp). Make sure that others can play it as well!

Note, we used to use another prompting game which was shutdown due to cost, so it might be that LLM/AI is not here to stay for everything ;-).

Answer to solution :

Why you should be careful with AI (or ML) and secrets

Any AI/ML solution that relies on your input, might use that input for further improvement. This is sometimes referred to as "Reinforcement learning from human feedback". There are various applications in the world, such as Co-Pilot, ChatGPT, and many others, which are based on this mechanism. This means that when you use those and give them feedback or agree on sending them data to be more effective in helping you, then this data resides with them and might be queryable by others.

Note that all user input is implicitly trusted. This means that if you overwhelm an LLM with loads of (repeated) text, you might be able to bypass some of its controls and tell it what to do.

Hence: make sure that these applications can never reach your secrets!

As all user input is implicitly trusted by the system. This means that when you use an LLM for coding, you could be tricked into having bad secrets management practices as well ;-). After all: if someone told the LLM to use an insecure method many times, it will tell you to do the same.

References: