We do a lot of processing on our backend to prevent against prompt injection, but there definitely still is some risk. We can do better on as is always the case.
Need to read up on how CaMel does it. Do you have any good links?
Here’s a paper offering a survey of different mitigation techniques, including CaMeL. Design Patterns for Securing LLM Agents against Prompt Injections (2025):
https://arxiv.org/abs/2506.08837
Need to read up on how CaMel does it. Do you have any good links?