Project Roma: Deterministic security for AI agents
AI agents perform consequential actions while processing data from various sources, including trusted collaborators and the public Web. It is crucial that AI agents handle this data with care: confidential data must be adequately protected, and untrusted data must not derail the agent’s behavior. However, AI agents rely on models that can behave unpredictably and are susceptible to manipulation. This makes them vulnerable to attacks such as indirect prompt injection attacks, which can steal and corrupt data.
Project Roma aims to create a system-level, deterministic protective layer that hardens AI agents, providing strong security and confidentiality guarantees even when the models they use misbehave.