|
As enterprises rapidly adopt AI agents (e.g., Salesforce’s Agentforce), a critical risk emerges: misconfigured or compromised agents performing anomalous, potentially harmful, data operations. Millie unveils an original, practical methodology for detecting such threats using unsupervised machine learning.
Drawing from a real-world Proof-of-Concept, Millie demonstrates how behavioral profiling—analyzing features engineered from system logs like data access patterns, query syntax (SOQL keyword analysis), and IP usage, along with signals from the content moderation mechanisms embedded within the LLM guardrails such prompt injection detection and toxicity scoring—can distinguish risky agent actions. Explore the creation of 30+ behavioral features and the application of KMeans clustering to identify agents exhibiting statistically significant deviations, serving as an early warning for misuse or overpermissive configurations. Millie will share insights into observed differences between AI agent and human user profiles, and challenges like crucial data gaps that impact comprehensive monitoring.
More info here.
|