Side-channel risks in model inference refer to vulnerabilities where attackers extract sensitive information by analyzing indirect data, such as the time taken for a model to process inputs (timing attacks) or observing output patterns (token leakage). These risks can reveal private details about the model’s structure, training data, or user queries, potentially compromising confidentiality and privacy even if direct access to the model’s internals or data is restricted.
Side-channel risks in model inference refer to vulnerabilities where attackers extract sensitive information by analyzing indirect data, such as the time taken for a model to process inputs (timing attacks) or observing output patterns (token leakage). These risks can reveal private details about the model’s structure, training data, or user queries, potentially compromising confidentiality and privacy even if direct access to the model’s internals or data is restricted.
What are side-channel risks in model inference?
Side-channel risks occur when attackers infer sensitive information from indirect signals, such as how long inference takes or patterns in produced tokens, rather than the direct outputs alone.
What is a timing attack in this context?
A timing attack analyzes the time a model takes to process inputs to deduce hidden details about the inputs, model internals, or data, exploiting data-dependent timing variations.
What is token leakage in generative AI?
Token leakage is when generated tokens or output patterns reveal private information from prompts, inputs, or training data, potentially exposing sensitive details to an observer.
How can these inference-side risks be mitigated?
Mitigations include constant-time or data-independent inference, reducing data-dependent branches, adding noise, rate limiting and monitoring, prompt filtering, data minimization, differential privacy, secure enclaves, and strict access controls.