[IT.T.12] A Man-in-the-Middle (MitM) attack in the context of machine learning is a type of adversarial attack where an attacker stealthily intercepts and alters the communication between two parties (e.g., a data source and a machine learning classifier) to deliver malicious payloads or manipulate the data, with the aim of compromising the integrity or availability of the ML system.
System Asset: Machine learning model.
Business Asset: input data.
Security Criteria: integrity.
Vulnerabilities:
Threat agent: white-box and black-box scenarios. In the white-box scenario, the attacker is assumed to have complete knowledge of the target machine learning model, its architecture, parameters, utilized training data, and the learning algorithm. In a black-box scenario, the attacker has no knowledge of the target model's architecture, parameters, or training data. The attacker is assumed to be only able to interact with the model by sending it inputs and observing the outputs.
Attack methods:
Impact and harm: Negates the integrity of the targeted machine learning model. This leads to misclassification of benign and malicious inputs.
Security requirement: The machine learning system must be resistant to adversarial attacks.
Security controls: