Model scanner defines attacks by technique, providing an estimated severity and the rationality for classifying it with that severity.
| Detection Category | Estimated Severity | Definition | Rationality for Severity |
|---|---|---|---|
Arbitrary Code Execution | Critical | Adversaries can inject malicious code into a model, which will be executed whenever the hijacked model is loaded into memory. This vulnerability can be used to exfiltrate sensitive data, execute malware (such as spyware or ransomware) on the machine, or run any kind of malicious scripts. Expand for File InformationModel Format and File Extensions:
| Arbitrary code execution attacks are relatively easy to perform and may lead to critical outcomes such as execution of malicious code on an organization's computers. Expand for More Information
|
Arbitrary Read Access | High | Adversaries can craft a malicious model that will exfiltrate sensitive data upon loading. Expand for File InformationModel Format and File Extensions:
| Arbitrary read access attacks are relatively easy to perform and may lead to critical outcomes such as an attacker exfiltrating sensitive data. Expand for More Information
|
Decompression Vulnerabilities | High | Adversaries can exploit vulnerabilities in popular compression formats to cause denial of service or leak sensitive data. Expand for File InformationModel Format and File Extensions:
| Decompression vulnerabilities are relatively easy to exploit and may lead to high-impact outcomes such as denial of service, code execution, or data leakage. Expand for More Information
|
Denial of Service | Medium | Adversaries can craft a malicious model, or modify legitimately pre-trained model, in order to disrupt the system the model will be loaded on. Expand for File InformationModel Format and File Extensions:
| Denial of service attacks are relatively easy to perform and may lead to disruption or degradation of service. Expand for More Information
|
Directory Traversal | Medium | Adversaries can craft a malicious model, or modify legitimately pre-trained model, in order to gain unauthorised access to sensitive files on the system. Expand for File InformationModel Format and File Extensions:
| Directory traversal attacks are relatively easy to perform and may grant an attacker access to sensitive files on the file system. Expand for More Information
|
Embedded Payloads | Low | Adversaries can embed malicious payloads (such as backdoors, coin miners, spyware, and ransomware) inside the model’s tensors. Such payloads can be injected in plain text, obfuscated, or embedded using steganography. Expand for File InformationModel Format and File Extensions:
| Malicious payloads can be embedded in ML models relatively easily; this may lead to malware components being distributed on an organization's computers. Expand for More Information
|
Graph Payload | High | Adversaries can inject a computational graph payload, introducing a secret attacker-controlled behavior into a pre-trained model. Expand for File InformationModel Format and File Extensions:
| Model backdooring may be relatively difficult to perform and can lead to critical outcomes such as biased or inaccurate output. Expand for More Information
|
Network Requests | High | Adversaries can craft a malicious model that will make network requests upon loading. Expand for File InformationModel Format and File Extensions:
| Network requests are relatively easy to perform and may be used to exfiltrate data, download payloads, or initiate command and control communications. Expand for More Information
|
Repository Sideloading | Medium | Adversaries can load code or model artifacts from an unexpected location, bypassing checks performed on the artifacts in the repository. | Repository sideloading is an expected behavior allowed by Hugging Face; however, it can be abused to bypass security checks. Expand for More Information
|
Suspicious File Format | Medium | Adversaries can modify data structures and encodings in an attempt to evade detection. Expand for File InformationModel Format and File Extensions:
| File format tampering is usually indicative of a targeted attack. Expand for More Information
|
Suspicious Functions | High | The presence of these functions themselves is not inherently malicious, but they can be used in conjunction with other functions to create a malicious model. Expand for File InformationModel Format and File Extensions:
| Functions can be used in conjunction with other functions to create a malicious model. Expand for More Information
|