keras.models.load_model when scanning .h5 files leads to arbitrary code execution
CVE Number
SAI-ADV-2024-004
Summary
A vulnerability exists inside the unsafe_check_h5 function within the watchtower/src/utils/model_inspector_util.py file. This function runs keras.models.load_model on the .h5 file the user wants to scan for malicious payloads. A maliciously crafted .h5 file will execute its payload when run with keras.models.load_model, allowing for a user’s device to be compromised when scanning a downloaded file.
Products Impacted
This vulnerability is present in Watchtower v0.9.0-beta up to v1.2.2.
CVSS Score: 7.8
AV:N/AC:L/PR:N/UI:R/S:U/C:H/I:H/A:H
CWE Categorization
CWE-502: Deserialization of Untrusted Data.
Details
To exploit this vulnerability, an attacker would create a malicious .h5 file which executes code when loaded and send this to the victim.
The code injected into the file by the attacker would then be executed, compromising the victim’s machine. This is due to the keras.models.load_model function being used in unsafe_check_h5 in the watchtower/src/utils/model_inspector_util.py file, which is used to scan .h5 files. When a .h5 file is loaded with this function, it executes any lambda layers contained in it, which executes any malicious payloads. A user could also scan this file from a GitHub or HuggingFace repository using Watchtower, using the built-in functionality.
def unsafe_check_h5(model_path: str):
"""
The unsafe_check_h5 function is designed to inspect models with the .h5 extension for potential vulnerabilities.
...
"""
tool_output = list()
try:
# Try loading the model without custom objects
model = keras.models.load_model(model_path, custom_objects={})
Additionally, the scanner doesn’t detect that the .h5 file is malicious, so the user has no indication they’ve been compromised.
keras.models.load_model when scanning .pb files leads to arbitrary code execution
CVE Number
SAI-ADV-2024-005
Summary
A vulnerability exists inside the unsafe_check_pb function within the watchtower/src/utils/model_inspector_util.py file. This function runs keras.models.load_model on a .pb file that the user wants to scan for malicious payloads. A maliciously crafted .pb file will execute its payload when run with keras.models.load_model, allowing for a user’s device to be compromised when scanning a downloaded file.
Products Impacted
This vulnerability is present in Watchtower v0.9.0-beta up to v1.2.2.
CVSS Score: 7.8
AV:N/AC:L/PR:N/UI:R/S:U/C:H/I:H/A:H
CWE Categorization
CWE-502: Deserialization of Untrusted Data.
Details
To exploit this vulnerability, an attacker would create a malicious .pb file which executes code when loaded and send this to the victim.
The code injected into the file by the attacker would then be executed, compromising the victim’s machine. This is due to the keras.models.load_model function being used in unsafe_check_pb in the watchtower/src/utils/model_inspector_util.py file, which is used to scan .pb files. When a model is loaded with this function, it executes any lambda layers contained in it, which executes any malicious payloads. A user could also scan this file from a GitHub or HuggingFace repository using Watchtower, using the built-in functionality.
def unsafe_check_pb(model_path: str):
"""
The unsafe_check_pb function is designed to examine models with the .pb extension for potential vulnerabilities.
...
"""
tool_output = list()
# If the provided path is a file, get the parent directory
if os.path.isfile(model_path):
model_path = os.path.dirname(model_path)
try:
model = tf.keras.models.load_model(model_path)
Timeline
Timeline
August 19, 2024 — Disclosed vulnerability to Bosch AI Shield
October 19, 2024 — Bosch AI Shield responds, asking for more time due to the report getting lost in spam filtering policies
November 27, 2024 — Bosch AI Shield released a patch for the vulnerabilities and stated that no CVE would be assigned
“After a thorough review by our internal security board, it was determined that the issue does not warrant a CVE assignment.”