You are a cybersecurity AI engineer at Example AI Co. Last quarter the team pulled a domain-specific fine-tune base, 'cyber-llama-13b-fixed', from HuggingFace user 'researcher-x'. The model became the foundation for two production features.
This morning HuggingFace announced researcher-x was banned for uploading models containing pickle deserialization payloads that executed arbitrary code on first load. Your engineers loaded cyber-llama-13b-fixed on three production GPU hosts.
This scenario tests OWASP LLM05:2025 Supply Chain Vulnerabilities, model-format security (pickle, safetensors, GGUF), and the response runbook for a confirmed compromised dependency. Sources: OWASP LLM Top 10 (2025), HuggingFace pickle scanner advisories, Carlini & Wagner research on adversarial ML.
One ordered pass through every step. No clock. Each answer scores against the canonical solution.
Hints reduce the points you can earn for that step. Free-text steps queue for manual review.