Skip to main content
BackFine-tuning related (Harmful fine-tuning of open-weights models)
Home/Risks/Gipiškis2024/Fine-tuning related (Harmful fine-tuning of open-weights models)

Fine-tuning related (Harmful fine-tuning of open-weights models)

Sub-category
Risk Domain

Using AI systems to develop cyber weapons (e.g., by coding cheaper, more effective malware), develop new or enhance existing weapons (e.g., Lethal Autonomous Weapons or chemical, biological, radiological, nuclear, and high-yield explosives), or use weapons to cause mass harm.

"Models with publicly available weights can be fine-tuned for harmful activities by bad actors, using significantly fewer resources (in terms of time and money) compared to the original training cost [115, 78]."(p. 14)

Other risks from Gipiškis2024 (144)