You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for the wonderful initial work on harmful fine-tuning. We recently noticed a huge amount of papers coming out on the harmful fine-tutning attacks for LLMs. We have pre-printed a survey paper summarizing the existing following-up paper on this issue.
It would be nice if the authors can incorporate our survey in the readme, as this probably can attract more people to work on this important topic. But no pressure if you feel it is inappropriate.
Thanks,
Tiansheng Huang
The text was updated successfully, but these errors were encountered:
Hi authors,
Thanks for the wonderful initial work on harmful fine-tuning. We recently noticed a huge amount of papers coming out on the harmful fine-tutning attacks for LLMs. We have pre-printed a survey paper summarizing the existing following-up paper on this issue.
Harmful Fine-tuning Attacks and Defenses for Large Language Models: A Survey https://arxiv.org/abs/2409.18169
Repo: https://github.com/git-disl/awesome_LLM-harmful-fine-tuning-papers
It would be nice if the authors can incorporate our survey in the readme, as this probably can attract more people to work on this important topic. But no pressure if you feel it is inappropriate.
Thanks,
Tiansheng Huang
The text was updated successfully, but these errors were encountered: