Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

A survey on a line of work following (Qi. et al. 2023) #8

Open
huangtiansheng opened this issue Oct 5, 2024 · 0 comments
Open

A survey on a line of work following (Qi. et al. 2023) #8

huangtiansheng opened this issue Oct 5, 2024 · 0 comments

Comments

@huangtiansheng
Copy link

huangtiansheng commented Oct 5, 2024

Hi authors,

Thanks for the wonderful initial work on harmful fine-tuning. We recently noticed a huge amount of papers coming out on the harmful fine-tutning attacks for LLMs. We have pre-printed a survey paper summarizing the existing following-up paper on this issue.

Harmful Fine-tuning Attacks and Defenses for Large Language Models: A Survey https://arxiv.org/abs/2409.18169

Repo: https://github.com/git-disl/awesome_LLM-harmful-fine-tuning-papers

It would be nice if the authors can incorporate our survey in the readme, as this probably can attract more people to work on this important topic. But no pressure if you feel it is inappropriate.

Thanks,
Tiansheng Huang

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant