[NLPCC 2024] Shared Task 10: Regulating Large Language Models
-
Updated
Jun 12, 2024
[NLPCC 2024] Shared Task 10: Regulating Large Language Models
Detoxifying Online Discourse: A Guided Response Generation Approach for Reducing Toxicity in User-Generated Text
Fine-tuning FLAN-T5 with PPO and PEFT to generate less toxic text summaries. This notebook leverages Meta AI's hate speech reward model and utilizes RLHF techniques for improved safety.
PMLDL Assignment 1
Add a description, image, and links to the detoxification topic page so that developers can more easily learn about it.
To associate your repository with the detoxification topic, visit your repo's landing page and select "manage topics."