![]() ![]() ![]() Meta says it did not remove toxic data from the data set, because leaving it in might help LLaMA 2 detect hate speech better, and removing it could risk accidentally filtering out some demographic groups. The company says it did not use Meta user data in LLaMA 2, and excluded data from sites it knew had lots of personal information.ĭespite that, LLaMA 2 still spews offensive, harmful, and otherwise problematic language, just like rival models. Al-Dahle says there were two sources of training data: data that was scraped online, and a data set fine-tuned and tweaked according to feedback from human annotators to behave in a more desirable way. The model was trained on 40% more data than its predecessor. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |