Skip to content

Natural Language Processing (NLP) and Large Language Models (LLM) and LLM with Society, Bias and toxicity

Notifications You must be signed in to change notification settings

YanSte/NLP-LLM-Society-Bias-Toxicity

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

| NLP | LLM | Society | Bias Toxicity |

Natural Language Processing (NLP) and Large Language Models (LLM), LLM with Society, Bias and toxicity

Learning

| Overview

This notebook covers various aspects related to language models, with a focus on societal implications. Here's an overview of the main sections:

Learning Objectives

In this section, learning objectives are outlined, emphasizing the following points:

  1. Understanding representation bias in training data.
  2. Using Hugging Face to calculate toxicity scores.
  3. Using SHAP to generate explanations for model output.
  4. Exploring the latest research advancements in model explanation: contrastive explanation.

About

Natural Language Processing (NLP) and Large Language Models (LLM) and LLM with Society, Bias and toxicity

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published