🔬
The Fair-ly Project
  • Welcome to RumorMill
    • Recent Papers Timeline
  • Fair-ly Toolkit
    • Chrome Extension
    • Python Package
      • TextAnalyzer Pipeline
      • MultimodalAnalyzer Pipeline
    • Hosted APIs
  • Research
    • Sequence Classification
      • Binary
      • Multi-class
    • Named-Entity Recognition
      • Token Classification
    • Multimodal
      • Image + Text Pair Classification
    • Datasets
      • News Media Bias Plus (2024)
      • BEADs Dataset (2024)
      • GUS Dataset (2024)
      • BABE Dataset (2022)
  • Learn
    • Blog Posts
      • Training a model for multi-label NER
      • Binary Classification w/ BERT
  • Join the Project
    • To Do List
    • Discord Server
    • GitHub Repo
  • Misc
    • Privacy Policy
Powered by GitBook
On this page
  • 🛠️ RumorMill Toolkit
  • 🧠 Learn
Edit on GitHub

Welcome to RumorMill

A directory and implementation of SOTA bias detection research papers, all in one place.

NextRecent Papers Timeline

Last updated 4 months ago

RumorMill was founded and is maintained by ML researchers of and .

RumorMill is an open-source collection of resources, such as:

  • Research papers

  • Blogs and videos

  • Datasets and models

Ethos: Bias-detection research should be accessible to users and developers of all levels.


🛠️ RumorMill Toolkit

Here are a few tools we built for using state-of-the-art models in practice, something for everyone :).

Fair-ly Extension (Pending Renaming)

Our Chrome extension, , is a showcase of SOTA models. Anyone can run a bias analysis pipeline/dashboard on their webpage, no code required.

It was created to open-the-door for new people to bias detection technology, by demonstrating it's strengths and weaknesses. The tasks it's intended to preform are:

Try this interactive demo for a quick look:

Install Our Package:

pip install the-fairly-project

How to Use The Pipeline:

from fairly import TextAnalyzer

analyzer = TextAnalyzer(bias="ternary", classes=True, top_k_classes=3, ner="gus")
result = analyzer.analyze("Tall people are so clumsy.")

Example Response:

{
'text': {
    'text': 'Tall people are so clumsy.', 
    'label': 'Slightly Biased', 
    'score': 0.6829080581665039, 
    'aspects': {
      'physical': 0.9650779366493225, 
      'gender': 0.024978743866086006, 
      'socioeconomic': 0.023334791883826256
    }
}, 
'ner': [
    {'token': 'tall', 'labels': ['B-STEREO', 'B-GEN', 'B-UNFAIR']}, 
    {'token': 'people', 'labels': ['I-STEREO', 'I-GEN', 'I-UNFAIR']}, 
    {'token': 'are', 'labels': ['I-STEREO']}, 
    {'token': 'so', 'labels': ['I-STEREO']}, 
    {'token': 'clumsy', 'labels': ['I-STEREO', 'B-UNFAIR', 'I-UNFAIR']},
    {'token': '.', 'labels': ['I-STEREO', 'I-UNFAIR']}
]
}

🧠 Learn

(sentence -> biased/fair).

(sentence -> gender bias, racial bias, ...).

of generalizations, unfairness, and stereotypes.

Walk through and run all the pipelines in this

If you're interested in contributing to the open-source tool-kit, check out our and join our .

Binary bias classification
Bias aspect classification
Token classification
Google Colab Notebook 💻

TextAnalyzer Docs

MultimodalAnalyzer Docs

Binary Classification API

Types-of-bias Classification API

GUS-Net (Token Classification) API

GitHub
Discord

Recent Papers

Papers to cite ;)

Binary Classification

Classifying text sequences as "Biased" or "Fair."

Multi-Class Classification

Classifying text sequences into more specific classes.

Named-Entity Recognition

Classifying tokens (words) that contain bias.

Multimodal Classification

Classifying image and text pairs for bias.

Discord

Ask questions or share a project.

Ethical Spectacle Research
The Vector Institute
Fair-ly
Fair-ly - Chrome Web Store
Logo