Skip to content
#

fairness-ai

Here are 136 public repositories matching this topic...

Responsible AI Toolbox is a suite of tools providing model and data exploration and assessment user interfaces and libraries that enable a better understanding of AI systems. These interfaces and libraries empower developers and stakeholders of AI systems to develop and monitor AI more responsibly, and take better data-driven actions.

  • Updated Nov 26, 2024
  • TypeScript

WEFE: The Word Embeddings Fairness Evaluation Framework. WEFE is a framework that standardizes the bias measurement and mitigation in Word Embeddings models. Please feel welcome to open an issue in case you have any questions or a pull request if you want to contribute to the project!

  • Updated Jun 18, 2024
  • Python

Improve this page

Add a description, image, and links to the fairness-ai topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the fairness-ai topic, visit your repo's landing page and select "manage topics."

Learn more