Projects

RESEARCH, INFRASTRUCTURE & ORGANIZATION
Research

Evaluation Cards

This project addresses the need for a structured and systematic approach to documenting AI model evaluations through the creation of "evaluation cards," focusing specifically on technical base syst...

Chairs: Avijit Ghosh, Anka Reuel

Infrastructure

Every Eval Ever

Every Eval Ever is a standardized schema for AI evaluation results, promoting interoperability, reproducibility, and transparency across the ecosystem.

Chairs: Jan Batzner, Leshem Choshen, Avijit Ghosh

Research

Benchmark Saturation

This project aims to investigate how to systematically characterize the complexity and behavior of AI benchmarks over time, with the overarching goal of informing more robust benchmark design. The ...

Chairs: Anka Reuel, Mubashara Akhtar

Infrastructure

Evaluation Harness and Tutorials

The Eleuther Harness Tutorials project is designed to lower the barrier to entry for using the LM Evaluation Harness, making it easier for researchers and practitioners to onboard, evaluate, and co...

Chairs: Baber Abbasi, Stella Biderman

Research

Evaluation Science

Recognizing the current lack of robustness in GPAI risk evaluations and the resulting limitations for informed decision-making and societal preparedness, this project aims to establish a scientific...

Chairs: Subho Majumdar, Patricia Paskov

Organization

Outreach & Research Engagement

Building upon the momentum of NeurIPS 2024, we aim to cultivate a QueerInAI-esque presence at other relevant venues by organizing social events and short talks to broaden our reach and foster commu...

Chairs: Jennifer Mickel, Usman Gohar

Get Involved

Join Our Community

Researchers, practitioners, and students are welcome to contribute to our mission. Send us an email to learn more about getting involved.

[email protected]

Hosted By