What is FHIBE?

The Fair Human-Centric Image Benchmark* (pronounced Fee-Bee) is the first publicly available, globally diverse, consent-driven dataset built specifically to evaluate fairness in human-centric computer vision tasks. While AI vision powers everything from autonomous vehicles to smartphones, most datasets still suffer from bias, lack of diversity, and ethical concerns around consent and compensation.

Why a Fairness Benchmark?

Bias evaluation is essential for ethical AI. Yet most evaluation datasets fall short: lacking consent, or detailed annotations. FHIBE fills that gap as a responsibly curated, consent-based dataset designed to advance fairness in computer vision. How FHIBE is different:

Consent
Compensation
Diversity
Comprehensive Annotations
Real-World Utility

Consent

Every dataset participant gave informed, revocable consent. Data collection is designed to comply with data protection laws.

Compensation

All image subjects, annotators, and QA reviewers were fairly paid at or above local minimum wage. Participants can withdraw their data at any time without affecting compensation.

Diversity

FHIBE maximizes diversity across demographics, appearance, poses, and environments. It includes detailed data on age, pronouns, ancestry, skin tone, eye color, hair type, and visible markers to support granular bias analysis.

Comprehensive Annotations

FHIBE includes Pixel-level labels, 33 keypoints, and 28 segmentation categories—linked to anonymized annotator IDs.

Real-World Utility

FHIBE is designed for fairness evaluation across pose estimation, face/body detection, segmentation, synthesis, and vision-language models.

*FHIBE is strictly for fairness/bias purposes – FHIBE cannot be used for training, with the narrow exception of training bias mitigation tools.

A Fair Reflection - a short film

Sony’s Quest to Make AI Models Work for Everyone

Many ethical issues with computer vision start at the data layer when data is often not collected in an ethical and responsible way. Ethical data collection is an extremely challenging issue and FHIBE is the first dataset of its kind, a new benchmark for the community. It’s important that the whole industry builds AI on a foundation of ethical data. Watch our short film to learn more.

Dataset Details

More than 10,000 Images from Nearly 2,000 People

FHIBE includes 10,318 diverse, consented images—each featuring one or two people, captured across 81 jurisdictions using varied environments and camera types. Most participants contributed around six images each, all provided directly by the individuals depicted.

FHIBE includes two-face focused assets:

  • Unaligned face-cropped dataset
  • Aligned face-cropped dataset

To protect privacy, all non-consensual individuals have been fully anonymized, and all personally identifiable information (PII) has been redacted, minimizing re-identification risk.

Comprehensive Annotations and Rich Metadata

  • Self-reported image subject attributes (not guessed)
  • Face and person bounding boxes
  • 33 body and facial keypoints
  • 28 segmentation categories
  • Anonymized annotator IDs

Evaluating Fairness in Computer Vision with FHIBE

FHIBE supports fairness evaluation across human pose estimation, face and body detection, segmentation, verification, image editing and synthesis, and vision-language models.

Human Pose Estimation

FHIBE supports the evaluation of human pose estimation models by providing 2D keypoint annotations (x, y coordinates) that capture the geometric structure of human bodies and faces. The dataset includes 33 keypoints. Keypoints are localized for major landmarks, such as the right eye inner corner, nose, right hip, and left foot index.

Face and Body Detection

FHIBE enables the evaluation of face detection and person detection models by providing bounding boxes with precise (x, y) coordinates for each image subject. These annotations allow for accurate localization of both the human body and face within an image.

Segmentation

FHIBE enables the evaluation of segmentation models by providing pixel-wise annotations that partition the human body and face into 28 distinct segmentation categories (e.g., face skin, inner mouth, bag, right shoe). Each pixel is assigned a label, ensuring precise differentiation of anatomical regions and accessories.

Identity Verification

FHIBE supports the evaluation of face verification models, which determine whether two images show the same person. To ensure reliable comparisons, FHIBE includes an aligned face-cropped dataset, providing standardized images for testing identity verification systems.

Image Editing and Synthesis

FHIBE supports the evaluation of image restoration and super-resolution models, assessing their ability to enhance or reconstruct human-centric images. Image restoration recovers degraded images by removing noise, correcting distortions, and restoring lost details. Super-resolution sharpens low-resolution images, improving clarity while preserving the subject’s identity and attributes.

Vision-Language Models

FHIBE enables the evaluation of vision-language models (VLMs) in image comprehension and recognition tasks, including visual question answering (VQA). By providing diverse, annotated human-centric images, FHIBE helps assess bias in how VLMs reason about people.


Frequently Asked Questions

Dataset Access and Usage
Dataset Creation and Ethics
Maintenance and Updates
Dataset Scope and Attributes
Safety and Security
Dataset Limitations
illustration of Sony AI FHIBE team

FHIBE is a Project of Sony AI

Over the course of three years, a global team of Sony AI researchers, engineers, policy specialists, and project managers worked to develop the rigorous procedures for data collection, annotation, and validation that became FHIBE. Their work was further supported by legal, privacy, IT, and QA specialists. The Sony AI Ethics research team is leading the way towards more ethical AI that protects the interests of AI users and creators by pushing the industry to ensure AI technologies are fair, transparent, and accountable.

Sony AI, a division of Sony Research, was founded as a subsidiary of Sony Group on April 1, 2020, with the mission to "Unleash human imagination and creativity with AI."

Learn more about Ethics Research at Sony AI.