Dataset Preview
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Job manager crashed while running this job (missing heartbeats).
Error code:   JobManagerCrashedError

Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.

image
image
label
class label
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
End of preview.

CrowdHuman: A Benchmark for Detecting Human in a Crowd

CrowdHuman is a benchmark dataset to better evaluate detectors in crowd scenarios. The CrowdHuman dataset is large, rich-annotated and contains high diversity. CrowdHuman contains 15000, 4370 and 5000 images for training, validation, and testing, respectively. There are a total of 470K human instances from train and validation subsets and 23 persons per image, with various kinds of occlusions in the dataset. Each human instance is annotated with a head bounding-box, human visible-region bounding-box and human full-body bounding-box. We hope our dataset will serve as a solid baseline and help promote future research in human detection tasks.

image/png Volume, density and diversity of different human detection datasets. For fair comparison, we only show the statistics of training subset.

πŸ” Samples

image/png image/png image/png
image/png image/png image/png
image/png image/png image/png

πŸ“ Files

  • CrowdHuman_train01.zip
  • CrowdHuman_train02.zip
  • CrowdHuman_train03.zip
  • CrowdHuman_val.zip
  • CrowdHuman_test.zip
  • annotation_train.odgt
  • annotation_val.odgt

πŸ–¨ Data Format

We support annotation_train.odgt and annotation_val.odgt which contains the annotations of our dataset.

What is odgt?

odgt is a file format that each line of it is a JSON, this JSON contains the whole annotations for the relative image. We prefer using this format since it is reader-friendly.

Annotation format

JSON{
    "ID" : image_filename,
    "gtboxes" : [gtbox], 
}

gtbox{
    "tag" : "person" or "mask", 
    "vbox": [x, y, w, h],
    "fbox": [x, y, w, h],
    "hbox": [x, y, w, h],
    "extra" : extra, 
    "head_attr" : head_attr, 
}

extra{
    "ignore": 0 or 1,
    "box_id": int,
    "occ": int,
}

head_attr{
    "ignore": 0 or 1,
    "unsure": int,
    "occ": int,
}
  • Keys in extra and head_attr are optional, it means some of them may not exist
  • extra/head_attr contains attributes for person/head
  • tag is mask means that this box is crowd/reflection/something like person/... and need to be ignore(the ignore in extra is 1)
  • vbox, fbox, hbox means visible box, full box, head box respectively

⚠️ Terms of use:

by downloading the image data you agree to the following terms:

  1. You will use the data only for non-commercial research and educational purposes.
  2. You will NOT distribute the above images.
  3. Megvii Technology makes no representations or warranties regarding the data, including but not limited to warranties of non-infringement or fitness for a particular purpose.
  4. You accept full responsibility for your use of the data and shall defend and indemnify Megvii Technology, including its employees, officers and agents, against any and all claims arising from your use of the data, including but not limited to your use of any copies of copyrighted images that you may create from the data.

πŸ† Related Challenge

πŸ“š Citaiton

Please cite the following paper if you use our dataset.

@article{shao2018crowdhuman,
  title={CrowdHuman: A Benchmark for Detecting Human in a Crowd},
  author={Shao, Shuai and Zhao, Zijian and Li, Boxun and Xiao, Tete and Yu, Gang and Zhang, Xiangyu and Sun, Jian},
  journal={arXiv preprint arXiv:1805.00123},
  year={2018}
}

πŸ‘₯ People

Downloads last month
108