Deep Learning Datasets

CVonline: Image Databases


This is a collated list of image and video databases that people
have found useful for computer vision research and algorithm evaluation.

An important article
How Good Is My Test Data? Introducing Safety Analysis for Computer Vision
(by Zendel, Murschitz, Humenberger, and Herzner)
introduces a methodology for ensuring that your dataset has sufficient
variety that algorithm results on the dataset are representative of
the results that one could expect in a real setting.
In particular, the team have produced a
Checklist of potential
hazards (imaging situations) that may cause algorithms to have problems.
Ideally, test datasets should have examples of the relevant hazards.

Index by Topic

  1. Action Databases
  2. Attribute recognition
  3. Autonomous Driving
  4. Biological/Medical
  5. Camera calibration
  6. Face and Eye/Iris Databases
  7. Fingerprints
  8. General Images
  9. General RGBD and depth datasets
  10. General Videos
  11. Hand, Hand Grasp, Hand Action and Gesture Databases
  12. Image, Video and Shape Database Retrieval
  13. Object Databases
  14. People (static and dynamic), human body pose
  15. People Detection and Tracking Databases (See also Surveillance)
  16. Remote Sensing
  17. Scenes or Places, Scene Segmentation or Classification
  18. Segmentation
  19. Simultaneous Localization and Mapping
  20. Surveillance and Tracking (See also People)
  21. Textures
  22. Urban Datasets
  23. Vision and Natural Language
  24. Other Collection Pages
  25. Miscellaneous Topics

Another helpful site is the YACVID page.


Action Databases

See also:
Action Recognition’s dataset summary with league tables (Gall, Kuehne, Bhattarai).

  1. 20bn-Something-Something – densely-labeled video clips that show humans performing predefined basic actions with everyday objects (Twenty Billion Neurons GmbH)
  2. 3D online action dataset – There are seven action categories (Microsoft and Nanyang Technological University)
  3. 50 Salads – fully annotated 4.5 hour dataset of RGB-D video + accelerometer data, capturing 25 people preparing two mixed salads each (Dundee University, Sebastian Stein)
  4. ActivityNet – A Large-Scale Video Benchmark for Human Activity Understanding (200 classes, 100 videos per class, 648 video hours) (Heilbron, Escorcia, Ghanem and Niebles)
  5. Action Detection in Videos – MERL Shopping Dataset consists of 106 videos, each of which is a sequence about 2 minutes long (Michael Jones, Tim Marks)
  6. Actor and Action Dataset – 3782 videos, seven classes of actors performing eight different actions (Xu, Hsieh, Xiong, Corso)
  7. An analyzed collation of various labeled video datasets for action recognition (Kevin Murphy)
  8. ASLAN Action similarity labeling challenge database (Orit Kliper-Gross)
  9. Attribute Learning for Understanding Unstructured Social Activity – Database of videos containing 10 categories of unstructured social events to recognise, also annotated with 69 attributes. (Y. Fu Fudan/QMUL, T. Hospedales Edinburgh/QMUL)
  10. AVA: A Video Dataset of Atomic Visual Action80 atomic visual actions in 430 15-minute movie clips Google Machine Perception Research Group)
  11. BEHAVE Interacting Person Video Data with markup (Scott Blunsden, Bob Fisher, Aroosha Laghaee)
  12. BU-action Datasets – Three image action datasets (BU101, BU101-unfiltered, BU203-unfiltered) that have 1:1 correspondence with classes of the video datasets UCF101 and ActivityNet. (S. Ma, S. A. Bargal, J. Zhang, L. Sigal, S. Sclaroff.)
  13. Berkeley MHAD: A Comprehensive Multimodal Human Action Database (Ferda Ofli)
  14. Berkeley Multimodal Human Action Database – five different modalities to expand the fields of application (University of California at Berkeley and Johns Hopkins University)
  15. Breakfast dataset – It’s a dataset with 1712 video clips showing 10 kitchen activities, which are hand segmented into 48 atomic action classes . (H. Kuehne, A. B. Arslan and T. Serre )
  16. Bristol Egocentric Object Interactions Dataset – Contains videos shot from a first-person (egocentric) point of view of 3-5 users performing tasks in six different locations (Dima Damen, Teesid Leelaswassuk and Walterio Mayol-Cuevas, Bristol University)
  17. Brown Breakfast Actions Dataset – 70 hours, 4 million frames of 10 different breakfast preparation activities (Kuehne, Arslan and Serre)
  18. CAD-120 dataset – focuses on high level activities and object interactions (Cornell University)
  19. CAD-60 dataset – The CAD-60 and CAD-120 data sets comprise of RGB-D video sequences of humans performing activities (Cornell University)
  20. CVBASE06: annotated sports videos (Janez Pers)
  21. Charades Dataset – 10,000 videos from 267 volunteers, each annotated with multiple activities, captions, objects, and temporal localizations. (Sigurdsson, Varol, Wang, Laptev, Farhadi, Gupta)
  22. Composable activities dataset – Different combinations of 26 atomic actions formed 16 activity classes which were performed by 14 subjects and annotations were provided (Pontificia Universidad Catolica de Chile and Universidad del Norte)
  23. Cornell Activity Datasets CAD 60, CAD 120 (Cornell Robot Learning Lab)
  24. DMLSmartActions dataset – Sixteen subjects performed 12 different actions in a natural manner (University of British Columbia)
  25. Depth-included Human Action video dataset – It contains 23 different actions (CITI in Academia Sinica)
  26. DogCentric Activity Dataset – first-person videos taken from a camera mounted on top of a *dog* (Michael Ryoo)
  27. Edinburgh ceilidh overhead video data – 16 ground-truthed dances viewed from overhead, where the 10 dancers follow a structured dance pattern (2 different dances). The dataset is useful for highly structured behavior understanding (Aizeboje, Fisher)
  28. EPIC-KITCHENS – egocentric video recorded by 32 participants in their native kitchen environments, non-scripted daily activities, 11.5M frames, 39.6K frame-level action segments and 454.2K object bounding boxes (Damen, Doughty, Fidler, et al)
  29. EPFL crepe cooking videos – 6 types of structured cooking activity (12) videos in 1920×1080 resolution (Lee, Ognibene, Chang, Kim and Demiris)
  30. ETS Hockey Game Event Data Set – This data set contains footage of two hockey games captured using fixed cameras. (M.-A. Carbonneau, A. J. Raymond, E. Granger, and G. Gagnon)
  31. FCVID: Fudan-Columbia Video Dataset – 91,223 Web videos annotated manually according to 239 categories (Jiang, Wu, Wang, Xue, Chang)
  32. G3D – synchronised video, depth and skeleton data for 20 gaming actions captured with Microsoft Kinect (Victoria Bloom)
  33. G3Di – This dataset contains 12 subjects split into 6 pairs (Kingston University)
  34. Gaming 3D dataset – real-time action recognition in gaming scenario (Kingston University)
  35. Georgia Tech Egocentric Activities – Gaze(+) – videos of where people look at and their gaze location (Fathi, Li, Rehg)
  36. HMDB: A Large Human Motion Database (Serre Lab)
  37. Hollywood 3D dataset – 650 3D video clips, across 14 action classes (Hadfield and Bowden)
  38. Human Actions and Scenes Dataset (Marcin Marszalek, Ivan Laptev, Cordelia Schmid)
  39. Hollywood Extended – 937 video clips with a total of 787720 frames containing sequences of 16 different actions from 69 Hollywood movies. (Bojanowski, Lajugie, Bach, Laptev, Ponce, Schmid, and Sivic)
  40. HumanEva: Synchronized Video and Motion Capture Dataset for Evaluation of Articulated Human Motion (Brown University)
  41. I-LIDS video event image dataset (Imagery library for intelligent detection systems) (Paul Hosner)
  42. I3DPost Multi-View Human Action Datasets (Hansung Kim)
  43. IAS-lab Action dataset – contain sufficient variety of actions and number of people
    performing the actions (IAS Lab at the University of Padua)
  44. ICS-FORTH MHAD101 Action Co-segmentation – 101 pairs of long-term action sequences that share one or multiple common actions to be co-segmented, contains both 3d skeletal and video related frame-based features (Universtiy of Crete and FORTH-ICS, K. Papoutsakis)
  45. INRIA Xmas Motion Acquisition Sequences (IXMAS) (INRIA)
  46. InfAR Dataset -Infrared Action Recognition at Different Times Neurocomputing(Chenqiang Gao, Yinhe Du, Jiang Liu, Jing Lv, Luyu Yang, Deyu Meng, Alexander G. Hauptmann)
  47. JHMDB: Joints for the HMDB dataset (J-HMDB) based on 928 clips from HMDB51 comprising 21 action categories (Jhuang, Gall, Zuffi, Schmid and Black)
  48. JPL First-Person Interaction dataset – 7 types of human activity videos taken from a first-person viewpoint (Michael S. Ryoo, JPL)
  49. Jena Action Recognition Dataset – Aibo dog actions (Korner and Denzler)
  50. K3Da – Kinect 3D Active dataset – K3Da (Kinect 3D active) is a realistic clinically relevant human action dataset containing skeleton, depth data and associated participant information (D. Leightley, M. H. Yap, J. Coulson, Y. Barnouin and J. S. McPhee)
  51. Kinetics Human Action Video Dataset – 300,000 video clips, 400 human action classe, 10 second clips, single action per clip (Kay, Carreira, et al)
  52. KIT Robo-Kitchen Activity Data Set – 540 clips of 17 people performing 12 complex
    kitchen activities.(L. Rybok, S. Friedberger, U. D. Hanebeck, R. Stiefelhagen)
  53. KTH human action recognition database (KTH CVAP lab)
  54. Karlsruhe Motion, Intention, and Activity Data set (MINTA) – 7 types of activities of daily living including fully motion primitive segments.(D. Gehrig, P. Krauthausen, L. Rybok, H. Kuehne, U. D. Hanebeck, T. Schultz, R. Stiefelhagen)
  55. LIRIS Human Activities Dataset – contains (gray/rgb/depth) videos showing people performing various activities (Christian Wolf, et al, French National Center for Scientific Research)
  56. MEXaction2 action detection and localization dataset – to support the development and evaluation of methods for ‘spotting’ instances of short actions in a relatively large video database: 77 hours, 117 videos (Michel Crucianu and Jenny Benois-Pineau)
  57. MPII Cooking Activities Dataset for fine-grained cooking activity recognition, which also includes the continuous pose estimation challenge (Rohrbach, Amin, Andriluka and Schiele)
  58. MPII Cooking 2 Dataset – A large dataset of fine-grained cooking activities, an extension of the MPII Cooking Activities Dataset. (Rohrbach, Rohrbach, Regneri, Amin, Andriluka, Pinkal, Schiele)
  59. MSR-Action3D – benchmark RGB-D action dataset (Microsoft Research Redmond and University of Wollongong)
  60. MSRActionPair dataset – : Histogram of Oriented 4D Normals for Activity Recognition from Depth Sequences (University of Central Florida and Microsoft)
  61. MSRC-12 Kinect gesture data set – 594 sequences and 719,359 frames from people performing 12 gestures (Microsoft Research Cambridge)
  62. MSRC-12 dataset – sequences of human movements, represented as body-part locations, and the associated gesture (Microsoft Research Cambridge and University of Cambridge)
  63. MSRDailyActivity3D Dataset – There are 16 activities (Microsoft and the Northwestern University)
  64. ManiAc RGB-D action dataset: different manipulation actions, 15 different versions, 30 different objects manipulated, 20 long and complex chained manipulation sequences (Eren Aksoy)
  65. Mivia dataset – It consists of 7 high-level actions
    performed by 14 subjects. (Mivia Lab at the University of Salemo)
  66. MuHAVi – Multicamera Human Action Video Data (Hossein Ragheb)
  67. Multi-modal action detection (MAD) Dataset – It contains 35 sequential actions performed by 20 subjects. (CarnegieMellon University)
  68. Multiview 3D Event dataset – This dataset includes 8 categories of events performed by 8 subjects (University of California at Los Angles)
  69. NTU RGB+D Action Recognition Dataset – NTU RGB+D is a large scale dataset for human action recognition(Amir Shahroudy)
  70. Northwestern-UCLA Multiview Action 3D – There are 10 action categories:(Northwestern University and University of California at Los Angles)
  71. Oxford TV based human interactions (Oxford Visual Geometry Group)
  72. Parliament – The Parliament dataset is a collection of 228 video sequences, depicting political speeches in the Greek parliament(Michalis Vrigkas, Christophoros Nikou, Ioannins A. kakadiaris)
  73. RGB-D activity dataset – Each video in the dataset contains 2-7 actions involving interaction with different objects. (Cornell University and Stanford University)
  74. RGBD-Action-Completion-2016 – This dataset includes 414 complete/incomplete object interaction sequences, spanning six actions and presenting RGB, depth and skeleton data. (Farnoosh Heidarivincheh, Majid Mirmehdi, Dima Damen)
  75. RGBD-SAR Dataset – RGBD-SAR Dataset (University of Electronic Science and Technology of China and Microsoft)
  76. Rochester Activities of Daily Living Dataset (Ross Messing)
  77. SBU Kinect Interaction Dataset – It contains eight types of interactions (Stony Brook University)
  78. SBU-Kinect-Interaction dataset v2.0 – It comprises of RGB-D video sequences of humans performing interaction activities (Kiwon Yun etc.)
  79. SDHA Semantic Description of Human Activities 2010 contest – Human Interactions (Michael S. Ryoo, J. K. Aggarwal, Amit K. Roy-Chowdhury)
  80. SDHA Semantic Description of Human Activities 2010 contest – aerial views (Michael S. Ryoo, J. K. Aggarwal, Amit K. Roy-Chowdhury)
  81. SFU Volleyball Group Activity Recognition – 2 levels annotations dataset (9 players’ actions and 8 scene’s activity) for volleyball videos.(M. Ibrahim, S. Muralidharan, Z. Deng, A. Vahdat, and G. Mori / Simon Fraser University)
  82. SYSU 3D Human-Object Interaction Dataset – Forty subjects perform 12 distinct activities (Sun Yat-sen University)
  83. ShakeFive Dataset – contains only two actions, namely hand shake and high
    five. (Universiteit Utrecht)
  84. ShakeFive2 – A dyadic human interaction dataset with limb level annotations on 8 classes in 153 HD videos(Coert van Gemeren, Ronald Poppe, Remco Veltkamp)
  85. Sports Videos in the Wild (SVW) – SVW is comprised of 4200 videos captured solely with smartphones by users of Coach Eye smartphone app, a leading app for sports training developed by TechSmith corporation.(Seyed Morteza Safdarnejad, Xiaoming Liu)
  86. Stanford Sport Events dataset (Jia Li)
  87. The Leeds Activity Dataset–Breakfast (LAD–Breakfast) – composed of 15 annotated videos, representing five different people having breakfast or other simple meal; (John Folkesson et al.)
  88. THUMOS – Action Recognition in Temporally Untrimmed Videos! – 430 hours of video data and 45 million frames (Gorban, Idrees, Jiang, Zamir, Laptev Shah, Sukthanka)
  89. TUM Kitchen Data Set of Everyday Manipulation Activities (Moritz Tenorth, Jan Bandouch)
  90. TV Human Interaction Dataset (Alonso Patron-Perez)
  91. The Falling Detection dataset – Six subjects in two sceneries performed a series of actions continuously (University of Texas)
  92. The TJU dataset – contains 22 actions performed by 20 subjects in two different environments; a total of
    1760 sequences. (Tianjin University)
  93. UCF-iPhone Data Set – 9 Aerobic actions were recorded from (6-9) subjects using the Inertial Measurement Unit (IMU) on an Apple iPhone 4 smartphone. (Corey McCall, Kishore Reddy and Mubarak Shah)
  94. The UPCV action dataset – The dataset consists of 10 actions performed by 20 subjects twice. (University of Patras)
  95. UC-3D Motion Database – Available data types encompass high resolution Motion Capture, acquired with MVN Suit from Xsens and Microsoft Kinect RGB and depth images.(Institute of Systems and Robotics, Coimbra, Portugal)
  96. UCF 101 action dataset 101 action classes, over 13k clips and 27 hours of video data (Univ of Central Florida)
  97. UCFKinect – The dataset is composed of 16 actions (University of Central Florida Orlando)
  98. UCLA Human-Human-Object Interaction (HHOI) Dataset Vn1 – Human interactions in RGB-D videos (Shu, Ryoo, and Zhu)
  99. UCLA Human-Human-Object Interaction (HHOI) Dataset Vn2 – Human interactions in RGB-D videos (version 2) (Shu, Gao, Ryoo, and Zhu)
  100. UCR Videoweb Multi-camera Wide-Area Activities Dataset (Amit K. Roy-Chowdhury)
  101. UTD-MHAD – Eight subjects performed 27 actions four times. (University of Texas at Dallas)
  102. UTKinect dataset – Ten types of human actions were performed twice by 10 subjects (University of Texas)
  103. UWA3D Multiview Activity Dataset – Thirty activities were performed by 10 individuals (University of Western Australia)
  104. Univ of Central Florida – 50 Action Category Recognition in Realistic Videos (3 GB) (Kishore Reddy)
  105. Univ of Central Florida – ARG Aerial camera, Rooftop camera and Ground camera (UCF Computer Vision Lab)
  106. Univ of Central Florida – Feature Films Action Dataset (Univ of Central Florida)
  107. Univ of Central Florida – Sports Action Dataset (Univ of Central Florida)
  108. Univ of Central Florida – YouTube Action Dataset (sports) (Univ of Central Florida)
  109. Unsegmented Sports News Videos – Database of 74 sports news videos tagged with 10 categories of sports. Designed to test multi-label video tagging. (T. Hospedales, Edinburgh/QMUL)
  110. Utrecht Multi-Person Motion Benchmark (UMPM). – a collection of video recordings of people together with a ground truth based on motion capture data.(N.P. van der Aa, X. Luo, G.J. Giezeman, R.T. Tan, R.C. Veltkamp.)
  111. VIRAT Video Dataset – event recognition from two broad categories of activities (single-object and two-objects) which involve both human and vehicles. (Sangmin Oh et al)
  112. Verona Social interaction dataset (Marco Cristani)
  113. ViHASi: Virtual Human Action Silhouette Data (userID: VIHASI password: virtual$virtual) (Hossein Ragheb, Kingston University)
  114. Videoweb (multicamera) Activities Dataset (B. Bhanu, G. Denina, C. Ding, A. Ivers, A. Kamal, C. Ravishankar, A. Roy-Chowdhury, B. Varda)
  115. WVU Multi-view action recognition dataset (Univ. of West Virginia)
  116. WorkoutSU-10 Kinect dataset for exercise actions (Ceyhun Akgul)
  117. WorkoutSU-10 dataset – contains exercise actions selected by professional trainers for therapeutic purposes. (Sabanci University)
  118. Wrist-mounted camera video dataset – object manipulation (Ohnishi, Kanehira, Kanezaki, Harada)
  119. YouCook – 88 open-source YouTube cooking videos with annotations (Jason Corso)
  120. YouTube-8M Dataset -A Large and Diverse Labeled Video Dataset for Video Understanding Research(Google Inc.)

Attribute recognition

  1. Attribute Learning for Understanding Unstructured Social Activity – Database of videos containing 10 categories of unstructured social events to recognise, also annotated with 69 attributes. (Y. Fu Fudan/QMUL, T. Hospedales Edinburgh/QMUL)
  2. BirdsThis database contains 600 images (100 samples each) of six different classes of birds.(Svetlana Lazebnik, Cordelia Schmid, and Jean Ponce)
  3. ButterfliesThis database contains 619 images of seven different classes of butterflies. (Svetlana Lazebnik, Cordelia Schmid, and Jean Ponce)
  4. CALVIN research group datasets – object detection with eye tracking, imagenet bounding boxes, synchronised activities, stickman and body poses, youtube objects, faces, horses, toys, visual attributes, shape classes (CALVIN ggroup)
  5. CelebA – Large-scale CelebFaces Attributes Dataset(Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang)
  6. DukeMTMC-attribute – 23 pedestrian attributes for DukeMTMC-reID (Lin, Zheng, Zheng, Wu and Yang)
  7. HAT database of 27 human attributes (Gaurav Sharma, Frederic Jurie)
  8. LFW-10 dataset for learning relative attributes – A dataset of 10,000 pairs of face images with instance-level annotations for 10 attributes.(CVIT, IIIT Hyderabad. )
  9. Market-1501-attribute – 27 visual attributes for 1501 shoppers (Lin, Zheng, Zheng, Wu and Yang)
  10. Person Recognition in Personal Photo Collections – we introduced three harder splits for evaluation and long-term attribute annotations and per-photo timestamp metadata.(Oh, Seong Joon and Benenson, Rodrigo and Fritz, Mario and Schiele, Bernt)
  11. Visual Attributes Dataset visual attribute annotations for over 500 object classes (animate and inanimate) which are all represented in ImageNet. Each object class is annotated with visual attributes based on a taxonomy of 636 attributes (e.g., has fur, made of metal, is round).
  12. The Visual Privacy (VISPR) Dataset – Privacy Multilabel Dataset (22k images, 68 privacy attributes) (Orekondy, Schiele, Fritz)
  13. WIDER Attribute Dataset – WIDER Attribute is a large-scale human attribute dataset, with 13789 images belonging to 30 scene categories, and 57524 human bounding boxes each annotated with 14 binary attributes.(Li, Yining and Huang, Chen and Loy, Chen Change and Tang, Xiaoou)

Autonomous Driving

  1. AMUSE -The automotive multi-sensor (AMUSE) dataset taken in real traffic scenes during multiple test drives. (Philipp Koschorrek etc.)
  2. Autonomous Driving – semantic segmentation,pedestrian detection,virtual-world data,far infrared,stereo,driver monitoring(CVC research center and the UAB and UPC universities)
  3. Joint Attention in Autonomous Driving (JAAD) – The dataset includes instances of pedestrians and cars intended primarily for the purpose of behavioural studies and detection in the context of autonomous driving.(Iuliia Kotseruba, Amir Rasouli and John K. Tsotsos)
  4. LISA Vehicle Detection Dataset – colour first person driving video under various lighting and traffic conditions (Sivaraman, Trivedi)
  5. Lost and Found Dataset – The Lost and Found Dataset addresses the problem of detecting unexpected small road hazards (often caused by lost cargo) for autonomous driving applications. (Sebastian Ramos, Peter Pinggera, Stefan Gehrig, Uwe Franke, Rudolf Mester, Carsten Rother)
  6. SYNTHIA – Large set (~half million) of virtual-world images for training autonomous cars to see. (ADAS Group at Computer Vision Center)
  7. The SYNTHetic collection of Imagery and Annotations – the purpose of aiding semantic segmentation and related scene understanding problems in the context of driving scenarios (Computer vision center,UAB)

Biological/Medical

  1. 2008 MICCAI MS Lesion Segmentation Challenge (National Institutes of Health Blueprint for Neuroscience Research)
  2. ASU DR-AutoCC Data – a Multiple-Instance Learning feature space for a diabetic retinopathy classification dataset (Ragav Venkatesan, Parag Chandakkar, Baoxin Li – Arizona State University)
  3. Aberystwyth Leaf Evaluation Dataset – Timelapse plant images with hand marked up leaf-level segmentations for some time steps, and biological data from plant sacrifice. (Bell, Jonathan; Dee, Hannah M.)
  4. Annotated Spine CT Database for Benchmarking of Vertebrae Localization, 125 patients, 242 scans (Ben Glockern)
  5. BRATS – the identification and segmentation of tumor structures in multiparametric magnetic resonance images of the brain (TU Munchen etc.)
  6. Cholec80: 80 gallbladder laparoscopic videos annotated with phase and tool information. (Andru Putra Twinanda)
  7. CRCHistoPhenotypes – Labeled Cell Nuclei Data – colorectal cancer?histology images?consisting of nearly 30,000 dotted nuclei with over 22,000 labeled with the cell type (Rajpoot + Sirinukunwattana)
  8. CREMI: MICCAI 2016 Challenge – 6 volumes of electron microscopy of neural tissue,neuron and synapse segmentation, synaptic partner annotation. (Jan Funke, Stephan Saalfeld, Srini Turaga, Davi Bock, Eric Perlman)
  9. Cavy Action Dataset – 16 sequences with 640 x 480 resolutions recorded at 7.5 frames per second (fps) with approximately 31621506 frames in total (272 GB) of interacting cavies (guinea pig) (Al-Raziqi and Denzler)
  10. Cell Tracking Challenge Datasets – 2D/3D time-lapse video sequences with ground truth(Ma et al., Bioinformatics 30:1609-1617, 2014)
  11. Computed Tomography Emphysema Database (Lauge Sorensen)
  12. CRIM13 Caltech Resident-Intruder Mouse dataset – 237 10 minute videos (25 fps) annotated with actions (13 classes) (Burgos-Artizzu, Dollár, Lin, Anderson and Perona)
  13. DIADEM: Digital Reconstruction of Axonal and Dendritic Morphology Competition (Allen Institute for Brain Science et al)
  14. DIARETDB1 – Standard Diabetic Retinopathy Database (Lappeenranta Univ of Technology)
  15. DRIVE: Digital Retinal Images for Vessel Extraction (Univ of Utrecht)
  16. DeformIt 2.0 – Image Data Augmentation Tool: Simulate novel images with ground truth segmentations from a single image-segmentation pair (Brian Booth and Ghassan Hamarneh)
  17. Deformable Image Registration Lab dataset – for objective and rigrorous evaluation of deformable image registration (DIR) spatial accuracy performance. (Richard Castillo et al.)
  18. DERMOFIT Skin Cancer Dataset – 1300 lesions from 10 classes captured under identical controlled conditions. Lesion segmentation masks are included (Fisher, Rees, Aldridge, Ballerini, et al)
  19. Dermoscopy images (Eric Ehrsam)
  20. EPT29.This database contains 4842 images of 1613 specimens of 29 taxa of EPTs:(Tom etc.)
  21. FIRE Fundus Image Registration Dataset – 134 retinal image pairs and groud truth for registration.(FORTH-ICS)
  22. IIT MBADA mice – mice behavioral data. FLIR A315, spacial resolution of 320×240px at 30fps, 50x50cm open arena, two experts for three different mice pairs, mice identities (Italian Inst. of Technology, PAVIS lab)
  23. IRMA(Image retrieval in medical applications) – This collection compiles anonymous radiographs (Deserno TM, Ott B)
  24. KID – A capsule endoscopy database for medical decision support (Anastasios Koulaouzidis and Dimitris
    Iakovidis)
  25. Leaf Segmentation ChallengeTobacco and arabidopsis plant images (Hanno Scharr, Massimo Minervini, Andreas Fischbach, Sotirios A. Tsaftaris)
  26. LITS Liver Tumor Segmentation – 130 3D CT scans with segmentations of the liver and liver tumor. Public benchmark with leaderboard at Codalab.org (Patrick Christ)
  27. MIT CBCL Automated Mouse Behavior Recognition datasets (Nicholas Edelman)
  28. MUCIC: Masaryk University Cell Image Collection – 2D/3D synthetic images of cells/tissues for benchmarking(Masaryk University)
  29. MiniMammographic Database (Mammographic Image Analysis Society)
  30. Moth fine-grained recognition – 675 similar classes, 5344 images (Erik Rodner et al)
  31. Mouse Embryo Tracking Database – cell division event detection (Marcelo Cicconet, Kris Gunsalus)
  32. OASIS – Open Access Series of Imaging Studies – 500+ MRI data sets of the brain (Washington University, Harvard University, Biomedical Informatics Research Network)
  33. Plant Phenotyping Datasets – plant data suitable for plant and leaf detection, segmentation, tracking, and species recognition (M. Minervini, A. Fischbach, H. Scharr, S. A. Tsaftaris)
  34. RatSI: Rat Social Interaction Dataset – 9 fully annotated (11 class) videos (15 minute, 25 FPS) of two rats interacting socially in a cage (Malte Lorbach, Noldus Information Technology)
  35. Retinal fundus images – Ground truth of vascular bifurcations and crossovers (Univ of Groningen)
  36. SCORHE – 1, 2 and 3 mouse behavior videos, 9 behaviors, (Ghadi H. Salem, et al, NIH)
  37. STructured Analysis of the Retina – DESCRIPTION(400+ retinal images, with ground truth segmentations and medical annotations)
  38. Spine and Cardiac data (Digital Imaging Group of London Ontario, Shuo Li)
  39. Stonefly9This database contains 3826 images of 773 specimens of 9 taxa of Stoneflies (Tom etc.)
  40. Synthetic Migrating Cells -Six artificial migrating cells (neutrophils) over 98 time frames, various levels of Gaussian/Poisson noise and different paths characteristics with ground truth. (Dr Constantino Carlos Reyes-Aldasoro et al.)
  41. Univ of Central Florida – DDSM: Digital Database for Screening Mammography (Univ of Central Florida)
  42. VascuSynth – 120 3D vascular tree like structures with ground truth (Mengliu Zhao, Ghassan Hamarneh)
  43. VascuSynth – Vascular Synthesizer generates vascular trees in 3D volumes. (Ghassan Hamarneh, Preet Jassi, Mengliu Zhao)
  44. York Cardiac MRI dataset (Alexander Andreopoulos)

Camera calibration

  1. Catadioptric camera calibration images (Yalin Bastanlar)
  2. GoPro-Gyro Dataset – This dataset consists of a number of wide-angle rolling shutter video sequences with corresponding gyroscope measurements (Hannes etc.)
  3. LO-RANSAC – LO-RANSAC library for estimation of homography and epipolar geometry(K. Lebeda, J. Matas and O. Chum)

Face and Eye/Iris Databases

  1. 300 Videos in the Wild (300-VW) – 68 Facial Landmark Tracking (Chrysos, Antonakos, Zafeiriou, Snape, Shen, Kossaifi, Tzimiropoulos, Pantic)
  2. 3D Mask Attack Database (3DMAD) – 76500 frames of 17 persons using Kinect RGBD with eye positions (Sebastien Marcel)
  3. 3D facial expression – Binghamton University 3D Static and Dynamic Facial Expression Databases (Lijun Yin, Jeff Cohn, and teammates)
  4. Audio-visual database for face and speaker recognition (Mobile Biometry MOBIO http://www.mobioproject.org/)
  5. BANCA face and voice database (Univ of Surrey)
  6. Binghampton Univ 3D static and dynamic facial expression database (Lijun Yin, Peter Gerhardstein and teammates)
  7. Binghamton-Pittsburgh 4D Spontaneous Facial Expression Database – consist of 2D spontaneous facial expression videos and FACS codes. (Lijun Yin et al.)
  8. BioID face database (BioID group)
  9. BioVid Heat Pain Database – This video (and biomedical signal) dataset contains facial and physiopsychological reactions of 87 study participants who were subjected to experimentally induced heat pain.(University of Magdeburg (Neuro-Information Technology group) and University of Ulm (Emotion Lab))
  10. Biometric databases – biometric databases related to iris recognition (Adam Czajka)
  11. Biwi 3D Audiovisual Corpus of Affective Communication – 1000 high quality, dynamic 3D scans of faces, recorded while pronouncing a set of English sentences.
  12. Bosphorus 3D/2D Database of FACS annotated facial expressions, of head poses and of face occlusions (Bogazici University)
  13. Caricature/Photomates dataset – a dataset with frontal faces and corresponding Caricature line drawings (Tayfun Akgul)
  14. CASIA-IrisV3 (Chinese Academy of Sciences, T. N. Tan, Z. Sun)
  15. CASIR Gaze Estimation Database – RGB and depth images (from Kinect V1.0) and ground truth values of facial features corresponding to experiments for gaze estimation benchmarking: (Filipe Ferreira etc.)
  16. CMU Facial Expression Database (CMU/MIT)
  17. The CMU Multi-PIE Face Database – more than 750,000 images of 337 people recorded in up to four sessions over the span of five months. (Jeff Cohn et al.)
  18. CMU Pose, Illumination, and Expression (PIE) Database (Simon Baker)
  19. CMU/MIT Frontal Faces (CMU/MIT)
  20. CMU/MIT Frontal Faces (CMU/MIT)
  21. CSSE Frontal intensity and range images of faces (Ajmal Mian)
  22. CelebA – Large-scale CelebFaces Attributes Dataset(Ziwei Liu, Ping Luo, Xiaogang Wang, Xiaoou Tang)
  23. Cohn-Kanade AU-Coded Expression Database – 500+ expression sequences of 100+ subjects, coded by activated Action Units (Affect Analysis Group, Univ. of Pittsburgh)
  24. Cohn-Kanade AU-Coded Expression Database – for research in automatic facial image analysis and synthesis and for perceptual studies (Jeff Cohn et al.)
  25. Columbia Gaze Data Set – 5,880 images of 56 people over 5 head poses and 21 gaze directions (Brian A. Smith, Qi Yin, Steven K. Feiner, Shree K. Nayar)
  26. Computer Vision Laboratory Face Database (CVL Face Database) – Database contains 798 images of 114 persons, with 7 images per person and is freely available for research purposes.(Peter Peer etc.)
  27. DISFA+:Extended Denver Intensity of Spontaneous Facial Action Database – an extension of DISFA (M.H. Mahoor)
  28. DISFA:Denver Intensity of Spontaneous Facial Action Database – a non-posed facial expression database for those who are interested in developing computer algorithms for automatic action unit detection and their intensities described by FACS. (M.H. Mahoor)
  29. EURECOM Facial Cosmetics Database – 389 images, 50 persons with/without make-up, annotations about the
    amount and location of applied makeup.(Jean-Luc DUGELAY et al)
  30. EURECOM Kinect Face Database – 52 people, 2 sessions, 9 variations, 6 facial landmarks.(Jean-Luc DUGELAY et al)
  31. EYEDIAP dataset – The EYEDIAP dataset was designed to train and evaluate gaze estimation algorithms from RGB and RGB-D data.It contains a diversity of participants, head poses, gaze targets and sensing conditions.(Kenneth Funes and Jean-Marc Odobez)
  32. FDDB: Face Detection Data set and Benchmark – studying unconstrained face detection (University of Massachusetts Computer Vision Laboratory)
  33. FG-Net Aging Database of faces at different ages (Face and Gesture Recognition Research Network)
  34. Face Recognition Grand Challenge datasets (FRVT – Face Recognition Vendor Test)
  35. Face Super-Resolution Dataset – Ground truth HR-LR face images captured with a dual-camera setup(Chengchao Qu etc.)
  36. FaceScrub – A Dataset With Over 100,000 Face Images of 530 People (50:50 male and female) (H.-W. Ng, S. Winkler)
  37. FaceTracer Database – 15,000 faces (Neeraj Kumar, P. N. Belhumeur, and S. K. Nayar)
  38. Facial Expression Dataset – This dataset consists of 242 facial videos (168,359 frames) recorded in real world conditions. (Daniel McDuff et al.)
  39. Florence 2D/3D Hybrid Face Dataset – bridges the gap between 2D, appearance-based recognition techniques, and fully 3D approaches (Bagdanov, Del Bimbo, and Masi)
  40. Facial Recognition Technology (FERET) Database (USA National Institute of Standards and Technology)
  41. Gi4E Database – eye-tracking database with 1300+ images acquired with a standard webcam, corresponding to different subjects gazing at different points on a screen, including ground-truth 2D iris and corner points (Villanueva, Ponz, Sesma-Sanchez, Mikel Porta, and Cabeza)
  42. Hannah and her sisters database – a dense audio-visual person-oriented ground-truth annotation of faces, speech segments, shot boundaries (Patrick Perez, Technicolor)
  43. Hong Kong Face Sketch Database
  44. IDIAP Head Pose Database (IHPD) – The dataset contains a set of meeting videos along with the head groundtruth of individual participants (around 128min)(Sileye Ba and Jean-Marc Odobez)
  45. IMDB-WIKI – 500k+ face images with age and gender labels (Rasmus Rothe, Radu Timofte, Luc Van Gool )
  46. Indian Movie Face database (IMFDB) – a large unconstrained face database consisting of 34512 images of 100 Indian actors collected from more than 100 videos (Vijay Kumar and C V Jawahar)
  47. Iranian Face Database – IFDB is the first image database in middle-east, contains color facial images with age, pose, and expression whose subjects are in the range of 2-85. (Mohammad Mahdi Dehshibi)
  48. Japanese Female Facial Expression (JAFFE) Database (Michael J. Lyons)
  49. LFW: Labeled Faces in the Wild – unconstrained face recognition
  50. Makeup Induced Face Spoofing (MIFS) – 107 makeup-transformations attempting to spoof a target identity. Also other datasets. (Antitza Dantcheva)
  51. Mexculture142 – Mexican Cultural heritage objects and eye-tracker gaze fixations (Montoya Obeso, Benois-Pineau, Garcia-Vazquez, Ramirez Acosta)
  52. MIT CBCL Face Recognition Database (Center for Biological and Computational Learning)
  53. MIT Collation of Face Databases (Ethan Meyers)
  54. MIT eye tracking database (1003 images) (Judd et al)
  55. MMI Facial Expression Database – 2900 videos and high-resolution still images of 75 subjects, annotated for FACS AUs.
  56. MORPH (Craniofacial Longitudinal Morphological Face Database) (University of North Carolina Wilmington)
  57. MPIIGaze dataset – 213,659 samples with eye images and gaze target under different illumination conditions and nature head movement, collected from 15 participants with their laptop during daily using. (Xucong Zhang, Yusuke Sugano, Mario Fritz, Andreas Bulling.)
  58. Manchester Annotated Talking Face Video Dataset (Timothy Cootes)
  59. MegaFace – 1 million faces in bounding boxes (Kemelmacher-Shlizerman, Seitz, Nech, Miller, Brossard)
  60. Music video dataset – 8 music videos from YouTube for developing multi-face tracking algorithms in unconstrained environments (Shun Zhang, Jia-Bin Huang, Ming-Hsuan Yang)
  61. NIST Face Recognition Grand Challenge (FRGC) (NIST)
  62. NIST mugshot identification database (USA National Institute of Standards and Technology)
  63. NRC-IIT Facial Video Database – this database contains pairs of short video clips each showing a face of a computer user sitting in front of the monitor exhibiting a wide range of facial expressions and orientations (Dmitry Gorodnichy)
  64. Notre Dame Iris Image Dataset (Patrick J. Flynn)
  65. Notre Dame face, IR face, 3D face, expression, crowd, and eye biometric datasets (Notre Dame)
  66. ORL face database: 40 people with 10 views (ATT Cambridge Labs)
  67. OUI-Adience Faces – unfiltered faces for gender and age classification plus 3D faces (OUI)
  68. Oxford: faces, flowers, multi-view, buildings, object categories, motion segmentation, affine covariant regions, misc (Oxford Visual Geometry Group)
  69. Pandora – POSEidon: Face-from-Depth for Driver Pose (Borghi, Venturelli, Vezzani, Cucchiara)
  70. PubFig: Public Figures Face Database (Neeraj Kumar, Alexander C. Berg, Peter N. Belhumeur, and Shree K. Nayar)
  71. Re-labeled Faces in the Wild – original images, but aligned using “deep funneling” method. (University of Massachusetts, Amherst)
  72. SCface – Surveillance Cameras Face Database (Mislav Grgic, Kresimir Delac, Sonja Grgic, Bozidar Klimpak)
  73. Salient features in gaze-aligned recordings of human visual input – TB of human gaze-contingent data “in the wild” (Frank Schumann etc.)
  74. SiblingsDB – The SiblingsDB contains two datasets depicting images of individuals related by sibling relationships. (Politecnico di Torino/Computer Graphics & Vision Group)
  75. Spontaneous Emotion Multimodal Database (SEM-db) – non-posed reactions to visual stimulus data recorded with HD RGB, depth and IR frames of the face, EEG signal and eye gaze data (Fernandez Montenegro, Gkelias, Argyriou)
  76. The UNBC-McMaster Shoulder Pain Expression Archive Database – Painful data: The UNBC-McMaster Shoulder Pain Expression Archive Database (Lucy et al.)
  77. Trondheim Kinect RGB-D Person Re-identification Dataset (Igor Barros Barbosa)
  78. UB KinFace Database – University of Buffalo kinship verification and recognition database
  79. UBIRIS: Noisy Visible Wavelength Iris Image Databases (University of Beira)
  80. UPNA Head Pose Database – head pose database, with 120 webcam videos containing guided-movement sequences and free-movement sequences, including ground-truth head pose and automatically annotated 2D facial points. (Ariz, Bengoechea, Villanueva, Cabeza)
  81. UPNA Synthetic Head Pose Database – a synthetic replica of the UPNA Head Pose Database, with 120 videos with their 2D ground truth landmarks projections, their corresponding head pose ground truth, 3D head models and camera parameters. (Larumbe, Segura, Ariz, Bengoechea, Villanueva, Cabeza)
  82. UTIRIS cross-spectral iris image databank (Mahdi Hosseini)
  83. VIPSL Database – VIPSL Database is for research on face sketch-photo synthesis and recognition, including 200 subjects (1 photo and 5 sketches per subject).(Nannan Wang)
  84. VT-KFER: A Kinect-based RGBD+Time Dataset for Spontaneous and Non-Spontaneous Facial Expression Recognition – 32 subjects, 1,956 sequences of RGBD, six facial expressions in 3 poses (Aly, Trubanova, Abbott, White, and Youssef)
  85. Washington Facial Expression Database (FERG-DB) – a database of 6 stylized (Maya) characters with 7 annotated facial expressions (Deepali Aneja, Alex Colburn, Gary Faigin, Linda Shapiro, and Barbara Mones)
  86. WIDER FACE: A Face Detection Benchmark – 32,203 images with 393,703 labeled faces, 61 event classes (Shuo Yang, Ping Luo, Chen Change Loy, Xiaoou Tang)
  87. XM2VTS Face video sequences (295): The extended M2VTS Database (XM2VTS) – (Surrey University)
  88. Yale Face Database – 11 expressions of 10 people (A. Georghaides)
  89. Yale Face Database B – 576 viewing conditions of 10 people (A. Georghaides)
  90. York Univ Eye Tracking Dataset (120 images) (Neil Bruce)
  91. YouTube Faces DB – 3,425 videos of 1,595 different people. (Wolf, Hassner, Maoz)
  92. Zurich Natural Image – the image material used for creating natural stimuli in a series of eye-tracking studies (Frey et al.)

Fingerprints

  1. FVC fingerpring verification competition 2002 dataset (University of Bologna)
  2. FVC fingerpring verification competition 2004 dataset (University of Bologna)
  3. Fingerprint Manual Minutiae Marker (FM3) Databases: – Fingerprint Manual Minutiae Marker (FM3) Databases( Mehmet Kayaoglu, Berkay Topcu and Umut Uludag)
  4. NIST fingerprint databases (USA National Institute of Standards and Technology)
  5. SPD2010 Fingerprint Singular Points Detection Competition (SPD 2010 committee)

General Images

  1. A Dataset for Real Low-Light Image Noise Reduction – It contains pixel and intensity aligned pairs of images corrupted by low-light camera noise and their low-noise counterparts. (J. Anaya, A. Barbu)
  2. A database of paintings related to Vincent van Gogh – This is the dataset VGDB-2016 built for the paper “From Impressionism to Expressionism: Automatically Identifying Van Gogh’s Paintings” (Guilherme Folego and Otavio Gomes and Anderson Rocha)
  3. AMOS: Archive of Many Outdoor Scenes (20+m) (Nathan Jacobs)
  4. Aerial images Building detection from aerial images using invariant color features and shadow information (Beril Sirmacek)
  5. AutoDA (Automatic Dataset Augmentation) – An automatically constructed image dataset including 12.5 million images with relevant textual information for the 1000 categories of ILSVRC2012 (Bai, Yang, Ma, Zhao)
  6. BGU Hyperspectral Image Database of Natural Scenes (Ohad Ben-Shahar and Boaz Arad)
  7. Brown Univ Large Binary Image Database (Ben Kimia)
  8. CMP Facade Database – Includes 606 rectified images of facades from various places with 12 architectural classes annotated.(Radim Tylecek)
  9. Caltech-UCSD Birds-200-2011 (Catherine Wah)
  10. Columbia Multispectral Image Database (F. Yasuma, T. Mitsunaga, D. Iso, and S.K. Nayar)
  11. DAQUAR (Visual Turing Challenge) – A dataset containing questions and answers about real-world indoor scenes.(Mateusz Malinowski, Mario Fritz)
  12. Darmstadt Noise Dataset – 50 pairs of real noisy images and corresponding ground truth images (RAW and sRGB) (Tobias Plotz and Stefan Roth)
  13. Dataset of American Movie Trailers 2010-2014 – Contains links to 474 hollywood movie
    trailers along with associated metadata (genre, budget, runtime, release, MPAA rating, screens released, sequel indicator) (USC Signal Analysis and Interpretation Lab)
  14. DSLR Photo Enhancement Dataset (DPED) – 22K photos taken synchronously in the wild by three smartphones and one DSLR camera, useful for comparing infered high quality images from multiple low quality images (Ignatov, Kobyshev, Timofte, Vanhoey, and Van Gool).
  15. Flickr-style – 80K Flickr photographs annotated with 20 curated style labels, and 85K paintings annotated with 25 style/genre labels (Sergey Karayev)
  16. Forth Multispectral Imaging Datasets – images from 23 spectral bands each from 5 paintings. Images are annotated with ground truth data. (Karamaoynas Polykarpos et al)
  17. General 100 Dataset – General-100 dataset contains 100 bmp-format images (with no
    compression), which are well-suited for super-resolution training(Dong, Chao and Loy, Chen Change and Tang, Xiaoou)
  18. GOPRO dataset – Blurred image dataset with sharp image ground truth (Nah, Kim, and Lee)
  19. HIPR2 Image Catalogue of different types of images (Bob Fisher et al)
  20. HPatches – A benchmark and evaluation of handcrafted and learned local descriptors (Balntas, Lenc, Vedaldi, Mikolajczyk)
  21. Hyperspectral images for spatial distributions of local illumination in natural scenes – Thirty calibrated hyperspectral radiance images of natural scenes with probe spheres embedded for local illumination estimation. (Nascimento, Amano & Foster)
  22. Hyperspectral images of natural scenes – 2002 (David H. Foster)
  23. Hyperspectral images of natural scenes – 2004 (David H. Foster)
  24. ISPRS multi-platform photogrammetry dataset – 1: Nadir and oblique aerial images plus 2: Combined UAV and terrestrial images (Francesco Nex and Markus Gerke)
  25. Image & Video Quality Assessment at LIVE – used to develop picture quality algorithms (the University of Texas at Austin)
  26. ImageNet Large Scale Visual Recognition Challenges
    – currently 200 object classes and 500+K images (Alex Berg, Jia Deng, Fei-Fei Li and others)
  27. ImageNet Linguistically organised (WordNet) Hierarchical Image Database – 10E7 images, 15K categories (Li Fei-Fei, Jia Deng, Hao Su, Kai Li)
  28. Improved 3D Sparse Maps for High-performance Structure from Motion with Low-cost Omnidirectional Robots – Evaluation Dataset – Data set used in research paper doi:10.1109/ICIP.2015.7351744 (Breckon, Toby P., Cavestany, Pedro)
  29. LabelMeFacade Database – 945 labeled building images (Erik Rodner et al)
  30. Local illumination hyperspectral radiance images – Thirty hyperspectral radiance images of natural scenes with embedded probe spheres for local illumination estimates(Sgio M. C. Nascimento, Kinjiro Amano, David H. Foster)
  31. McGill Calibrated Colour Image Database (Adriana Olmos and Fred Kingdom)
  32. Multiply Distorted Image Database -a database for evaluating the results of image quality assessment metrics on multiply distorted images.(Fei Zhou)
  33. NPRgeneral – A standardized collection of images for evaluating image stylization algorithms. (David Mould, Paul Rosin)
  34. NYU Symmetry Database – 176 single-symmetry and 63 multyple-symmetry images (Marcelo Cicconet and Davi Geiger)
  35. OTCBVS Thermal Imagery Benchmark Dataset Collection (Ohio State Team)
  36. PAnorama Sparsely STructured Areas Datasets – the PASSTA datasets used for evaluation of the image alignment (Andreas Robinson)
  37. See in the Dark – 77 Gb of dark images (Chen, Chen, Xu, and Koltun)
  38. Stanford Street View Image, Pose, and 3D Cities Dataset – a large scale dataset of street view images (25 million images and 118 matching image pairs) with their relative camera pose, 3D models of cities, and 3D metadata of images. (Zamir, Wekel, Agrawal, Malik, Savarese)
  39. TESTIMAGES – huge and free collection of sample images designed for analysis and quality assessment of different kinds of displays (i.e. monitors, televisions and digital cinema projectors) and image processing techniques. (Nicola Asuni)
  40. Time-Lapse Hyperspectral Radiance Images of Natural Scenes – Four time-lapse sequences of 7-9 calibrated hyperspectral radiance images of natural scenes taken over the day. (Foster, D.H., Amano, K., & Nascimento, S.M.C.)
  41. Time-lapse hyperspectral radiance images – Four time-lapse sequences of 7-9 calibrated hyperspectral images of natural scenes, spectra at 10-nm intervals(David H. Foster, Kinjiro Amano, Sgio M. C. Nascimento)
  42. Tiny Images Dataset 79 million 32×32 color images (Fergus, Torralba, Freeman)
  43. Visual Dialog – 120k human-human dialogs on COCO images, 10 rounds of QA per dialog (Das, Kottur, Gupta, Singh, Yadav, Moura, Parikh, Batra)
  44. Visual Question Answering – 254K imags, 764K questions, ground truth (Agrawal, Lu, Antol, Mitchell, Zitnick, Batra, Parikh)
  45. Visual Question Generation – 15k images (including both object-centric and event-centric images), 75k natural questions asked about the images which can evoke further conversation(Nasrin Mostafazadeh, Ishan Misra, Jacob Devlin, Margaret Mitchell, Xiao dong He, Lucy Vanderwende)
  46. VQA Human Attention – 60k human attention maps for visual question answering i.e. where humans choose to look to answer questions about images (Das, Agrawal, Zitnick, Parikh, Batra)
  47. Wild Web tampered image dataset – A large collection of tampered images from Web and social media sources, including ground-truth annotation masks for tampering localization (Markos Zampoglou, Symeon Papadopoulos)
  48. YFCC100M: The New Data in Multimedia Research – This publicly available curated dataset of 100 million photos and videos is free and legal for all.(Bart Thomee, Yahoo Labs and Flickr in San Francisco,etc.)

General RGBD and Depth Datasets

Note: there are 3D datasets elsewhere as well, e.g. in
Objects, Scenes, and Actions.

  1. 3D-Printed RGB-D Object Dataset – 5 objects with groundtruth CAD models and camera trajectories, recorded with various quality RGB-D sensors(Siemens & TUM)
  2. 3DCOMET – 3DCOMET is a dataset for testing 3D data compression methods.(Miguel Cazorla, Javier Navarrete,Vicente Morell, Miguel Cazorla, Diego Viejo, Jose Garcia-Rodriguez, Sergio Orts.)
  3. A Dataset for Non-Rigid Reconstruction from RGB-D Data – Eight scenes for reconstructing non-rigid geometry from RGB-D data, each containing several hundred frames along with our results. (Matthias Innmann, Michael Zollhoefer, Matthias Niessner, Christian Theobalt, Marc Stamminger)
  4. A Large Dataset of Object Scans – 392 objects in 9 casses, hundreds of frames each (Choi, Zhou, Miller, Koltun)
  5. Articulated Object Challenge – 4 articulated objects consisting of rigids parts connected by 1D revolute and prismatic joints, 7000+ RGBD images with annotations for 6D pose estimation(Frank Michel, Alexander Krull, Eric Brachmann, Michael. Y. Yang,Stefan Gumhold, Carsten Rother)
  6. BigBIRD – 100 objects with for each object, 600 3D point clouds and 600 high-resolution color images spanning all views (Singh, Sha, Narayan, Achim, Abbeel)
  7. CAESAR Civilian American and European Surface Anthropometry Resource Project – 4000 3D human body scans (SAE International)
  8. CIN 2D+3D object classification dataset – segmented color and depth images of objects from 18 categories of common household and office objects (Bjorn Browatzki et al)
  9. CoRBS – an RGB-D SLAM benchmark, providing the combination of real depth and color data together with a ground truth trajectory of the camera and a ground truth 3D model of the scene (Oliver Wasenmuller)
  10. CTU Garment Folding Photo Dataset – Color and depth images from various stages of garment folding.(Sushkov R., Melkumov I., Smutn y V. (Czech Technical University in Prague))
  11. CTU Garment Sorting Dataset – Dataset of garment images, detailed stereo images, depth images and weights.(Petrik V., Wagner L. (Czech Technical University in Prague))
  12. Clothing part dataset – The clothing part dataset consists of image and depth scans, acquired with a Kinect, of garments laying on a table, with over a thousand part annotations (collar, cuffs, hood, etc) using polygonal masks.(Arnau Ramisa, Guillem Aleny, Francesc Moreno-Noguer and Carme Torras)
  13. Cornell-RGBD-Dataset – Office Scenes (Hema Koppula)
  14. Delft Windmill Interior and Exterior Laser Scanning Point Clouds (Beril Sirmacek)
  15. Diabetes60 – RGB-D images of 60 western dishes, home made. Data was recorded using a Microsoft Kinect V2. (Patrick Christ and Sebastian Schlecht)
  16. EURECOM Kinect Face Database – 52 people, 2 sessions, 9 variations, 6 facial landmarks.(Jean-Luc DUGELAY et al)
  17. EURECOM Kinect Face Database – 52 people, 2 sessions, 9 variations, 6 facial landmarks.(Jean-Luc DUGELAY et al)
  18. G4S meta rooms – RGB-D data 150 sweeps with 18 images per sweep. (John Folkesson et al.)
  19. Georgiatech-Metz Symphony Lake Dataset – 5 million RGBD outdoor images over 4 years from 121 surveys of a lakeshore. (Griffith and Pradalier)
  20. Goldfinch: GOogLe image-search Dataset for FINe grained CHallenges – a largescale dataset for finegrained bird (11K species),butterfly (14K species), aircraft (409 types), and dog (515 breeds) recognition.(Jonathan Krause, Benjamin Sapp, Andrew Howard, Howard Zhou, Alexander Toshev, Tom Duerig, James Philbin, Li Fei-Fei)
  21. IMPART multi-view/multi-modal 2D+3D film production dataset – LIDAR, video, 3D models, spherical camera, RGBD, stereo, action, facial expressions, etc. (Univ. of Surrey)
  22. Industrial 3D Object Detection Dataset (MVTec ITODD) – depth and gray value data of 28 objects in 3500 labeled scenes for 3D object detection and pose estimation with a strong focus on industrial settings and applications (MVTec Software GmbH, Munich)
  23. Kinect v2 Dataset – Efficient Multi-Frequency Phase Unwrapping using Kernel Density Estimation (Felix etc.)
  24. McGill-Reparti Artificial Perception Database – RGBD data from four cameras and unfiltered Vicon skeletal data of two human subjects performing simulated assembly tasks on a car door (Andrew Phan, Olivier St-Martin Cormier, Denis Ouellet, Frank P. Ferrie).
  25. Meta rooms – RGB-D data comprised of 28 aligned depth camera images collected by having robot go to specific place and do 360 degrees of pan with various tilts. (John Folkesson et al.)
  26. MHT RGB-D – collected by a robot every 5 min over 16 days by the University of Lincoln. (John Folkesson et al.)
  27. Multi-sensor 3D Object Dataset for Object Recognition with Full Pose Estimation – Multi-sensor 3D Object Dataset for Object Recognition and Pose Estimation(Alberto Garcia-Garcia, Sergio Orts-Escolano, Sergiu Oprea,etc.)
  28. NTU RGB+D Action Recognition Dataset – NTU RGB+D is a large scale dataset for human action recognition(Amir Shahroudy)
  29. NYU Depth Dataset V2 – Indoor Segmentation and Support Inference from RGBD Images
  30. Oakland 3-D Point Cloud Dataset (Nicolas Vandapel)
  31. Pacman project – Synthetic RGB-D images of 400 objects from 20 classes. Generated from 3D mesh models (Vladislav Kramarev, Umit Rusen Aktas, Jeremy L. Wyatt.)
  32. RGB-D Part Affordance Dataset – RGB-D images and ground-truth affordance labels for 105 kitchen, workshop and garden tools, and 3 cluttered scenes (Myers, Teo, Fermuller, Aloimonos)
  33. SceneNN: A Scene Meshes Dataset with aNNotations – RGB-D scene dataset with 100+ indoor scenes, labeled triangular mesh, voxel and pixel. (Hua, Pham, Nguyen, Tran, Yu, and Yeung)
  34. Semantic-8: 3D point cloud classification with 8 classes (ETH Zurich)
  35. Small office data sets – Kinect depth images every 5 seconds beginning in April 2014 and on-going. (John Folkesson et al.)
  36. Stereo and ToF dataset with ground truth – The dataset contains 5 different scenes acquired with a Time-of-flight sensor and a stereo setup. Ground truth information is also provided.(Carlo Dal Mutto, Pietro Zanuttigh, Guido M. Cortelazzo)
  37. SYNTHIA – Large set (~half million) of virtual-world images for training autonomous cars to see. (ADAS Group at Computer Vision Center)
  38. TUM RGB-D Benchmark – Dataset and benchmark for the evaluation of RGB-D visual
    odometry and SLAM algorithms (Jorgen Sturm, Nikolas Engelhard, Felix Endres, Wolfram Burgard and Daniel Cremers)
  39. UC-3D Motion Database – Available data types encompass high resolution Motion Capture, acquired with MVN Suit from Xsens and Microsoft Kinect RGB and depth images.(Institute of Systems and Robotics, Coimbra, Portugal)
  40. USF Range Image Database – 400+ laser range finder and structured light camera images, many with ground truth segmentations (Adam et al.)
  41. Washington RGB-D Object Dataset – 300 common household objects and 14 scenes. (University of Washington and Intel Labs Seattle)
  42. Witham Wharf – For RGB-D of eight locations collect by robot every 10 min over ~10 days by the University of Lincoln. (John Folkesson et al.)

General Videos

  1. AlignMNIST – An artificially extended version of the MNIST handwritten
    dataset. (en Hauberg)
  2. DAVIS: Video Object Segmentation dataset 2016 – A Benchmark Dataset and Evaluation Methodology for Video Object Segmentation (F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross, and A. Sorkine-Hornung)
  3. DAVIS: Video Object Segmentation dataset 2017 – The 2017 DAVIS Challenge on Video Object Segmentation (J. Pont-Tuset, F. Perazzi, S. Caelles, P. Arbelaez, A. Sorkine-Hornung, and L. Van Gool)
  4. GoPro-Gyro Dataset – ego centric videos (Linkoping Computer Vision Laboratory)
  5. Image & Video Quality Assessment at LIVE – used to develop picture quality algorithms (the University of Texas at Austin)
  6. Large scale YouTube video dataset – 156,823 videos (2,907,447 keyframes) crawled from YouTube videos (Yi Yang)
  7. Movie Memorability Dataset – memorable movie clips and ground truth of detail memorability, 660 short movie excerpts extracted from 100 Hollywood-like movies (Cohendet, Yadati, Duong and Demarty)
  8. MovieQA – each machines to understand stories by answering questions about them. 15000 multiple choice QAs, 400+ movies.(M. Tapaswi, Y. Zhu, R. Stiefelhagen, A. Torralba, R. Urtasun, and S. Fidler)
  9. Multispectral visible-NIR video sequences – Annotated multispectral video, visible + NIR (LE2I, Universit de Bourgogne)
  10. Sports-1M – Dataset for sports video classification containing 487 classes and 1.2M videos.(Andrej Karpathy and George Toderici and Sanketh Shetty and Thomas Leung and Rahul Sukthankar and Li Fei-Fei.)
  11. Video Sequencesused for research on Euclidean upgrades based on minimal assumptions about the camera(Kenton McHenry)
  12. Video Stacking Dataset – A Virtual Tripod for Hand-held Video Stacking on Smartphones (Erik Ringaby etc.)
  13. YFCC100M videos – A benchmark on the video subset of YFCC100M which includes the videos, he video content features and the API to a sate-of-the-art video content engine.(Lu Jiang)
  14. YFCC100M: The New Data in Multimedia Research – This publicly available curated dataset of 100 million photos and videos is free and legal for all.(Bart Thomee, Yahoo Labs and Flickr in San Francisco,etc.)
  15. YouTube-BoundingBoxes – 5.6 million accurate human-annotated BB from 23 object classes tracked across frames, from 240,000 YouTube videos, with a strong focus on the person class (1.3 million boxes) (Real, Shlens, Pan, Mazzocchi, Vanhoucke, Khan, Kakarla et al)
  16. YouTube-8M – Dataset for video classification in the wild, containing pre-extracted frame level features from 8M videos, and 4800 classes.(Sami Abu-El-Haija, Nisarg Kothari, Joonseok Lee, Paul Natsev,George Toderici, Balakrishnan Varadarajan, Sudheendra Vijayanarasimhan)

Hand, Hand Grasp, Hand Action and Gesture Databases

  1. 11k Hands – 11,076 hand images (1600 x 1200 pixels) of 190 subjects, of varying ages between 18 – 75, with metadata (id, gender, age, skin color, handedness, which hand, accessories, etc). (Mahmoud Afifi)
  2. 20bn-Jester – densely-labeled video clips that show humans performing predefined hand gestures in front of a laptop camera or webcam (Twenty Billion Neurons GmbH)
  3. 3D Articulated Hand Pose Estimation with Single Depth Images (Tang, Chang, Tejani, Kim, Yu)
  4. A Dataset of Human Manipulation Actions – RGB-D of 25 objects and 6 actions (Alessandro Pieropan)
  5. A Hand Gesture Detection Dataset (Javier Molina et al)
  6. A-STAR Annotated Hand-Depth Image Dataset and its Performance Evaluation – depth data and data glove data, 29 images of 30 volunteers, Chinese number counting and American Sign Language (Xu and Cheng)
  7. Bosphorus Hand Geometry Database and Hand-Vein Database (Bogazici University)
  8. EgoGesture Dataset – First-person view gestures with 83 classes, 50 subjects, 6 scenes, 24161 RGB-D video samples (Zhang, Cao, Cheng, Lu)
  9. EgoHands – A large dataset with over 15,000 pixel-level-segmented hands recorded from egocentric cameras of people interacting with each other. (Sven Bambach)
  10. FORTH Hand tracking library (FORTH)
  11. General HANDS: general hand detection and pose challenge – 22 sequences with different gestures, activities and viewpoints (UC Irvine)
  12. Grasp UNderstanding (GUN-71) dataset – 12,000 first-person RGB-D images of object manipulation scenes annotated using a taxonomy of 71 fine-grained grasps.(Rogez, Supancic and Ramanan)
  13. Hand gesture and marine silhouettes (Euripides G.M. Petrakis)
  14. HandNet: annotated depth images of articulated hands 214971 annotated depth images of hands captured by a RealSense RGBD sensor of hand poses. Annotations: per pixel classes, 6D fingertip pose, heatmap. Train: 202198, Test: 10000, Validation: 2773. Recorded at GIP Lab, Technion.
  15. IDIAP Hand pose/gesture datasets (Sebastien Marcel)
  16. Kinect and Leap motion gesture recognition dataset – The dataset contains 1400 different gestures acquired with both the Leap Motion and the Kinect devices(Giulio Marin, Fabio Dominio, Pietro Zanuttigh)
  17. Kinect and Leap motion gesture recognition dataset – The dataset contains several different static gestures acquired with the Creative Senz3D camera.(A. Memo, L. Minto, P. Zanuttigh)
  18. LISA CVRR-HANDS 3D – 19 gestures performed by 8 subjects as car driver and passengers (Ohn-Bar and Trivedi)
  19. LISA Vehicle Detection Dataset – colour first person driving video under various lighting and traffic conditions (Sivaraman, Trivedi)
  20. MPI Dexter 1 Dataset for Evaluation of 3D Articulated Hand Motion Tracking – Dexter 1: 7 sequences of challenging, slow and fast hand motions, RGB + depth (Sridhar, Oulasvirta, Theobalt)
  21. MSR Realtime and Robust Hand Tracking from Depth – (Qian, Sun, Wei, Tang, Sun)
  22. Mobile and Webcam Hand images database – MOHI and WEHI – 200 people, 30 images each (Ahmad Hassanat)
  23. NYU Hand Pose Dataset – 8252 test-set and 72757 training-set frames of captured RGBD data with ground-truth hand-pose, 3 views (Tompson, Stein, Lecun, Perlin}
  24. PRAXIS gesture dataset – RGB-D upper-body data from 29 gestures, 64 volunteers, several repetitions, many volunteers have some cognitive impairment (Farhood Negin, INRIA)
  25. Rendered Handpose Dataset – Synthetic dataset for 2D/ 3D Handpose Estimation with RGB, depth, segmentation masks and 21 keypoints per hand (Christian Zimmermann and Thomas Brox)
  26. Sahand Dynamic Hand Gesture Database – This database contains 11 Dynamic gestures designed to convey the functions of mouse and touch screens to computers.(Behnam Maleki, Hossein Ebrahimnezhad)
  27. Sheffield gesture database – 2160 RGBD hand gesture sequences, 6 subjects, 10 gestures, 3 postures, 3 backgrounds, 2 illuminations (Ling Shao)
  28. UT Grasp Data Set – 4 subjects grasping a variety of objectss with a variety of grasps (Cai, Kitani, Sato)
  29. Yale human grasping data set – 27 hours of video with tagged grasp, object, and task data from two housekeepers and two machinists (Bullock, Feix, Dollar)

Image, Video and Shape Database Retrieval

  1. ANN_SIFT1M – 1M Flickr images encoded by 128D SIFT descriptors (Jegou et al)
  2. Brown Univ 25/99/216 Shape Databases (Ben Kimia)
  3. CIFAR-10 – 60K 32×32 images from 10 classes, with a 512D GIST descriptor (Alex Krizhevsky)
  4. CLEF-IP 2011 evaluation on patent images
  5. DeepFashion – Large-scale Fashion Database(Ziwei Liu, Ping Luo, Shi Qiu, Xiaogang Wang, Xiaoou Tang)
  6. EMODB – Thumbnails of images in the picsearch image search engine together with the picsearch emotion keywords (Reiner Lenz etc.)
  7. ETU10 Silhouette Dataset – The dataset consists of 720 silhouettes of 10
    objects, with 72 views per object.(M. Akimaliev and M.F. Demirci)
  8. Fish Shape Database – It’s a Fish Shape Database with 100, 2D point set shapes. (Adrian M. Peter)
  9. Flickr 30K – images, actions and captions (Peter Young et al)
  10. Flickr15k – Sketch based Image Retrieval (SBIR) Benchmark – Dataset of 330 sketches and 15,024 photos comprising 33 object categories,benchmark dataset commonly used to evaluate Sketch based Image Retrieval (SBIR) algorithms.(Hu and Collomosse, CVIU 2013)
  11. IAPR TC-12 Image Benchmark (Michael Grubinger)
  12. IAPR-TC12 Segmented and annotated image benchmark (SAIAPR TC-12): (Hugo Jair Escalante)
  13. ImageCLEF 2010 Concept Detection and Annotation Task (Stefanie Nowak)
  14. ImageCLEF 2011 Concept Detection and Annotation Task – multi-label classification challenge in Flickr photos
  15. METU Trademark datasetThe METU Dataset is composed of more than 900K real logos belonging to companies worldwide. (Usta Bilgi Sistemleri A.S. and Grup Ofis Marka Patent A.S)
  16. McGill 3D Shape Benchmark (Siddiqi, Zhang, Macrini, Shokoufandeh, Bouix, Dickinson)
  17. Multiview Stereo Evaluation – Each dataset is registered with a “ground-truth” 3D model acquired via a laser scanning process(Steve Seitz et al)
  18. SHREC 2016 – 3D Sketch-Based 3D Shape Retrieval – data to evaluate the performance of different 3D sketch-based 3D model retrieval algorithms using a hand-drawn 3D sketch query dataset on a generic 3D model dataset (Bo Li)
  19. NIST SHREC – 2014 NIST retrieval contest databases and links (USA National Institute of Standards and Technology)
  20. NIST SHREC – 2013 NIST retrieval contest databases and links (USA National Institute of Standards and Technology)
  21. NIST SHREC 2010 – Shape Retrieval Contest of Non-rigid 3D Models (USA National Institute of Standards and Technology)
  22. NIST TREC Video Retrieval Evaluation Database (USA National Institute of Standards and Technology)
  23. NUS-WIDE – 269K Flickr images annotated with 81 concept tags, enclded as a 500D BoVW descriptor (Chau et al)
  24. Princeton Shape Benchmark (Princeton Shape Retrieval and Analysis Group)
  25. Queensland cross media dataset – millions of images and text documents for “cross-media” retrieval (Yi Yang)
  26. SHREC Watertight Models Track (of SHREC 2007) 400 watertight 3D models (Daniela Giorgi)
  27. SHREC Partial Models Track (of SHREC 2007) 400 watertight 3D DB models and 30 reduced watertight query models (Daniela Giorgi)
  28. SBU Captions Dataset – image captions collected for 1 million images from Flickr (Ordonez, Kulkarni and Berg)
  29. Sketch me That Shoe – Sketch-based object retrieval in a fine-grained setting. Match sketches to specific shoes and chairs. (Qian Yu, QMUL, T. Hospedales Edinburgh/QMUL).
  30. TOSCA 3D shape database (Bronstein, Bronstein, Kimmel)
  31. YouTube-8M Dataset -A Large and Diverse Labeled Video Dataset for Video Understanding
    Research(Google Inc.)

Object Databases

  1. 2.5D/3D Datasets of various objects and scenes (Ajmal Mian)
  2. 3D Object Recognition Stereo DatasetThis dataset consists of 9 objects and 80 test images. (Akash Kushal and Jean Ponce)
  3. 3D Photography Dataseta collection of ten multiview data sets captured in our lab(Yasutaka Furukawa and Jean Ponce)
  4. 3D-Printed RGB-D Object Dataset – 5 objects with groundtruth CAD models and camera trajectories, recorded with various quality RGB-D sensors(Siemens & TUM)
  5. 3DNet Dataset – The 3DNet dataset is a free resource for object class recognition and 6DOF pose estimation from point cloud data. (John Folkesson et al.)
  6. Amsterdam Library of Object Images (ALOI): 100K views of 1K objects (University of Amsterdam/Intelligent Sensory Information Systems)
  7. ASU Office-Home Dataset – Object recognition dataset of everyday objects for domain adaptation (Venkateswara, Eusebio, Chakraborty, Panchanathan)
  8. B3DO: Berkeley 3-D Object Dataset – household object detection (Janoch et al)
  9. Bristol Egocentric Object Interactions Dataset – egocentric object interactions with synchronised gaze (Dima Damen)
  10. CORE image dataset – to help learn more detailed models and for exploring cross-category generalization in object recognition. (Ali Farhadi, Ian Endres, Derek Hoiem, and David A. Forsyth)
  11. CTU Color and Depth Image Dataset of Spread Garments – Images of spread garments with annotated corners.(Wagner, L., Krejov D., and Smutn V. (Czech Technical University in Prague))
  12. Caltech 101 (now 256) category object recognition database (Li Fei-Fei, Marco Andreeto, Marc’Aurelio Ranzato)
  13. Catania Fish Species Recognition – 15 fish species, with about 20,000 sample training images and additional test images (Concetto Spampinato)
  14. COCO-Stuff 10K v1.1 – 10K images labeled with ‘things’ and ‘stuff’ (Caesar, Uijlings, Ferrari)
  15. Columbia COIL-100 3D object multiple views (Columbia University)
  16. Densely sampled object views: 2500 views of 2 objects, eg for view-based recognition and modeling (Gabriele Peters, Universiteit Dortmund)
  17. Edinburgh Kitchen Utensil Database – 897 raw and binary images of 20 categories of kitchen utensil, a resource for training future domestic assistance robots (D. Fullerton, A. Goel, R. B. Fisher)
  18. EDUB-Obj – Egocentric dataset for object localization and segmentation.(Marc Bolaños and Petia Radeva.)
  19. Ellipse finding dataset (Dilip K. Prasad et al)
  20. GDXray:X-ray images for X-ray testing and Computer Vision – GDXray includes five groups of images: Castings, Welds*,Baggages, Nature and Settings. (Domingo Mery, Catholic University of Chile)
  21. GMU Kitchens Dataset – instance level annotation of 11 common household products from BigBird dataset across 9 different kitchens (George Mason University)
  22. Grasping In The Wild – Egocentric video dataset of natural everyday life objects. 16 objects in 7 kitchens. (Benois-Pineau, Larrousse, de Rugy)
  23. GRAZ-02 Database (Bikes, cars, people) (A. Pinz)
  24. GTSDB: German Traffic Sign Detection Benchmark (Ruhr-Universitat Bochum)
  25. ICubWorld – iCubWorld datasets are collections of images acquired by recording from the cameras of the iCub humanoid robot while it observes daily objects.(Giulia Pasquale, Carlo Ciliberto, Giorgio Metta, Lorenzo Natale, Francesca Odone and Lorenzo Rosasco.)
  26. Industrial 3D Object Detection Dataset (MVTec ITODD) – depth and gray value data of 28 objects in 3500 labeled scenes for 3D object detection and pose estimation with a strong focus on industrial settings and applications (MVTec Software GmbH, Munich)
  27. Instagram Food Dataset – A database of 800,000 food images and associated metadata posted to Instagram over 6 week period. Supports food type recognition and social network analysis. (T. Hospedales. Edinburgh/QMUL)
  28. Keypoint-5 dataset – a dataset of five kinds of furniture with their 2D keypoint labels (Jiajun Wu, Tianfan Xue, Joseph Lim, Yuandong Tian, Josh Tenenbaum, Antonio Torralba, Bill Freeman)
  29. KTH-3D-TOTAL – RGB-D Data with objects on desktops annotated. 20 Desks, 3 times per day, over 19 days. (John Folkesson et al.)
  30. LISA Traffic Light Dataset – 6 light classes in various lighting conditions (Jensen, Philipsen, Mogelmose, Moeslund, and Trivedi)
  31. LISA Traffic Sign Dataset – video of 47 US sign types with 7855 annotations on 6610 frames (Mogelmose, Trivedi, and Moeslund)
  32. Linkoping 3D Object Pose Estimation Database (Fredrik Viksten and Per-Erik Forssen)
  33. Linkoping Traffic Signs Dataset – 3488 traffic signs in 20K images (Larsson and Felsberg)
  34. Longterm Labeled – This dataset contains a subset of the observations from the longterm dataset (longterm dataset above). (John Folkesson et al.)
  35. MCIndoor20000 – 20,000 digital images from three different indoor object categories: doors, stairs, and hospital signs. (Bashiri, LaRose, Peissig, and Tafti)
  36. Mexculture142 – Mexican Cultural heritage objects and eye-tracker gaze fixations (Montoya Obeso, Benois-Pineau, Garcia-Vazquez, Ramirez Acosta)
  37. MIT CBCL Car Data (Center for Biological and Computational Learning)
  38. MIT CBCL StreetScenes Challenge Framework: (Stan Bileschi)
  39. Microsoft COCO – Common Objects in Context (Tsung-Yi Lin et al)
  40. Microsoft Object Class Recognition image databases (Antonio Criminisi, Pushmeet Kohli, Tom Minka, Carsten Rother, Toby Sharp, Jamie Shotton, John Winn)
  41. Microsoft salient object databases (labeled by bounding boxes) (Liu, Sun Zheng, Tang, Shum)
  42. Princeton ModelNet – 127,915 CAD Models, 662 Object Categories, 10 Categories with Annotated Orientation (Wu, Song, Khosla, Yu, Zhang, Tang, Xiao)
  43. Moving Labled – This dataset extends the longterm datatset with more locations within the same office environment at KTH. (John Folkesson et al.)
  44. NABirds Dataset – 70,000 annotated photographs of the 400 species of birds commonly observed in North America (Grant Van Horn)
  45. NEC Toy animal object recognition or categorization database (Hossein Mobahi)
  46. NORB 50 toy image database (NYU)
  47. NTU-VOI: NTU Video Object Instance Dataset – video clips with frame-level bounding box annotations of object instances for evaluating object instance search and localization in large scale videos.(Jingjing Meng, et. al.)
  48. Object Pose Estimation Database – This database contains 16 objects, each sampled at 5 degrees angle increments along two rotational axes (F. Viksten etc.)
  49. Object Recognition DatabaseThis database features modeling shots of eight objects and 51 cluttered test shots containing multiple objects.(Fred Rothganger, Svetlana Lazebnik, Cordelia Schmid, and Jean Ponce. )
  50. Osnabrück Synthetic Scalable Cube Dataset – 830000 different cubes captured from 12 different viewpoints for ANN training (Schöning, Behrens, Faion, Kheiri, Heidemann & Krumnack)
  51. PacMan datasets – RGB and 3D synthetic and real data for graspable cookware and crockery (Jeremy Wyatt)
  52. PACS (Photo Art Cartoon Sketch) – An object category recognition dataset dataset for testing domain generalisation: How well can a classifier trained on object images in one domain recognise objects in another domain? (Da Li QMUL, T. Hospedales. Edinburgh/QMUL)
  53. PASCAL 2007 Challange Image Database (motorbikes, cars, cows) (PASCAL Consortium)
  54. PASCAL 2008 Challange Image Database (PASCAL Consortium)
  55. PASCAL 2009 Challange Image Database (PASCAL Consortium)
  56. PASCAL 2010 Challange Image Database (PASCAL Consortium)
  57. PASCAL 2011 Challange Image Database (PASCAL Consortium)
  58. PASCAL 2012 Challange Image Database Category classification, detection, and segmentation, and still-image action
    classification (PASCAL Consortium)
  59. PASCAL Image Database (motorbikes, cars, cows) (PASCAL Consortium)
  60. PASCAL Parts dataset – PASCAL VOC with segmentation annotation for semantic parts of objects (Alan Yuille)
  61. PASCAL-Context dataset – annotations for 400+ additional categories (Alan Yuille)
  62. PASCAL 3D/Beyond PASCAL: A Benchmark for 3D Object Detection in the Wild – 12 class, 3000+ images each with 3D annotations (Yu Xiang, Roozbeh Mottaghi, Silvio Savarese)
  63. Physics 101 dataset – a video dataset of 101 objects in five different scenarios (Jiajun Wu, Joseph Lim, Hongyi Zhang, Josh Tenenbaum, Bill Freeman)
  64. Raindrop Detection – Improved Raindrop Detection using Combined Shape and Saliency Descriptors with Scene Context Isolation – Evaluation Dataset (Breckon, Toby P., Webster, Dereck D.)
  65. ReferIt Dataset (IAPRTC-12 and MS-COCO) – referring expressions for objects in images from the IAPRTC-12 and MS-COCO datasets (Kazemzadeh, Matten, Ordonez, and Berg)
  66. ShapeNet – 3D models of 55 common object categories with about 51K unique 3D models. Also 12K models over 270 categories. (Princeton, Stanford and TTIC)
  67. Swedish Leaf Dataset – These images contains leaves from 15 treeclasses (Oskar J. O. S?derkvist)
  68. T-LESS – An RGB-D dataset for 6D pose estimation of texture-less objects (Tomas Hodan, Pavel Haluza, Stepan Obdrzalek, Jiri Matas, Manolis Lourakis, Xenophon Zabulis)
  69. ToolArtec point clouds – 50 kitchen tool 3D scans (ply) from an Artec EVA scanner. See also ToolKinect – 13 scans using a Kinect 2 and ToolWeb – 116 point clouds of synthetic household tools with mass and affordance groundtruth for 5 tasks. (Paulo Abelha)
  70. TUW Object Instance Recognition Dataset – Annotations of object instances and their 6DoF pose for cluttered indoor scenes observed from various viewpoints and represented as Kinect RGB-D point clouds (Thomas, A. Aldoma, M. Zillich, M. Vincze)
  71. TUW dat sets – Several RGB-D Ground truth and annotated data sets from TUW. (John Folkesson et al.)
  72. UAH Traffic Signs Dataset (Arroyo etc.)
  73. UIUC Car Image Database (UIUC)
  74. UIUC Dataset of 3D object categories (S. Savarese and L. Fei-Fei)
  75. Venezia 3D object-in-clutter recognition and segmentation (Emanuele Rodola)
  76. Visual Attributes Dataset visual attribute annotations for over 500 object classes (animate and inanimate) which are all represented in ImageNet. Each object class is annotated with visual attributes based on a taxonomy of 636 attributes (e.g., has fur, made of metal, is round).
  77. Visual Hull Data Setsa collection of visual hull datasets (Svetlana Lazebnik, Yasutaka Furukawa, and Jean Ponce)
  78. YouTube-BoundingBoxes – 5.6 million accurate human-annotated BB from 23 object classes tracked across frames, from 240,000 YouTube videos, with a strong focus on the person class (1.3 million boxes) (Real, Shlens, Pan, Mazzocchi, Vanhoucke, Khan, Kakarla et al)

People (static and dynamic), human body pose

  1. BUFF dataset – About 10K scans of people in clothing and the estimated body shape of people underneath. Scans contain texture so synthetic videos/images are easy to generate.(Zhang, Pujades, Black and Pons-Moll)
  2. Dynamic Dyna – More than 40K 4D 60fps high resolution scans and models of people very accurately registered. Scans contain texture so synthetic videos/images are easy to generate. (Pons-Moll, Romero, Mahmood and Black)
  3. Dynamic Faust – More than 40K 4D 60fps high resolution scans of people very accurately registered. Scans contain texture so synthetic videos/images are easy to generate. (Bogo, Romero, Pons-Moll and Black)
  4. Extended Chictopia dataset – 14K image Chictopia dataset with additional processed annotations (face) and SMPL body model fits to the images. (Lassner, Pons-Moll and Gehler)
  5. Frames Labeled In Cinema (FLIC) – 20928 frames labeled with human pose (Sapp, Taskar)
  6. Leeds Sports Pose Dataset – 2000 pose annotated images of mostly sports people (Johnson, Everingham)
  7. Look into Person Dataset – 50,000 images with elaborated pixel-wise annotations with 19 semantic human part labels and 2D hposes with 16 key points (Gong, Liang, Zhang, Shen, Lin)
  8. MPII Human Pose Dataset – 25K images containing over 40K people with annotated body joints, 410 human activities {Andriluka, Pishchulin, Gehler, Schiele)
  9. MPII Human Pose Dataset – MPII Human Pose dataset is a de-facto standard benchmark for evaluation of articulated human pose estimation. (Mykhaylo Andriluka, Leonid Pishchulin, Peter Gehler, Bernt Schiele)
  10. People In Photo Albums – Social media photo dataset with images from Flickr, and manual annotations on person heads and their identities.(Ning Zhang and Manohar Paluri and Yaniv Taigman and Rob Fergus and Lubomir Bourdev)
  11. Person Recognition in Personal Photo Collections – we introduced three harder splits for evaluation and long-term attribute annotations and per-photo timestamp metadata.(Oh, Seong Joon and Benenson, Rodrigo and Fritz, Mario and Schiele, Bernt)
  12. Pointing’04 ICPR Workshop Head Pose Image Database
  13. TNT 15 dataset – Several sequences of video synchronised by 10 Inertial Sensors (IMU) worn at the extremities. (von Marcard, Pons-Moll and Rosenhahn)
  14. UC-3D Motion Database – Available data types encompass high resolution Motion Capture, acquired with MVN Suit from Xsens and Microsoft Kinect RGB and depth images.(Institute of Systems and Robotics, Coimbra, Portugal)
  15. United People (UP) Dataset – ˜8,000 images with keypoint and foreground segmentation annotations as well as 3D body model fits. (Lassner, Romero, Kiefel, Bogo, Black, Gehler)
  16. VGG Human Pose Estimation datasets including the BBC Pose (20 videos with an overlaid sign language interpreter), Extended BBC Pose (72 additional training videos), Short BBC Pose (5 one hour videos with sign language signers), and ChaLearn Pose (23 hours of Kinect data of 27 persons performing 20 Italian gestures). (Charles, Everingham, Pfister, Magee, Hogg, Simonyan, Zisserman)
  17. VRLF: Visual Lip Reading Feasibility – audio-visual corpus of 24 speakers recorded in Spanish (Fernandez-Lopez, Martinez and Sukno)

People Detection and Tracking Databases

  1. 3D KINECT Gender Walking data base (L. Igual, A. Lapedriza, R. Borràs from UB, CVC and UOC, Spain)
  2. AGORASET: a dataset for crowd video analysis (Nicolas Courty et al)
  3. CASIA gait database (Chinese Academy of Sciences)
  4. CAVIAR project video sequences with tracking and behavior ground truth (CAVIAR team/Edinburgh University – EC project IST-2001-37540)
  5. CMU Panoptic Studio Dataset – Multiple people social interaction dataset captured by 500+ synchronized video cameras, with 3D full body skeletons and calibration data. (H. Joo, T. Simon, Y. Sheikh)
  6. CUHK Crowd Dataset – 474 video clips from 215 crowded scenes (Shao, Loy, and Wang)
  7. CUHK01 Dataset : Person re-id dataset with 3, 884 images of 972 pedestrians (Rui Zhao et al)
  8. CUHK02 Dataset : Person re-id dataset with five camera view settings. (Rui Zhao et al)
  9. CUHK03 Dataset : Person re-id dataset with 13,164 images of 1,360 pedestrians (Rui Zhao et al)
  10. Caltech Pedestrian Dataset (P. Dollar, C. Wojek, B. Schiele and P. Perona)
  11. Daimler Pedestrian Detection Benchmark 21790 images with 56492 pedestrians plus empty scenes (D. M. Gavrila et al)
  12. Driver Monitoring Video Dataset (RobeSafe + Jesus Nuevo-Chiquero)
  13. DukeMTMC: Duke Multi-Target Multi-Camera tracking dataset – 8 cameras, 85 min, 2m frames, 2000 people of video (Ergys Ristani, Francesco Solera, Roger S. Zou, Rita Cucchiara, Carlo Tomasi)
  14. Edinburgh overhead camera person tracking dataset (Bob Fisher, Bashia Majecka, Gurkirt Singh, Rowland Sillito)
  15. GVVPerfcapEva – repository of human shape and performance capture data, including full body skeletal, hand tracking, body shape, face performance, interactions (Christian Theobalt)
  16. HAT database of 27 human attributes (Gaurav Sharma, Frederic Jurie)
  17. INRIA Person Dataset (Navneet Dalal)
  18. Immediacy Dataset – This dataset is designed for estimation personal relationships. (Xiao Chu et al.)
  19. Inria Dressed human bodies in motion benchmark – Benchmark containing 3D motion sequences of different subjects, motions, and clothing styles that allows to quantitatively measure the accuracy of body shape estimates.(Jinlong Yang, Jean-Sbastien Franco, Franck H=E9troy-Wheeler, and Stefanie Wuhrer)
  20. Izmir – omnidirectional and panoramic image dataset (with annotations) to be used for human and car detection (Yalin Bastanlar)
  21. Joint Attention in Autonomous Driving (JAAD) – The dataset includes instances of pedestrians and cars intended primarily for the purpose of behavioural studies and detection in the context of autonomous driving.(Iuliia Kotseruba, Amir Rasouli and John K. Tsotsos)
  22. JTL Stereo Tacking Dataset for Person Following Robots – 11 different indoor and outdoor places for the task of robots following people under challenging situations (Chen, Sahdev, Tsotsos)
  23. MAHNOB: MHI-Mimicry database – A 2 person, multiple camera and microphone database for studying mimicry in human-human interaction scenarios. (Sun, Lichtenauer, Valstar, Nijholt, and Pantic)
  24. MIT CBCL Pedestrian Data (Center for Biological and Computational Learning)
  25. MPI DYNA – A Model of Dynamic Human Shape in Motion (Max Planck Tubingen)
  26. MPI FAUST Dataset A data set containing 300 real, high-resolution human scans, with automatically computed ground-truth correspondences (Max Planck Tubingen)
  27. MPI JHMDB dataset – Joint-annotated Human Motion Data Base – 21 actions, 928 clips, 33183 frames (Jhuang, Gall, Zuffi, Schmid and Black)
  28. MPI MOSH Motion and Shape Capture from Markers. MOCAP data, 3D shape meshes, 3D high resolution scans. (Max Planck Tubingen)
  29. MVHAUS-PI – a multi-view human interaction recognition dataset (Saeid et al.)
  30. Market-1501 Dataset – 32,668 annotated bounding boxes of 1,501 identities from up to 6 cameras (Liang Zheng et al)
  31. Modena and Reggio Emilia first person head motion videos (Univ of Modena and Reggio Emilia)
  32. Multimodal Activities of Daily Living – including video, audio, physiological, sleep, motion and plug sensors. (Alexia Briasouli)
  33. Multiple Object Tracking Benchmark – A collection of datasets with ground truth, plus a performance league table (ETHZ, U. Adelaide, TU Darmstadt)
  34. Multispectral visible-NIR video sequences – Annotated multispectral video, visible + NIR (LE2I, Universit de Bourgogne)
  35. NYU Multiple Object Tracking Benchmark (Konrad Schindler et al)
  36. Occluded Articulated Human Body Dataset – Body pose extraction and tracking under occlusions, 6 RGB-D sequences in total (3500 frames) with one, two and three users, marker-based ground truth data(Markos Sigalas, Maria Pateraki, Panos Trahanias)
  37. OU-ISIR Gait Database – six video-based gait data sets, two inertial sensor-based gait datasets, and a gait-relevant biometric score data set (Yasushi Makihara)
  38. PARSE Dataset Additional Data – facial expression, gaze direction, and gender (Antol, Zitnick, Parikh)
  39. PARSE Dataset of Articulated Bodies – 300 images of humans and horses (Ramanan)
  40. PDbm: People Detection benchmark repository – realistic sequences, manually annotated people detection ground truth and a complete evaluation framework (García-Martín, Martínez, Bescós)
  41. PDds: A Person Detection dataset – several annotated surveillance sequences of different levels of complexity (García-Martín, Martínez, Bescós)
  42. PETS 2009 Crowd Challange dataset (Reading University & James Ferryman)
  43. PETS Winter 2009 workshop data (Reading University & James Ferryman)
  44. PETS: 2015 Performance Evaluation of Tracking and Surveillance (Reading University & James Ferryman)
  45. PETS: 2015 Performance Evaluation of Tracking and Surveillance (Reading University & Luis Patino)
  46. PETS 2016 datasets – multi-camera (including thermal cameras) video recordings of human behavior around a stationary vehicle and around a boat (Thomas Cane)
  47. PIROPO – People in Indoor ROoms with Perspective and Omnidirectional cameras, with more than 100,000 annotated frames (GTI-UPM, Spain)
  48. People-Art – a databased containing people labelled in photos and artwork (Qi Wu and Hongping Cai)
  49. Photo-Art-50 – a databased containing 50 object classes annoted in photos and artwork (Qi Wu and Hongping Cai)
  50. Pixel-based change detection benchmark dataset (Goyette et al)
  51. Precarious Dataset – unusual people detection dataset (Huang)
  52. RAiD – Re-Identification Across Indoor-Outdoor Dataset: 43 people, 4 cameras, 6920 images (Abir Das et al)
  53. SYNTHIA – Large set (~half million) of virtual-world images for training autonomous cars to see. (ADAS Group at Computer Vision Center)
  54. Shinpuhkan 2014 – A Person Re-identification dataset containing 22,000 images of 24 people captured by 16 cameras. (Yasutomo Kawanishi et al.)
  55. Stanford Structured Group Discovery dataset – Discovering Groups of People in Images (W. Choi et al)
  56. Temple Color 128 – Color Tracking Benchmark – Encoding Color Information for Visual Tracking (P. Liang, E. Blasch, H. Ling)
  57. Transient Biometrics Nails Dataset V01 (Igor Barros Barbosa)
  58. TUM Gait from Audio, Image and Depth (GAID) database – containing tracked RGB video, tracked depth video, and audio for 305 subjects (Babaee, Hofmann, Geiger, Bachmann, Schuller, Rigoll)
  59. TVPR (Top View Person Re-identification) dataset – person re-identification using an RGB-D camera in a Top-View configuration: indoor 23 sessions, 100 people, 8 days (Liciotti, Paolanti, Frontoni, Mancini and Zingaretti)
  60. UCLA Aerial Event Dataset – Human activities in aerial videos with annotations of people, objects, social groups, activities and roles (Shu, Xie, Rothrock, Todorovic, and Zhu)
  61. Univ of Central Florida – Crowd Dataset (Saad Ali)
  62. Univ of Central Florida – Crowd Flow Segmentation datasets (Saad Ali)
  63. VIPeR: Viewpoint Invariant Pedestrian Recognition – 632 pedestrian image pairs taken from arbitrary viewpoints under varying illumination conditions. (Gray, Brennan, and Tao)
  64. Visual object tracking challenge datasets – The VOT datasets is a collection of fully annotated visual object tracking datasets used in the single-target short-term visual object tracking challenges.(The VOT committee)
  65. WIDER Attribute Dataset – WIDER Attribute is a large-scale human attribute dataset, with 13789 images belonging to 30 scene categories, and 57524 human bounding boxes each annotated with 14 binary attributes.(Li, Yining and Huang, Chen and Loy, Chen Change and Tang, Xiaoou)
  66. WUds: Wheelchair Users Dataset – wheelchair users detection data, to extend people detection, providing a more general solution to detect people in environments such as independent and assisted living, hospitals, healthcare centers and senior residences (Martín-Nieto, García-Martín, Martínez)
  67. YouTube-BoundingBoxes – 5.6 million accurate human-annotated BB from 23 object classes tracked across frames, from 240,000 YouTube videos, with a strong focus on the person class (1.3 million boxes) (Real, Shlens, Pan, Mazzocchi, Vanhoucke, Khan, Kakarla et al)

Remote Sensing

  1. Brazilian Cerrado-Savanna Scenes Dataset – Composition of IR-R-G scenes taken by RapidEye sensor for vegetation classification in Brazilian Cerrado-Savanna. (K. Nogueira, J. A. dos Santos, T. Fornazari, T. S. Freire, L. P. Morellato, R. da S. Torres)
  2. Brazilian Coffee Scenes Dataset – Composition of IR-R-G scenes taken by SPOT sensor for identification of coffee crops in Brazilian mountains.(O. A. B. Penatti, K. Nogueira, J. A. dos Santos.)
  3. Building Detection Benchmark -14 images acquired from IKONOS (1 m) and QuickBird (60 cm)(Ali Ozgun Ok and Caglar Senaras)
  4. CBERS-2B, Landsat 5 TM, Geoeye, Ikonos-2 MS and ALOS-PALSAR – land-cover classification using optical images(D. Osaku et al. )
  5. FORTH Multispectral Imaging (MSI) datasets – 5 datasets for Multispectral Imaging (MSI), annotated with ground truth data (Polykarpos Karamaoynas)
  6. Furnas and Tiete – sediment yield classification( Pisani et al.)
  7. ISPRS 2D semantic labeling – Height models and true ortho-images with a ground sampling distance of 5cm have been prepared over the city of Potsdam/Germany (Franz Rottensteiner, Gunho Sohn, Markus Gerke, Jan D. Wegner)
  8. ISPRS 3D semantic labeling – nine class airborne laser scanning data (Franz Rottensteiner, Gunho Sohn, Markus Gerke, Jan D. Wegner)
  9. Inria Aerial Image Labeling Dataset – 9000 square kilometeres of color aerial imagery over U.S. and Austrian cities. (Emmanuel Maggiori, Yuliya Tarabalka, Guillaume Charpiat, Pierre Alliez.)
  10. Lampert’s Spectrogram Analysis – Passive sonar spectrogram images derived from time-series data, these spectrograms are generated from recordings of acoustic energy radiated from propeller and engine machinery in underwater sea recordings. (Thomas Lampert)
  11. Linkoping Thermal InfraRed dataset – The LTIR dataset is a thermal infrared dataset for evaluation of Short-Term Single-Object (STSO) tracking (Linkoping University)
  12. MUUFL Gulfport Hyperspectral and LiDAR data set – Co-registered aerial hyperspectral and lidar data over the University of Southern Mississippi Gulfpark campus containing several sub-pixel targets. (Gader, Zare, Close, Aitken, Tuell)
  13. UC Merced Land Use Dataset 21 class land use image dataset with 100 images per class, largely urban, 256×256 resolution, 1 foot pixels (Yang and Newsam)
  14. Zurich Summer dataset – t is intended for semantic segmentation of very high resolution satellite images of urban scenes, with incomplete ground truth (Michele Volpi and Vitto Ferrari.)
  15. Zurich Urban Micro Aerial Vehicle Dataset – time synchronized aerial high-resolution images of 2 km of Zurich, with associated other data (Majdik, Till, Scaramuzza)

Scenes or Places, Scene Segmentation or Classification

  1. Barcelona – 15,150 images, urban views of Barcelona (Tighe and Lazebnik)
  2. CMU Visual Localization Data Set – Dataset collected over the period of a year using the
    Navlab 11 equipped with IMU, GPS, INS, Lidars and cameras.(Hernan Badino, Daniel Huber and Takeo Kanade)
  3. COLD (COsy Localization Database) – place localization (Ullah, Pronobis, Caputo, Luo, and Jensfelt)
  4. DAVIS: Video Object Segmentation dataset 2016 – A Benchmark Dataset and Evaluation Methodology for Video Object Segmentation (F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross, and A. Sorkine-Hornung)
  5. DAVIS: Video Object Segmentation dataset 2017 – The 2017 DAVIS Challenge on Video Object Segmentation (J. Pont-Tuset, F. Perazzi, S. Caelles, P. Arbelaez, A. Sorkine-Hornung, and L. Van Gool)
  6. EDUB-Seg– Egocentric dataset for event segmentation.(Mariella Dimiccoli, Marc Bolaños, Estefania Talavera, Maedeh Aghaei, Stavri G. Nikolov, and Petia Radeva.)
  7. Fifteen Scene Categoriesa dataset of fifteen natural scene categories ( Fei-Fei Li and Aude Oliva)
  8. Geometric Context – scene interpretation images (Derek Hoiem)
  9. HyKo: A Spectral Dataset for Scene Understanding – The HyKo dataset was captured with compact, low-cost, snapshot mosaic (SSM) imaging cameras, which are able to capture a whole spectral cube in one shot recorded from a moving vehicle enabling hyperspectral scene analysis for road scene understanding. (Active Vision Group, University of Koblenz-Landau)
  10. Indoor Place Recognition Dataset for localization of Mobile Robots – The dataset contains 17 different places built from 2 different robots (virtualMe and pioneer) (Raghavender Sahdev, John K. Tsotsos.)
  11. Indoor Scene Recognition – 67 Indoor categories, 15620 images (Quattoni and Torralba)
  12. Intrinsic Images in the Wild (IIW) – Intrinsic Images in the Wild, is a large-scale, public dataset for evaluating intrinsic image decompositions of indoor scenes (Sean Bell, Kavita Bala, Noah Snavely)
  13. LM+SUN – 45,676 images, mainly urban or human related scenes (Tighe and Lazebnik)
  14. Maritime Imagery in the Visible and Infrared Spectrums – VAIS contains simultaneously acquired unregistered thermal and visible images of ships acquired from piers (Zhang, Choi, Daniilidis, Wolf, & Kanan)
  15. Materials in Context (MINC) – The Materials in Context Database (MINC) builds on OpenSurfaces, but includes millions of point annotations of material labels. (Sean Bell, Paul Upchurch, Noah Snavely, Kavita Bala)
  16. MIT Intrinsic Images – 20 objects
    {Roger Grosse, Micah K. Johnson, Edward H. Adelson, and William T. Freeman)
  17. NYU V2 Mixture of Manhattan Frames Dataset – We provide the Mixture of Manhattan Frames (MMF) segmentation and MF rotations on the full NYU depth dataset V2 by Silberman et al. (Straub, Julian and Rosman, Guy and Freifeld, Oren and Leonard, John J. and Fisher III, John W.)
  18. OpenSurfaces – OpenSurfaces consists of tens of thousands of examples of surfaces segmented from consumer photographs of interiors, and annotated with material parameters, texture information, and contextual information . (Kavita Bala et al.)
  19. Oxford Audiovisual Segmentation Dataset – Oxford Audiovisual Segmentation Dataset with Oxford Audiovisual Segmentation Dataset including audio recordings of objects being struck (Arnab, Sapienza, Golodetz, Miksik and Torr)
  20. Places 2 Scene Recognition database -365 scene categories and 8 millions of images (Zhou, Khosla, Lapedriza, Torralba and Oliva)
  21. Places Scene Recognition database – 205 scene categories and 2.5 millions of images (Zhou, Lapedriza, Xiao, Torralba, and Oliva)
  22. RGB-NIR Scene Dataset – 477 images in 9 categories captured in RGB and Near-infrared (NIR) (Brown and Susstrunk)
  23. RMS2017 – Reconstruction Meets Semantics outdoor dataset – 500 semantically annotated images with poses and point cloud from a real garden (Tylecek, Sattler)
  24. RMS2018 – Reconstruction Meets Semantics virtual dataset – 30k semantically annotated images with poses and point cloud from 6 virtual gardens (An, Tylecek)
  25. Southampton-York Natural Scenes Dataset 90 scenes, 25 indoor and outdoor scene categories, with spherical LiDAR, HDR intensity, stereo intensity panorama. (Adams, Elder, Graf, Leyland, Lugtigheid, Muryy)
  26. SUN 2012 – 16,873 fully annotated scene images for scene categorization (Xiao et al)
  27. SUN 397 – 397 scene categories for scene classification (Xiao et al)
  28. SUN RGB-D: A RGB-D Scene Understanding Benchmark Suite – 10,000 RGB-D images, 146,617 2D polygons and 58,657 3D bounding boxes (Song, Lichtenberg, and Xiao)
  29. SYNTHIA – Large set (~half million) of virtual-world images for training autonomous cars to see. (ADAS Group at Computer Vision Center)
  30. Sift Flow (also known as LabelMe Outdoor, LMO) – 2688 images, mainly outdoor natural and urban (Tighe and Lazebnik)
  31. Stanford Background Dataset – 715 images of outdoor scenes containing at least one foreground object (Gould et al)
  32. Surface detection – Real-time traversable surface detection by colour space fusion and temporal analysis – Evaluation Dataset (Breckon, Toby P., Katramados, Ioannis)
  33. ViDRILO – ViDRILO is a dataset containing 5 sequences of annotated RGB-D images acquired with a mobile robot in two office buildings under challenging lighting conditions.(Miguel Cazorla, J. Martinez-Gomez, M. Cazorla, I. Garcia-Varea and V. Morell.)

Segmentation (General)

  1. Shadow Detection/Texture Segmentation Computer Vision Dataset – Video based sequences for shadow detection/suppression, with ground truth (Newey, C., Jones, O., & Dee, H. M.)
  2. Aberystwyth Leaf Evaluation Dataset – Timelapse plant images with hand marked up leaf-level segmentations for some time steps, and biological data from plant sacrifice. (Bell, Jonathan; Dee, Hannah M.)
  3. Alpert et al. Segmentation evaluation database (Sharon Alpert, Meirav Galun, Ronen Basri, Achi Brandt)
  4. BMC (Background Model Challenge) – A dataset for comparing background subtraction algorithms, comp=
    osed of real and synthetic videos(Antoine)
  5. Berkeley Segmentation Dataset and Benchmark (David Martin and Charless Fowlkes)
  6. CAD 120 affordance dataset – Pixelwise affordance annotation in human context (Sawatzky, Srikantha, Gall)
  7. CTU Color and Depth Image Dataset of Spread Garments – Images of spread garments with annotated corners.(Wagner, L., Krejov D., and Smutn V. (Czech Technical University in Prague))
  8. CTU Garment Folding Photo Dataset – Color and depth images from various stages of garment folding.(Sushkov R., Melkumov I., Smutn y V. (Czech Technical University in Prague))
  9. DeformIt 2.0 – Image Data Augmentation Tool: Simulate novel images with ground truth segmentations from a single image-segmentation pair (Brian Booth and Ghassan Hamarneh)
  10. GrabCut Image database (C. Rother, V. Kolmogorov, A. Blake, M. Brown)
  11. ICDAR’15 Smartphone document capture and OCR competition – challenge 1 – videos of documents filmed by a user with a smartphone to simulate mobile document capture, and ground truth coordinates of the document corners to detect. (Burie, Chazalon, Coustaty, Eskenazi, Luqman, Mehri, Nayef, Ogier, Prum and Rusinol)
  12. Intrinsic Images in the Wild (IIW) – Intrinsic Images in the Wild, is a large-scale, public dataset for evaluating intrinsic image decompositions of indoor scenes (Sean Bell, Kavita Bala, Noah Snavely)
  13. LabelMe images database and online annotation tool (Bryan Russell, Antonio Torralba, Kevin Murphy, William Freeman)
  14. LITS Liver Tumor Segmentation – 130 3D CT scans with segmentations of the liver and liver tumor. Public benchmark with leaderboard at Codalab.org (Patrick Christ)
  15. Materials in Context (MINC) – The Materials in Context Database (MINC) builds on OpenSurfaces, but includes millions of point annotations of material labels. (Sean Bell, Paul Upchurch, Noah Snavely, Kavita Bala)
  16. OpenSurfaces – OpenSurfaces consists of tens of thousands of examples of surfaces segmented from consumer photographs of interiors, and annotated with material parameters, texture information, and contextual information . (Kavita Bala et al.)
  17. Osnabrück gaze tracking data – 318 video sequences from several different gaze tracking data sets with polygon based object annotation (Schöning, Faion, Heidemann, Krumnack, Gert, Açik, Kietzmann, Heidemann & König)
  18. PetroSurf3D – 26 high resolution (sub-millimeter accuracy) 3D scans of rock art with pixelwise labeling of petroglyphs for segmentation(Poier, Seidl, Zeppelzauer, Reinbacher, Schaich, Bellandi, Marretta, Bischof)
  19. SYNTHIA – Large set (~half million) of virtual-world images for training autonomous cars to see. (ADAS Group at Computer Vision Center)
  20. Stony Brook University Shadow Dataset (SBU-Shadow5k) – Large scale shadow detection dataset from a wide variety of scenes and photo types, with human annotations (Tomas F.Y. Vicente, Le Hou, Chen-Ping Yu, Minh Hoai, Dimitris Samaras)

Simultaneous Localization and Mapping

  1. Event-Camera Data for Pose Estimation, Visual Odometry, and SLAMThe data also include intensity images, inertial measurements, and ground truth from a motion-capture system. (ETH)
  2. RAWSEEDS SLAM benchmark datasets (Rawseeds Project)
  3. TUM RGB-D Benchmark – Dataset and benchmark for the evaluation of RGB-D visual
    odometry and SLAM algorithms (BCrgen Sturm, Nikolas Engelhard, Felix Endres, Wolfram Burgard and Daniel Cremers)
  4. Visual Odometry / SLAM Evaluation – The odometry benchmark consists of 22 stereo sequences (Andreas Geiger and Philip Lenz and Raquel Urtasun)

Surveillance and Tracking

  1. ATOMIC GROUP ACTIONS dataset – (Ricky J. Sethi et al.)
  2. AVSS07: Advanced Video and Signal based Surveillance 2007 datasets (Andrea Cavallaro)
  3. Activity modeling and abnormality detection dataset – The dataset containes a 45 minutes video with annotated anomalies.(Jagan Varadarajan and Jean-Marc Odobez)
  4. Background subtraction – a list of datasets about background subtraction(Thierry BOUWMANS )
  5. CAMO-UOW Dataset – 10 high resolution videos captured in real scenes for camouflaged background subtraction (Shuai Li and Wanqing Li)
  6. CMUSRD: Surveillance Research Dataset – multi-camera video for indoor surveillance scenario (K. Hattori, H. Hattori, et al)
  7. DukeMTMC: Duke Multi-Target Multi-Camera tracking dataset – 8 cameras, 85 min, 2m frames, 2000 people of video (Ergys Ristani, Francesco Solera, Roger S. Zou, Rita Cucchiara, Carlo Tomasi)
  8. DukeMTMC-reID – A subset of the DukeMTMC for image-based person re-identification (8 cameras,16,522 training images of 702 identities, 2,228 query images of the other 702 identities and 17,661 gallery images.) (Zheng, Zheng, and Yang)
  9. ETISEO Video Surveillance Download Datasets (INRIA Orion Team and others)
  10. HDA+ Multi-camera Surveillance Dataset – video from a network of 18 heterogeneous cameras (different resolutions and frame rates) distributed over 3 floors of a research institute with 13 fully labeled sequences, 85 persons, and 64028 bounding boxes of persons. (D. Figueira, M. Taiana, A. Nambiar, J. Nascimento and A. Bernardino)
  11. Human click data – 20K human clicks on a tracking target (including click errors) (Zhu and Porikli)
  12. Immediacy Dataset – This dataset is designed for estimation personal relationships. (Xiao Chu et al.)
  13. MAHNOB Databases -including Laughter Database,HCI-tagging Database,MHI-Mimicry Database( M. Pantic. etc.)
  14. MVHAUS-PI – a multi-view human interaction recognition dataset (Saeid et al.)
  15. Multispectral visible-NIR video sequences – Annotated multispectral video, visible + NIR (LE2I, Universit de Bourgogne)
  16. Openvisor – Video surveillance Online Repository (Univ of Modena and Reggio Emilia)
  17. Parking-Lot dataset – Parking-Lot dataset is a car dataset which focus on moderate and heavily occlusions on cars in the parking lot scenario.(B. Li, T.F. Wu and S.C. Zhu)
  18. Pornography Database – The Pornography database is a pornography detection dataset containing nearly 80 hours of 400 pornographic and 400 non-pornographic videos extracted from pornography websites and Youtube. (Avila, Thome, Cord, Valle, de Araujo)
  19. Princeton Tracking Benchmark – 100 RGBD tracking datasets (Song and Xiao)
  20. QMUL Junction Dataset 1 and 2 – Videos of busy road junctions. Supports anomaly detection tasks. (T. Hospedales Edinburgh/QMUL)
  21. Queen Mary Multi-Camera Distributed Traffic Scenes Dataset (QMDTS) – The QMDTS is collected from urban surveillance environment for the study of surveillance behaviours in distributed scenes.(Dr. Xun Xu. Prof. Shaogang Gong and Dr. Timothy Hospedales)
  22. SALSA: Synergetic sociAL Scene Analysis – A Novel Dataset for Multimodal Group Behavior Analysis(Xavier Alameda-Pineda etc.)
  23. SBMnet (Scene Background Modeling.NET) – A dataset for testing background estimation algorithms(Jodoin, Maddalena, and Petrosino)
  24. SBM-RGBD dataset – 35 Kinect indoor RGBD videos to evaluate and compare scene background modelling methods for moving object detection (Camplani, Maddalena, Moyà Alcover, Petrosino, Salgado)
  25. SCOUTER – video surveillance ground truthing (shifting perspectives, different setups/lighting conditions, large variations of subject). 30 videos and approximately 36,000 manually labeled frames. (Catalin Mitrea)
  26. SJTU-BESTOne surveillance-specified datasets platform with realistic, on-using camera-captured, diverse set of surveillance images and videos (Shanghai Jiao Tong University)
  27. SPEVI: Surveillance Performance EValuation Initiative (Queen Mary University London)
  28. Shinpuhkan 2014 – A Person Re-identification dataset containing 22,000 images of 24 people captured by 16 cameras. (Yasutomo Kawanishi et al.)
  29. Stanford Drone Dataset – 60 images and videos of various types of agents (not just pedestrians, but also bicyclists, skateboarders, cars, buses, and golf carts) that navigate in a real world outdoor environment such as a university campus (Robicquet, Sadeghian, Alahi, Savarese)
  30. Tracking in extremely cluttered scenes – this single object tracking dataset has 28 highly cluttered sequences with per frame annotation(Jingjing Xiao,Linbo Qiao,Rustam Stolkin,Ale Leonardis)
  31. UCLA Aerial Event Dataset – Human activities in aerial videos with annotations of people, objects, social groups, activities and roles (Shu, Xie, Rothrock, Todorovic, and Zhu)
  32. UCSD Anomaly Detection Dataset – a stationary camera mounted at an elevation, overlooking pedestrian walkways, with unusual pedestrian or non-pedestrian motion.
  33. UCSD trajectory clustering and analysis datasets – (Morris and Trivedi)
  34. USC Information Sciences Institute’s ATOMIC PAIR ACTIONS dataset – (Ricky J. Sethi et al.)
  35. Udine Trajectory-based anomalous event detection dataset – synthetic trajectory datasets with outliers (Univ of Udine Artificial Vision and Real Time Systems Laboratory)
  36. Visual Tracker Benchmark – 100 object tracking sequences with ground truth with Visual Tracker Benchmark evaluation, including tracking results from a number of trackers (Wu, Lim, Yang)
  37. WIDER Attribute Dataset – WIDER Attribute is a large-scale human attribute dataset, with 13789 images belonging to 30 scene categories, and 57524 human bounding boxes each annotated with 14 binary attributes.(Li, Yining and Huang, Chen and Loy, Chen Change and Tang, Xiaoou)

Textures

  1. Brodatz Texture, Normalized Brodatz Texture, Colored Brodatz Texture, Multiband Brodatz Texture 154 new images plus 112 original images with various transformations (A. Safia, D. He)
  2. Color texture images by category (textures.forrest.cz)
  3. Columbia-Utrecht Reflectance and Texture Database (Columbia & Utrecht Universities)
  4. DynTex: Dynamic texture database (Renaud Piteri, Mark Huiskes and Sandor Fazekas)
  5. Intrinsic Images in the Wild (IIW) – Intrinsic Images in the Wild, is a large-scale, public dataset for evaluating intrinsic image decompositions of indoor scenes (Sean Bell, Kavita Bala, Noah Snavely)
  6. KTH TIPS & TIPS2 textures – pose/lighting/scale variations (Eric Hayman)
  7. Materials in Context (MINC) – The Materials in Context Database (MINC) builds on OpenSurfaces, but includes millions of point annotations of material labels. (Sean Bell, Paul Upchurch, Noah Snavely, Kavita Bala)
  8. OpenSurfaces – OpenSurfaces consists of tens of thousands of examples of surfaces segmented from consumer photographs of interiors, and annotated with material parameters, texture information, and contextual information . (Kavita Bala et al.)
  9. Oulu Texture Database (Oulu University)
  10. Oxford Describable Textures Dataset – 5640 images in 47 categories (M.Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, A. Vedaldi)
  11. Prague Texture Segmentation Data Generator and Benchmark (Mikes, Haindl)
  12. Salzburg Texture Image Database (STex) – a large collection of 476 color texture image that have been captured around Salzburg, Austria. (Roland Kwitt and Peter Meerwald)
  13. Texture DatabaseThe texture database features 25 texture classes, 40 samples each(Svetlana Lazebnik, Cordelia Schmid, and Jean Ponce)
  14. Uppsala texture dataset of surfaces and materials – fabrics, grains, etc.
  15. Vision Texture (MIT Media Lab)

Urban Datasets

  1. Barcelona – 15,150 images, urban views of Barcelona (Tighe and Lazebnik)
  2. CMP Facade Database – Includes 606 rectified images of facades from various places with 12 architectural classes annotated.(Radim Tylecek)
  3. LM+SUN – 45,676 images, mainly urban or human related scenes (Tighe and Lazebnik)
  4. MIT CBCL StreetScenes Challenge Framework: (Stan Bileschi)
  5. Queen Mary Multi-Camera Distributed Traffic Scenes Dataset (QMDTS) – The QMDTS is collected from urban surveillance environment for the study of surveillance behaviours in distributed scenes.(Dr. Xun Xu. Prof. Shaogang Gong and Dr. Timothy Hospedales)
  6. Robust Global Translations with 1DSfMthe numerical data describing global structure from motion problems for each dataset (Kyle Wilson and Noah Snavely)
  7. Sift Flow (also known as LabelMe Outdoor, LMO) – 2688 images, mainly outdoor natural and urban (Tighe and Lazebnik)
  8. Street-View Change Detection with Deconvolutional Networks – Database with aligned image pairs from street-view imagery with structural,lighting, weather and seasonal changes.(Pablo F. Alcantarilla, Simon Stent, German Ros, Roberto Arroyo and Riccardo Gherardi)
  9. SydneyHouse – Streetview house images with accurate 3D house shape, facade object label, dense point correspondence, and annotation toolbox.(Hang Chu, Shenlong Wang, Raquel Urtasun,Sanja Fidler)
  10. Traffic Signs Dataset – recording sequences from over 350 km of Swedish highways and city roads (Fredrik Larsson)

Vision and Natural Language

  1. Dataset of Structured Queries and Spatial Relations – Dataset of structured queries about images with the emphasise on spatial relations.(Mateusz Malinowski, Mario Fritz)
  2. Hannah and her sisters database – a dense audio-visual person-oriented ground-truth annotation of faces, speech segments, shot boundaries (Patrick Perez, Technicolor)
  3. Large scale Movie Description Challenge (LSMDC) – A large scale dataset and challenge for movie description, including over 128K video-sentence pairs, mainly sourced from Audio Description (also known as DVS). (Rohrbach, Torabi, Rohrbach, Tandon, Pal, Larochelle, Courville and Schiele)
  4. MPI Movie Description dataset – text and video – A dataset of movie clips associated with natural language descriptions sourced from movie scripts and Audio Description. (Rohrbach, Rohrbach, Tandon and Schiele)
  5. TACoS Multi-Level Corpus – Dataset of cooking videos associated with natural language descriptions at three levels of detail (long, short and single sentence). (Rohrbach, Rohrbach, Qiu, Friedrich, Pinkal and Schiele)
  6. TGIF – 100K animated GIFs from Tumblr and 120K natural language descriptions (Li, Song, Cao, Tetreault, Goldberg, Jaimes, Luo)
  7. Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations – Visual Genome is a dataset, a knowledge base, an ongoing effort to connect structured image concepts to language.(Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li Jia-Li, David Ayman Shamma, Michael Bernstrein, Li Fei-Fei)

Other Collections

  1. 4D Light Field Dataset – 24 synthetic scenes with 9x9x512x512x3 input images, depth and disparity ground truth, camera parameters, and evaluation masks. (Katrin Honauer, Ole Johannsen, Daniel Kondermann, Bastian Goldluecke)
  2. AMADI_LontarSet – Balinese Palm Leaf Manuscript Images Dataset for Binarization, Query-by-Example Word Spotting, and Isolated Character Recognition of Balinese Script. (The AMADI Project et al.)
  3. Annotated Web Ears Dataset (AWE Dataset) – All images were acquired by cropping ears from images from the internet of known persons.( Ziga Emersic, Vitomir Struc and Peter Peer)
  4. Biometrics Evaluation and Testing – Evaluation of identification technologies, including Biometrics( European computing e-infrastructure)
  5. CALVIN research group datasets – object detection with eye tracking, imagenet bounding boxes, synchronised activities, stickman and body poses, youtube objects, faces, horses, toys, visual attributes, shape classes (CALVIN ggroup)
  6. CANTATA Video and Image Database Index site (Multitel)
  7. Chinese University of Hong Kong datasets – Face sketch, face alignment, image search, public square observation, occlusion, central station, MIT single and multiple camera trajectories, person re-identification (Multimedia lab)
  8. Computer Vision Homepage list of test image databases (Carnegie Mellon Univ)
  9. Computer Vision Lab OCR DataBase (CVL OCR DB) – CVL OCR DB is a public annotated image dataset of 120 binary annotated images of text in natural scenes. (Andrej Ikica and Peter Peer.)
  10. ETHZ various, including ETH 3D head pose, BIWI audiovisual data, ETHZ shape classes, BIWI walking pedestrians, pedestrians, buildings, 4D MRI, personal events, liver untrasound, Food 101 (ETH Zurich, Computer Vision Lab)
  11. General 100 Dataset – General-100 dataset contains 100 bmp-format images (with no
    compression), which are well-suited for super-resolution training(Dong, Chao and Loy, Chen Change and Tang, Xiaoou)
  12. HKU-IS -4447 images with pixel labeling groundtruth for salient object detection.(Guanbin Li, Yizhou Yu)
  13. High-res 3D-Models – it includes high-res renderings of these data-sets. ( Hubert etc.)
  14. I3 – Yahoo Flickr Creative Commons 100M – This dataset contains a list of photos and videos. (B. Thomee, D.A. Shamma, G. Friedland et al.)
  15. IDIAP dataset collection – 26 different datasets – multimodal, attack, biometric, cursive characters, discourse, eye gaze, posters, maya codex, MOBIO, face spoofing, game playing, finger vein, youtube-personality traits (IDIAP team)
  16. Kinect v2 dataset – Dataset for evaluating unwrapping in kinect2 depth decoding (Felix etc.)
  17. Laval HDR Sky Database – The database contains 800 hemispherical, full HDR photos of the sky that can be used for outdoor lighting analysis. (Jean-Francois Lalonde et al.)
  18. Leibe’s Collection of people/vehicle/object databases (Bastian Leibe)
  19. Lotus Hill Image Database Collection with Ground Truth (Sealeen Ren, Benjamin Yao, Michael Yang)
  20. MIT Saliency Benchmark dataset – collection (pointers to 23 datasets) (Bylinskii, Judd, Borji, Itti, Durand, Oliva, Torralba}
  21. Michael Firman’s List of RGBD datasets
  22. Msspoof:2D multi-spectral face spoofing – Presentation attack (spoofing) dataset with samples from both real data subjects and spoofed data subjects performed with paper to a NIR and VIS camera(Idiap research institute)
  23. Multiview Stereo Evaluation – Each dataset is registered with a “ground-truth” 3D model acquired via a laser scanning process(Steve Seitz et al)
  24. Oxford Misc, including Buffy, Flowers, TV characters, Buildings, etc (Oxford Visual geometry Group)
  25. PEIPA Image Database Summary (Pilot European Image Processing Archive)
  26. PalmVein spoofing – Presentation attack (spoofing) dataset with samples from spoofed data subjects (corresponding to VERA Palmvein) performed with paper(Idiap research institute)
  27. RSBA dataset – Sequences for evaluating rolling shutter bundle adjustment (Per-Erik etc.)
  28. Replay Attack:2D face spoofing – Presentation attack (spoofing) dataset with samples from both real data subjects and spoofed data subjects performed with paper, photos and videos from a mobile device to a laptop.(Idiap research institute)
  29. Replay Mobile:2D face spoofing – Presentation attack (spoofing) dataset with samples from both real data subjects and spoofed data subjects performed with paper, photos and videos to/from a mobile device.(Idiap research institute)
  30. Synthetic Sequence Generator – Synthetic Sequence Generator (G. Hamarneh)
  31. The Event-Camera Dataset – This presents the world’s first collection of datasets with an event-based camera for high-speed robotics (E. Mueggler, H. Rebecq, G. Gallego, T. Delbruck, D. Scaramuzza)
  32. The world from a cat perspective – videos recorded from the head of a freely behaving cat (Belinda Y. Betsch, Wolfgang Einh?user)
  33. USC Annotated Computer Vision Bibliography database publication summary (Keith Price)
  34. USC-SIPI image databases: texture, aerial, favorites (eg. Lena) (USC Signal and Image Processing Institute)
  35. Univ of Bern databases on handwriting, online documents, string edit and graph matching (Univ of Bern, Computer Vision and Artificial Intelligence)
  36. VERA Fingervein spoofing – Presentation attack (spoofing) dataset with samples from spoofed data subjects (corresponding to VERA Fingervein) performed with paper(Idiap research institute)
  37. VERA Fingervein – Fingervein dataset with data subjects recorded with a open fingervein sensor(Idiap research institute)
  38. VERA PalmVein:PalmVein – Palmvein dataset with data subjects recorded with a open palmvein sensor(Idiap research institute)
  39. Vehicle Detection in Aerial Imagery – VEDAI is a dataset for Vehicle Detection in Aerial Imagery, provided as a tool to benchmark automatic target recognition algorithms in unconstrained environments. (Sebastien Razakarivony and Frederic Jurie)
  40. Video Stacking Dataset – Dataset for evaulating video stacking on cell-phones (Erik Ringaby etc.)
  41. Wrist-mounted camera video dataset – Activities of Daily Living videos captured from a wrist- mounted camera and a head-mounted camera(Katsunori Ohnishi, Atsushi Kanehira,Asako Kanezaki, Tatsuya Harada)
  42. Yummly-10k dataset – The goal was to understand human perception, in this case of food taste similarity.(SE(3) Computer Vision Group at Cornell Tech)

Miscellaneous

  1. ICDAR’15 Smartphone document capture and OCR competition – challenge 2 – pictures of documents captured with smartphones under various conditions of perspective, lighting, etc. The ground truth is the textual content which should be extracted. (Burie, Chazalon, Coustaty, Eskenazi, Luqman, Mehri, Nayef, Ogier, Prum and Rusinol)
  2. Intrinsic Images in the Wild (IIW) – Intrinsic Images in the Wild, is a large-scale, public dataset for evaluating intrinsic image decompositions of indoor scenes (Sean Bell, Kavita Bala, Noah Snavely)
  3. LabelMe images database and online annotation tool (Bryan Russell, Antonio Torralba, Kevin Murphy, William Freeman)
  4. Materials in Context (MINC) – The Materials in Context Database (MINC) builds on OpenSurfaces, but includes millions of point annotations of material labels. (Sean Bell, Paul Upchurch, Noah Snavely, Kavita Bala)
  5. OpenSurfaces – OpenSurfaces consists of tens of thousands of examples of surfaces segmented from consumer photographs of interiors, and annotated with material parameters, texture information, and contextual information . (Kavita Bala et al.)
  6. Osnabrück gaze tracking data – 318 video sequences from several different gaze tracking data sets with polygon based object annotation (Schöning, Faion, Heidemann, Krumnack, Gert, Açik, Kietzmann, Heidemann & König)
  7. PetroSurf3D – 26 high resolution (sub-millimeter accuracy) 3D scans of rock art with pixelwise labeling of petroglyphs for segmentation(Poier, Seidl, Zeppelzauer, Reinbacher, Schaich, Bellandi, Marretta, Bischof)
  8. SYNTHIA – Large set (~half million) of virtual-world images for training autonomous cars to see. (ADAS Group at Computer Vision Center)
  9. Stony Brook University Shadow Dataset (SBU-Shadow5k) – Large scale shadow detection dataset from a wide variety of scenes and photo types, with human annotations (Tomas F.Y. Vicente, Le Hou, Chen-Ping Yu, Minh Hoai, Dimitris Samaras)

Simultaneous Localization and Mapping

  1. Event-Camera Data for Pose Estimation, Visual Odometry, and SLAMThe data also include intensity images, inertial measurements, and ground truth from a motion-capture system. (ETH)
  2. RAWSEEDS SLAM benchmark datasets (Rawseeds Project)
  3. TUM RGB-D Benchmark – Dataset and benchmark for the evaluation of RGB-D visual
    odometry and SLAM algorithms (BCrgen Sturm, Nikolas Engelhard, Felix Endres, Wolfram Burgard and Daniel Cremers)
  4. Visual Odometry / SLAM Evaluation – The odometry benchmark consists of 22 stereo sequences (Andreas Geiger and Philip Lenz and Raquel Urtasun)

Surveillance and Tracking

  1. ATOMIC GROUP ACTIONS dataset – (Ricky J. Sethi et al.)
  2. AVSS07: Advanced Video and Signal based Surveillance 2007 datasets (Andrea Cavallaro)
  3. Activity modeling and abnormality detection dataset – The dataset containes a 45 minutes video with annotated anomalies.(Jagan Varadarajan and Jean-Marc Odobez)
  4. Background subtraction – a list of datasets about background subtraction(Thierry BOUWMANS )
  5. CMUSRD: Surveillance Research Dataset – multi-camera video for indoor surveillance scenario (K. Hattori, H. Hattori, et al)
  6. DukeMTMC: Duke Multi-Target Multi-Camera tracking dataset – 8 cameras, 85 min, 2m frames, 2000 people of video (Ergys Ristani, Francesco Solera, Roger S. Zou, Rita Cucchiara, Carlo Tomasi)
  7. DukeMTMC-reID – A subset of the DukeMTMC for image-based person re-identification (8 cameras,16,522 training images of 702 identities, 2,228 query images of the other 702 identities and 17,661 gallery images.) (Zheng, Zheng, and Yang)
  8. ETISEO Video Surveillance Download Datasets (INRIA Orion Team and others)
  9. HDA+ Multi-camera Surveillance Dataset – video from a network of 18 heterogeneous cameras (different resolutions and frame rates) distributed over 3 floors of a research institute with 13 fully labeled sequences, 85 persons, and 64028 bounding boxes of persons. (D. Figueira, M. Taiana, A. Nambiar, J. Nascimento and A. Bernardino)
  10. Human click data – 20K human clicks on a tracking target (including click errors) (Zhu and Porikli)
  11. Immediacy Dataset – This dataset is designed for estimation personal relationships. (Xiao Chu et al.)
  12. MAHNOB Databases -including Laughter Database,HCI-tagging Database,MHI-Mimicry Database( M. Pantic. etc.)
  13. MVHAUS-PI – a multi-view human interaction recognition dataset (Saeid et al.)
  14. Multispectral visible-NIR video sequences – Annotated multispectral video, visible + NIR (LE2I, Universit de Bourgogne)
  15. Openvisor – Video surveillance Online Repository (Univ of Modena and Reggio Emilia)
  16. Parking-Lot dataset – Parking-Lot dataset is a car dataset which focus on moderate and heavily occlusions on cars in the parking lot scenario.(B. Li, T.F. Wu and S.C. Zhu)
  17. Pornography Database – The Pornography database is a pornography detection dataset containing nearly 80 hours of 400 pornographic and 400 non-pornographic videos extracted from pornography websites and Youtube. (Avila, Thome, Cord, Valle, de Araujo)
  18. Princeton Tracking Benchmark – 100 RGBD tracking datasets (Song and Xiao)
  19. QMUL Junction Dataset 1 and 2 – Videos of busy road junctions. Supports anomaly detection tasks. (T. Hospedales Edinburgh/QMUL)
  20. Queen Mary Multi-Camera Distributed Traffic Scenes Dataset (QMDTS) – The QMDTS is collected from urban surveillance environment for the study of surveillance behaviours in distributed scenes.(Dr. Xun Xu. Prof. Shaogang Gong and Dr. Timothy Hospedales)
  21. SALSA: Synergetic sociAL Scene Analysis – A Novel Dataset for Multimodal Group Behavior Analysis(Xavier Alameda-Pineda etc.)
  22. SBMnet (Scene Background Modeling.NET) – A dataset for testing background estimation algorithms(Jodoin, Maddalena, and Petrosino)
  23. SBM-RGBD dataset – 35 Kinect indoor RGBD videos to evaluate and compare scene background modelling methods for moving object detection (Camplani, Maddalena, Moyà Alcover, Petrosino, Salgado)
  24. SCOUTER – video surveillance ground truthing (shifting perspectives, different setups/lighting conditions, large variations of subject). 30 videos and approximately 36,000 manually labeled frames. (Catalin Mitrea)
  25. SJTU-BESTOne surveillance-specified datasets platform with realistic, on-using camera-captured, diverse set of surveillance images and videos (Shanghai Jiao Tong University)
  26. SPEVI: Surveillance Performance EValuation Initiative (Queen Mary University London)
  27. Shinpuhkan 2014 – A Person Re-identification dataset containing 22,000 images of 24 people captured by 16 cameras. (Yasutomo Kawanishi et al.)
  28. Stanford Drone Dataset – 60 images and videos of various types of agents (not just pedestrians, but also bicyclists, skateboarders, cars, buses, and golf carts) that navigate in a real world outdoor environment such as a university campus (Robicquet, Sadeghian, Alahi, Savarese)
  29. Tracking in extremely cluttered scenes – this single object tracking dataset has 28 highly cluttered sequences with per frame annotation(Jingjing Xiao,Linbo Qiao,Rustam Stolkin,Ale Leonardis)
  30. UCLA Aerial Event Dataset – Human activities in aerial videos with annotations of people, objects, social groups, activities and roles (Shu, Xie, Rothrock, Todorovic, and Zhu)
  31. UCSD Anomaly Detection Dataset – a stationary camera mounted at an elevation, overlooking pedestrian walkways, with unusual pedestrian or non-pedestrian motion.
  32. UCSD trajectory clustering and analysis datasets – (Morris and Trivedi)
  33. USC Information Sciences Institute’s ATOMIC PAIR ACTIONS dataset – (Ricky J. Sethi et al.)
  34. Udine Trajectory-based anomalous event detection dataset – synthetic trajectory datasets with outliers (Univ of Udine Artificial Vision and Real Time Systems Laboratory)
  35. Visual Tracker Benchmark – 100 object tracking sequences with ground truth with Visual Tracker Benchmark evaluation, including tracking results from a number of trackers (Wu, Lim, Yang)
  36. WIDER Attribute Dataset – WIDER Attribute is a large-scale human attribute dataset, with 13789 images belonging to 30 scene categories, and 57524 human bounding boxes each annotated with 14 binary attributes.(Li, Yining and Huang, Chen and Loy, Chen Change and Tang, Xiaoou)

Textures

  1. Brodatz Texture, Normalized Brodatz Texture, Colored Brodatz Texture, Multiband Brodatz Texture 154 new images plus 112 original images with various transformations (A. Safia, D. He)
  2. Color texture images by category (textures.forrest.cz)
  3. Columbia-Utrecht Reflectance and Texture Database (Columbia & Utrecht Universities)
  4. DynTex: Dynamic texture database (Renaud Piteri, Mark Huiskes and Sandor Fazekas)
  5. Intrinsic Images in the Wild (IIW) – Intrinsic Images in the Wild, is a large-scale, public dataset for evaluating intrinsic image decompositions of indoor scenes (Sean Bell, Kavita Bala, Noah Snavely)
  6. KTH TIPS & TIPS2 textures – pose/lighting/scale variations (Eric Hayman)
  7. Materials in Context (MINC) – The Materials in Context Database (MINC) builds on OpenSurfaces, but includes millions of point annotations of material labels. (Sean Bell, Paul Upchurch, Noah Snavely, Kavita Bala)
  8. OpenSurfaces – OpenSurfaces consists of tens of thousands of examples of surfaces segmented from consumer photographs of interiors, and annotated with material parameters, texture information, and contextual information . (Kavita Bala et al.)
  9. Oulu Texture Database (Oulu University)
  10. Oxford Describable Textures Dataset – 5640 images in 47 categories (M.Cimpoi, S. Maji, I. Kokkinos, S. Mohamed, A. Vedaldi)
  11. Prague Texture Segmentation Data Generator and Benchmark (Mikes, Haindl)
  12. Salzburg Texture Image Database (STex) – a large collection of 476 color texture image that have been captured around Salzburg, Austria. (Roland Kwitt and Peter Meerwald)
  13. Texture DatabaseThe texture database features 25 texture classes, 40 samples each(Svetlana Lazebnik, Cordelia Schmid, and Jean Ponce)
  14. Uppsala texture dataset of surfaces and materials – fabrics, grains, etc.
  15. Vision Texture (MIT Media Lab)

Urban Datasets

  1. Barcelona – 15,150 images, urban views of Barcelona (Tighe and Lazebnik)
  2. CMP Facade Database – Includes 606 rectified images of facades from various places with 12 architectural classes annotated.(Radim Tylecek)
  3. LM+SUN – 45,676 images, mainly urban or human related scenes (Tighe and Lazebnik)
  4. MIT CBCL StreetScenes Challenge Framework: (Stan Bileschi)
  5. Queen Mary Multi-Camera Distributed Traffic Scenes Dataset (QMDTS) – The QMDTS is collected from urban surveillance environment for the study of surveillance behaviours in distributed scenes.(Dr. Xun Xu. Prof. Shaogang Gong and Dr. Timothy Hospedales)
  6. Robust Global Translations with 1DSfMthe numerical data describing global structure from motion problems for each dataset (Kyle Wilson and Noah Snavely)
  7. Sift Flow (also known as LabelMe Outdoor, LMO) – 2688 images, mainly outdoor natural and urban (Tighe and Lazebnik)
  8. Street-View Change Detection with Deconvolutional Networks – Database with aligned image pairs from street-view imagery with structural,lighting, weather and seasonal changes.(Pablo F. Alcantarilla, Simon Stent, German Ros, Roberto Arroyo and Riccardo Gherardi)
  9. SydneyHouse – Streetview house images with accurate 3D house shape, facade object label, dense point correspondence, and annotation toolbox.(Hang Chu, Shenlong Wang, Raquel Urtasun,Sanja Fidler)
  10. Traffic Signs Dataset – recording sequences from over 350 km of Swedish highways and city roads (Fredrik Larsson)

Vision and Natural Language

  1. Dataset of Structured Queries and Spatial Relations – Dataset of structured queries about images with the emphasise on spatial relations.(Mateusz Malinowski, Mario Fritz)
  2. Hannah and her sisters database – a dense audio-visual person-oriented ground-truth annotation of faces, speech segments, shot boundaries (Patrick Perez, Technicolor)
  3. Large scale Movie Description Challenge (LSMDC) – A large scale dataset and challenge for movie description, including over 128K video-sentence pairs, mainly sourced from Audio Description (also known as DVS). (Rohrbach, Torabi, Rohrbach, Tandon, Pal, Larochelle, Courville and Schiele)
  4. MPI Movie Description dataset – text and video – A dataset of movie clips associated with natural language descriptions sourced from movie scripts and Audio Description. (Rohrbach, Rohrbach, Tandon and Schiele)
  5. TACoS Multi-Level Corpus – Dataset of cooking videos associated with natural language descriptions at three levels of detail (long, short and single sentence). (Rohrbach, Rohrbach, Qiu, Friedrich, Pinkal and Schiele)
  6. TGIF – 100K animated GIFs from Tumblr and 120K natural language descriptions (Li, Song, Cao, Tetreault, Goldberg, Jaimes, Luo)
  7. Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations – Visual Genome is a dataset, a knowledge base, an ongoing effort to connect structured image concepts to language.(Ranjay Krishna, Yuke Zhu, Oliver Groth, Justin Johnson, Kenji Hata, Joshua Kravitz, Stephanie Chen, Yannis Kalantidis, Li Jia-Li, David Ayman Shamma, Michael Bernstrein, Li Fei-Fei)

Other Collections

  1. 4D Light Field Dataset – 24 synthetic scenes with 9x9x512x512x3 input images, depth and disparity ground truth, camera parameters, and evaluation masks. (Katrin Honauer, Ole Johannsen, Daniel Kondermann, Bastian Goldluecke)
  2. AMADI_LontarSet – Balinese Palm Leaf Manuscript Images Dataset for Binarization, Query-by-Example Word Spotting, and Isolated Character Recognition of Balinese Script. (The AMADI Project et al.)
  3. Annotated Web Ears Dataset (AWE Dataset) – All images were acquired by cropping ears from images from the internet of known persons.( Ziga Emersic, Vitomir Struc and Peter Peer)
  4. Biometrics Evaluation and Testing – Evaluation of identification technologies, including Biometrics( European computing e-infrastructure)
  5. CALVIN research group datasets – object detection with eye tracking, imagenet bounding boxes, synchronised activities, stickman and body poses, youtube objects, faces, horses, toys, visual attributes, shape classes (CALVIN ggroup)
  6. CANTATA Video and Image Database Index site (Multitel)
  7. Chinese University of Hong Kong datasets – Face sketch, face alignment, image search, public square observation, occlusion, central station, MIT single and multiple camera trajectories, person re-identification (Multimedia lab)
  8. Computer Vision Homepage list of test image databases (Carnegie Mellon Univ)
  9. Computer Vision Lab OCR DataBase (CVL OCR DB) – CVL OCR DB is a public annotated image dataset of 120 binary annotated images of text in natural scenes. (Andrej Ikica and Peter Peer.)
  10. ETHZ various, including ETH 3D head pose, BIWI audiovisual data, ETHZ shape classes, BIWI walking pedestrians, pedestrians, buildings, 4D MRI, personal events, liver untrasound, Food 101 (ETH Zurich, Computer Vision Lab)
  11. General 100 Dataset – General-100 dataset contains 100 bmp-format images (with no
    compression), which are well-suited for super-resolution training(Dong, Chao and Loy, Chen Change and Tang, Xiaoou)
  12. HKU-IS -4447 images with pixel labeling groundtruth for salient object detection.(Guanbin Li, Yizhou Yu)
  13. High-res 3D-Models – it includes high-res renderings of these data-sets. ( Hubert etc.)
  14. I3 – Yahoo Flickr Creative Commons 100M – This dataset contains a list of photos and videos. (B. Thomee, D.A. Shamma, G. Friedland et al.)
  15. IDIAP dataset collection – 26 different datasets – multimodal, attack, biometric, cursive characters, discourse, eye gaze, posters, maya codex, MOBIO, face spoofing, game playing, finger vein, youtube-personality traits (IDIAP team)
  16. Kinect v2 dataset – Dataset for evaluating unwrapping in kinect2 depth decoding (Felix etc.)
  17. Laval HDR Sky Database – The database contains 800 hemispherical, full HDR photos of the sky that can be used for outdoor lighting analysis. (Jean-Francois Lalonde et al.)
  18. Leibe’s Collection of people/vehicle/object databases (Bastian Leibe)
  19. Lotus Hill Image Database Collection with Ground Truth (Sealeen Ren, Benjamin Yao, Michael Yang)
  20. MIT Saliency Benchmark dataset – collection (pointers to 23 datasets) (Bylinskii, Judd, Borji, Itti, Durand, Oliva, Torralba}
  21. Michael Firman’s List of RGBD datasets
  22. Msspoof:2D multi-spectral face spoofing – Presentation attack (spoofing) dataset with samples from both real data subjects and spoofed data subjects performed with paper to a NIR and VIS camera(Idiap research institute)
  23. Multiview Stereo Evaluation – Each dataset is registered with a “ground-truth” 3D model acquired via a laser scanning process(Steve Seitz et al)
  24. Oxford Misc, including Buffy, Flowers, TV characters, Buildings, etc (Oxford Visual geometry Group)
  25. PEIPA Image Database Summary (Pilot European Image Processing Archive)
  26. PalmVein spoofing – Presentation attack (spoofing) dataset with samples from spoofed data subjects (corresponding to VERA Palmvein) performed with paper(Idiap research institute)
  27. RSBA dataset – Sequences for evaluating rolling shutter bundle adjustment (Per-Erik etc.)
  28. Replay Attack:2D face spoofing – Presentation attack (spoofing) dataset with samples from both real data subjects and spoofed data subjects performed with paper, photos and videos from a mobile device to a laptop.(Idiap research institute)
  29. Replay Mobile:2D face spoofing – Presentation attack (spoofing) dataset with samples from both real data subjects and spoofed data subjects performed with paper, photos and videos to/from a mobile device.(Idiap research institute)
  30. Synthetic Sequence Generator – Synthetic Sequence Generator (G. Hamarneh)
  31. The Event-Camera Dataset – This presents the world’s first collection of datasets with an event-based camera for high-speed robotics (E. Mueggler, H. Rebecq, G. Gallego, T. Delbruck, D. Scaramuzza)
  32. The world from a cat perspective – videos recorded from the head of a freely behaving cat (Belinda Y. Betsch, Wolfgang Einh?user)
  33. USC Annotated Computer Vision Bibliography database publication summary (Keith Price)
  34. USC-SIPI image databases: texture, aerial, favorites (eg. Lena) (USC Signal and Image Processing Institute)
  35. Univ of Bern databases on handwriting, online documents, string edit and graph matching (Univ of Bern, Computer Vision and Artificial Intelligence)
  36. VERA Fingervein spoofing – Presentation attack (spoofing) dataset with samples from spoofed data subjects (corresponding to VERA Fingervein) performed with paper(Idiap research institute)
  37. VERA Fingervein – Fingervein dataset with data subjects recorded with a open fingervein sensor(Idiap research institute)
  38. VERA PalmVein:PalmVein – Palmvein dataset with data subjects recorded with a open palmvein sensor(Idiap research institute)
  39. Vehicle Detection in Aerial Imagery – VEDAI is a dataset for Vehicle Detection in Aerial Imagery, provided as a tool to benchmark automatic target recognition algorithms in unconstrained environments. (Sebastien Razakarivony and Frederic Jurie)
  40. Video Stacking Dataset – Dataset for evaulating video stacking on cell-phones (Erik Ringaby etc.)
  41. Wrist-mounted camera video dataset – Activities of Daily Living videos captured from a wrist- mounted camera and a head-mounted camera(Katsunori Ohnishi, Atsushi Kanehira,Asako Kanezaki, Tatsuya Harada)
  42. Yummly-10k dataset – The goal was to understand human perception, in this case of food taste similarity.(SE(3) Computer Vision Group at Cornell Tech)

Miscellaneous

  1. 3D mesh watermarking benchmark dataset (Guillaume Lavoue)
  2. 4D Light Field Dataset – 24 synthetic scenes with 9x9x512x512x3 input images, depth and disparity ground truth, camera parameters, and evaluation masks. (Katrin Honauer, Ole Johannsen, Daniel Kondermann, Bastian Goldluecke)
  3. A Dataset for Real Low-Light Image Noise Reduction – It contains pixel and intensity aligned pairs of images corrupted by low-light camera noise and their low-noise counterparts. (J. Anaya, A. Barbu)
  4. AMADI_LontarSet – Balinese Palm Leaf Manuscript Images Dataset for Binarization, Query-by-Example Word Spotting, and Isolated Character Recognition of Balinese Script. (The AMADI Project et al.)
  5. Active Appearance Models datasets (Mikkel B. Stegmann)
  6. Aircraft tracking (Ajmal Mian)
  7. Annotated Web Ears Dataset (AWE Dataset) – All images were acquired by cropping ears from images from the internet of known persons.(Ziga Emersic, Vitomir Struc and Peter Peer)
  8. CITIUS Video Database – A database of 72 videos with eye-tracking data=
    for evaluate dynamic saliency visual models.(Xose)
  9. CVSSP 3D data repository – The datasets are designed to evaluate general
    multi-view reconstruction algorithms. (Armin Mustafa, Hansung Kim, Jean-Yves Guillemaut and Adrian Hilton)
  10. California-ND – 701 photos from a personal photo collection, including many challenging real-life non-identical near-duplicates (Vassilios Vonikakis)
  11. Cambridge Motion-based Segmentation and Recognition Dataset (Brostow, Shotton, Fauqueur, Cipolla)
  12. Catadioptric camera calibration images (Yalin Bastanlar)
  13. Chars74K dataset – 74 English and Kannada characters (Teo de Campos – t.decampos@surrey.ac.uk)
  14. Coin Image Dataset – The coin image dataset is a dataset of 60 classes of Roman Republican coins (Sebastian Zambanini, Klaus Vondrovec)
  15. Columbia Camera Response Functions: Database (DoRF) and Model (EMOR) (M.D. Grossberg and S.K. Nayar)
  16. Columbia Database of Contaminants’ Patterns and Scattering Parameters (Jinwei Gu, Ravi Ramamoorthi, Peter Belhumeur, Shree Nayar)
  17. COVERAGE – copy-move forged (CMFD) images and their originals with similar but genuine objects (SGOs), which highlight and address tamper detection ambiguity of popular methods, caused by self-similarity within natural images (Wen, Zhu, Subramanian, Ng, Shen, and Winkler)
  18. Crime Scene Footwear Impression Database – crime scene and reference foorware impression images (Adam Kortylewski)
  19. D-HAZY – : A DATASET TO EVALUATE QUANTITATIVELY DEHAZING ALGORITHMS (Cosmin Ancuti et al.)
  20. DR(eye)VE – A driver’s attention dataset (University of Modena and Reggio Emilia)
  21. DTU controlled motion and lighting image dataset (135K images) (Henrik Aanaes)
  22. Database for Visual Eye Movements (DOVES) – A set of eye movements collected from 29 human observers as they viewed 101 natural calibrated images. (van der Linde, I., Rajashekar, U., Bovik, A. C. etc.)
  23. DeformIt 2.0 – Image Data Augmentation Tool: Simulate novel images with ground truth segmentations from a single image-segmentation pair (Brian Booth and Ghassan Hamarneh)
  24. Dense outdoor correspondence ground truth datasets, for optical flow and local keypoint evaluation (Christoph Strecha)
  25. EISATS: .enpeda.. Image Sequence Analysis Test Site (Auckland University Multimedia Imaging Group)
  26. Featureless object tracking – This dataset contains several videosequences with limited texture, intended for visual tracking, including manually annotated per-frame pose.(Lebeda, Hadfield, Matas, Bowden)
  27. FlickrLogos-32 – 8240 images of 32 product logos (Stefan Romberg)
  28. General 100 Dataset – General-100 dataset contains 100 bmp-format images (with no
    compression), which are well-suited for super-resolution training(Dong, Chao and Loy, Chen Change and Tang, Xiaoou)
  29. Geometry2view – This dataset contains image pairs for 2-view geometry computation, including manually annotated point coordinates.(Lebeda, Matas, Chum)
  30. Hannover Region Detector Evaluation Data Set – Feature detector evaluation sequences in multiple image resolutions from 1.5 up to 8 megapixels (Kai Cordes)
  31. Hillclimb and CubicGlobe datasets – a video of a rally car, separated into several independent shots (for visual tracking and modelling). (Lebeda, Hadfield, Bowden)
  32. Houston Multimodal Distracted Driving Dataset – 68 volunteers that drove the same simulated highway under four different conditions (Dcosta, Buddharaju, Khatri, and Pavlidis)
  33. I3 – Yahoo Flickr Creative Commons 100M – This dataset contains a list of photos and videos. (B. Thomee, D.A. Shamma, G. Friedland et al.)
  34. ICDAR’15 Smartphone document capture and OCR competition – challenge 2 – pictures of documents captured with smartphones under various conditions of perspective, lighting, etc. The ground truth is the textual content which should be extracted. (Burie, Chazalon, Coustaty, Eskenazi, Luqman, Mehri, Nayef, Ogier, Prum and Rusinol)
  35. IISc – Dissimilarity between Isolated Objects (IISc-DIO) – The dataset has a total of 26,675 perceived dissimilarity measurements made on 269 human subjects using a Visual Search task with a diverse set of objects.(RT Pramod & SP Arun, IISc)
  36. INRIA feature detector evaluation sequences (Krystian Mikolajczyk)
  37. INRIA’s PERCEPTION’s database of images and videos gathered with several synchronized and calibrated cameras (INRIA Rhone-Alpes)
  38. Image/video quality assessment database summary (Stefan Winkler)
  39. KITTI dataset for stereo, optical flow and visual odometry (Geiger, Lenz, Urtasun)
  40. LFW-10 dataset for learning relative attributes – A dataset of 10,000 pairs of face images with
    instance-level annotations for 10 attributes.(CVIT, IIIT Hyderabad. )
  41. Large scale 3D point cloud data from terrestrial LiDAR scanning (Andreas Nuechter)
  42. Light-field Material Dataset – 1.2k annotated images of 12 material classes taken with the Lytro ILLUM camera(Ting-Chun Wang, Jun-Yan Zhu, Ebi Hiroaki,Manmohan Chandraker, Alexei Efros, Ravi Ramamoorthi)
  43. Linkoping Rolling Shutter Rectification Dataset (Per-Erik Forssen and Erik Ringaby)
  44. LIRIS-ACCEDE Dataset – a collection of video excerpts with a large content diversity annotated along affective dimensions (Technicolor)
  45. MARIS Portofino dataset – A dataset of underwater stereo images depicting cylindrical pipe objects and collected to test object detection and pose estimation algorithms. (RIMLab (Robotics and Intelligent Machines Laboratory), University of Parma.)
  46. MPI Sintel Flow Dataset A data set for the evaluation of optical flow derived from the open source 3D animated short film, Sintel. It has been extended for Stereo and disparity, Depth and camera motion, and Segmentation. (Max Planck Tubingen)
  47. MPI-Sintel optical flow evaluation dataset (Michael Black)
  48. MSR-VTT – video to text database of 200K+ video clip/sentence pairs
  49. Middlebury College stereo vision research datasets (Daniel Scharstein and Richard Szeliski)
  50. Modelling of 2D Shapes with Ellipses – he dataset contains 4,526 2D shapes included in standard as well as in home-build datasets.(Costas Panagiotakis and Antonis Argyros)
  51. Multi-FoV – photo-realistic video sequences that allow benchmarking of the impact of the Field-of-View (FoV) of the camera on various vision tasks. (Zhang, Rebecq, Forster, Scaramuzza)
  52. Multiview Stereo Evaluation – Each dataset is registered with a “ground-truth” 3D model acquired via a laser scanning process(Steve Seitz et al)
  53. Multiview stereo images with laser based groundtruth (ESAT-PSI/VISICS,FGAN-FOM,EPFL/IC/ISIM/CVLab)
  54. NCI Cancer Image Archive – prostate images (National Cancer Institute)
  55. NIST 3D Interest Point Detection (Helin Dutagaci, Afzal Godil)
  56. NRCS natural resource/agricultural image database (USDA Natural Resources Conservation Service)
  57. Object Removal – Generalized Dynamic Object Removal for Dense Stereo Vision Based Scene Mapping using Synthesised Optical Flow – Evaluation Dataset (Hamilton, O.K., Breckon, Toby P.)
  58. Occlusion detection test data (Andrew Stein)
  59. OSIE – Object and Semantic Images and Eye-tracking – 700 images, 5551 segmented objects, eye tracking data (Xu, Jiang, Wang, Kankanhalli, Zhao)
  60. OTIS: Open Turbulent Image Set – several sequences (either static or dynamic) of long distance imaging through a turbulent atmosphere (Jerome Gilles, Nicholas B. Ferrante)
  61. PHOS (illumination invariance dataset) – 15 scenes captured under different illumination conditions * 15 images (Vassilios Vonikakis)
  62. PRINTART: Artistic images of prints of well known paintings, including detail annotations. A benchmark for automatic annotation and retrieval tasks with this database was published at ECCV. (Nuno Miguel Pinho da Silva)
  63. Pics ‘n’ Trails – Dataset of Continuously archived GPS and digital photos (Gamhewage Chaminda de Silva)
  64. Pitt Image and Video Advertisement Understanding – rich annotations encompassing the topic and sentiment of the ads, questions and answers describing what actions the viewer is prompted to take and the reasoning that the ad presents to persuade the viewer (Hussain, Zhang, Zhang, Ye, Thomas, Agha, Ong, Kovashka (University of Pittsburgh)>
  65. RAWSEEDS SLAM benchmark datasets (Rawseeds Project)
  66. ROMA (ROad MArkings) : Image database for the evaluation of road markings extraction algorithms (Jean-Philippe Tarel, et al)
  67. Robotic 3D Scan Repository – 3D point clouds from robotic experiments of scenes (Osnabruck and Jacobs Universities)
  68. Rolling Shutter Rectification Dataset – Rectifying rolling shutter video from hand-held devices (Per-Erik etc.)
  69. SALICON – Saliency in Context eye tracking dataset c. 1000 images with eye-tracking data in 80 image classes (Jiang, Huang, Duan, Zhao)
  70. SIDIRE: Synthetic Image Dataset for Illumination Robustness Evaluation – SIDIRE is a freely available image dataset which provides synthetically generated images allowing to investigate the influence of illumination changes on object appearance (Sebastian Zambanini)
  71. Scripps Plankton Camera System – thousands of images of c. 50 classes of plankton and other small marine objects (Jaffe et al)
  72. Smartphone document capture and OCR 2015 – Quality Assessment – pictures of documents captured with smartphones under various conditions perspective, lighting, etc. It also features text ground truth and OCR accuracies to train and test document image quality assessment systems. (Nayef, Luqman, Prum, Eskenazi, Chazalon, and Ogier)
  73. Smartphone document capture and OCR 2017 – mobile video capture – video recording of documents, along with the reference ground truth image to reconstruct using the video stream. (Chazalon, Gomez-Krämer, Burie, Coustaty, Eskenazi, Luqman, Nayef, Rusiñol, Sidère, and Ogier)
  74. Stony Brook Univeristy Real-World Clutter Dataset (SBU-RwC90) – Images of different level of clutterness, ranked by humans (Chen-Ping Yu, Dimitris Samaras, Gregory Zelinsky)
  75. Street-View Change Detection with Deconvolutional Networks – Database with aligned image pairs from street-view imagery with structural,lighting, weather and seasonal changes.(Pablo F. Alcantarilla, Simon Stent, German Ros, Roberto Arroyo and Riccardo Gherardi)
  76. SydneyHouse – Streetview house images with accurate 3D house shape, facade object label, dense point correspondence, and annotation toolbox.(Hang Chu, Shenlong Wang, Raquel Urtasun,Sanja Fidler)
  77. Technicolor Interestingness Dataset – a collection of movie excerpts and key-frames and their corresponding ground-truth files based on the classification into interesting and non-interesting samples (Technicolor)
  78. Technicolor Hannah Dataset – 153,825 frames from the movie “Hannah and her sisters” annotated for several types of audio and visual information (Technicolor)
  79. Technicolor HR-EEG4EMO Dataset – EEG and other physiological recordings of 40 subjects collected during the viewing of neutral and emotional videos (Technicolor)
  80. Technicolor VSD Violent Scenes Dataset – a collection of ground-truth files based on the extraction of violent events in movies (Technicolor)
  81. TMAGIC dataset – Several videosequences for visual tracking, containing strong out-of-plane rotation(Lebeda, Hadfield, Bowden)
  82. TUM RGB-D Benchmark – Dataset and benchmark for the evaluation of RGB-D visual
    odometry and SLAM algorithms (BCrgen Sturm, Nikolas Engelhard, Felix Endres, Wolfram Burgard and Daniel Cremers)
  83. The Conflict Escalation Resolution (CONFER) Database – 120 audio-visual episodes (~142 mins) of naturalistic interactions from televised political debates, annotated frame-by-frame in terms of real-valued conflict intensity. (Christos Georgakis, Yannis Panagakis, Stefanos Zafeiriou,Maja Pantic)
  84. The Open Video Project (Gary Marchionini, Barbara M. Wildemuth, Gary Geisler, Yaxiao Song)
  85. The Toulouse Vanishing Points Dataset – a dataset of Manhattan scenes for vanishing point estimation which also provide, for each image, the IMU data of the camera orientation.(Vincent Angladon and Simone Gasparini)
  86. UCL Ground Truth Optical Flow Dataset (Oisin Mac Aodha)
  87. Univ of Genoa Datasets for disparity and optic flow evaluation (Manuela Chessa)
  88. Validation and Verification of Neural Network Systems (Francesco Vivarelli)
  89. Very Long Baseline Interferometry Image Reconstruction Dataset (MIT CSAIL)
  90. Virtual KITTI – 40 high-resolution videos (17,008 frames) generated from five different virtual worlds, for : object detection and multi-object tracking, scene-level and instance-level semantic segmentation, optical flow, and depth estimation (Gaidon, Wang, Cabon, Vig)
  91. Visual Object Tracking challenge – This challenge is held annually as an ICCV/ECCV workshop, with a new dataset and an updated evaluation kit every year.(Kristan et al.)
  92. WHOI-Plankton – 3.5 million images of microscopic marine plankton on 103 categories (Olson, Sosik)
  93. WILD: Weather and Illumunation Database (S. Narasimhan, C. Wang. S. Nayar, D. Stolyarov, K. Garg, Y. Schechner, H. Peri)
  94. YACCLAB dataset – YACCLAB dataset includes both synthetic and real binary images(Grana, Costantino; Bolelli, Federico; Baraldi, Lorenzo; Vezzani, Roberto)
  95. YtLongTrack – This dataset contains two video sequences with challenges such as low quality, extreme length and full occlusions, including manually annotated per-frame pose.(Lebeda, Hadfield, Matas, Bowden)

Acknowledgements: Many thanks to all of the contributors for their suggestions of databases.
Can PU was very helpful with the updating of this web page.

Source – http://homepages.inf.ed.ac.uk/rbf/CVonline/Imagedbase.htm