HAUTDatasets

The various shared datasets are listed below as links leading to sections with their details and illustrations:

CD-MaJ | SVMIS+ | FULLSCAN | PanoraMIS | ConveJRL | ArUcOmni | LFMIS | AFAMIS

HAUTCD-MaJ: the Color-Depth dataset of MIS lab and JRL

The CD-MaJ dataset was created for the evaluation of visual Simulatenous Localization and Mapping with RGB-D camera of various fields-of-view and resolutions with respect to ground truth measurements obtained by motion capture. It was generated with an Azure Kinect camera mounted on a Pioneer 3AT mobile robot moving in the indoor area covered by an Optitrack system which readings are synchronized with camera captures.

Pioneer 3AT mobile robot embedding an Azure Kinect RGB-D camera    Color image    Depth image

Seven RGB-D image sequences are shared from images captured while the mobile robot was moving. Various modes of the Azure Kinect cameras are considered: narrow or wide depth camera field-of-view, high or low resolution, at 15 or 30 images per second (except for wide field-of-view 30 images per second that is not possible). The motion capture data synchronized (software via ROS) with the image capture are also provided in rosbags.

More details on the camera frames and pose representation are within the CD-MaJ dataset archive (15.4 GB) to be downloaded from: CD-MaJ shared directory.

This dataset is published with the following paper:

Eva Goichon, Guillaume Caron, Pascal Vasseur, Fumio Kanehiro. On camera model conversions. in Proc. of IEEE International Conference on Robotics and Automation, May 2024, Yokohama, Japan. PDF

HAUTSVMIS+: Spherical Vision dataset onboard hexarotor

The SVMIS+ dataset, created for the evaluation of visual gyroscopes in outdoor environment with respect to ground truth measurements, was generated with a Ricoh Theta S dual-fisheye camera mounted on a DJI Matrice 600 Pro hexarotor drone with additional onboard computer to record images synchronized with IMU and GPS readings.

Matrice 600 Pro drone embedding a dual fisheye camera    Equirectangular image 1074

An equirectangular image sequence is shared from images captured while the drone was flying, thus rotating and translating the camera (4'36'' video at 10 FPS). The drone flight data (IMU and GPS measurements) synchronized (software via ROS) with the image capture are also provided in the form of panoramic equirectangular camera poses, meaning with the sky in the top part of the image and the ground in the bottom part of the image, when the drone is static (horizon line straight and horizontal in the image).

More details on the camera frames and camera pose representation are within the SVMIS+ dataset archive (3.1 GB) to be downloaded from: SVMISplus_er_pano.zip.

This dataset is published with the following paper:

Bruno Berenguel-Baeta, Antoine André, Guillaume Caron, Jesus Bermudez-Cameo and Josechu Guerrero, Visual Gyroscope: Combination of Deep Learning Features and Direct Alignment for Panoramic Stabilization, in Proc. of IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (OmniCV), pp.6444-6447, June 2023, Vancouver, Canada. PDF

HAUTFULLSCAN: FullScan project dataset

The FULLSCAN dataset, created for studying the nD reconstruction of stained-glasses, was generated with seven different sensors:

  • Leica C10 scan station
  • Faro Focus 3D laser scanner
  • Ricoh Theta V dual-fisheye camera
  • Trioptics PolarCam 4D Technology V polarimetric camera
  • Raytrix R8 plenoptic live camera
  • Ocean Optics Maya2000 Pro with Ocular Robotics RobotEye REHS25 hyperspectral imager
  • Texas Instruments OPT8241-CDK-EVM multi-frequency time-of-flight camera
Dual fisheye camera on a tripod near a window of the Amiens Cathedral    Wisk-broom hyperspectral camera scanning a stained-glass window in the Amiens Cathedral

Data were acquired and structured by MIS (UPJV, France), OMI (NAIST, Japan), VIBOT (UBFC, France) and IRSEEM (Esigelec, France).

For the moment (June 2022), only hyperspectral data (raw and corrected) of the REHS sensor (NAIST) and full-view equirectangular images of the Ricoh Theta V camera (MIS) are shared for three of the main stained-glass windows of the Choir and Transept parts of the Amiens cathedral triforium:

Stained-glass ThetaV REHS
params raw comp
III
XVIII
SouthPortal

Stained-glass III represents St Peter, XVIII represents St Bishop.

In the HSI data from the REHS sensor, "params" are the angular directions of captured pixels and the wavelengths of the 2068 channels (time is provided for information), "raw" stands for the raw data captured (hyperspectral cube) and "comp" stands for the compensated illumination variation HSI data. When using these spectral data, please cite:

T. Funatomi, T. Ogawa, K. Tanaka, H. Kubo, G. Caron, E. Mouaddib, Y. Matsushita, Y. Mukaigawa, Eliminating Temporal Illumination Variations in Whisk-broom Hyperspectral Imaging, International Journal of Computer Vision, IJCV, vol. 130, 1310-1324, 2022. PDF

When using the equirectangular images, please cite:

G. Caron, S. J. Kessy, Y. Mukaigawa, T. Funatomi, Direct alignment of narrow field-of-view hyperspectral data and full-view RGB image, IEEE International Conference on Image Processing, ICIP, 2022.

C++ code to densely align compensated HSI data with RGB equirectangular images is also provided at: github.com/jrl-umi3218/hsrgbalign.

HAUTPanoraMIS: Panoramic Vision of the MIS laboratory

The PanoraMIS dataset is the gathering and extension of previous OVMIS and SVMIS datasets (kept at the bottom of this page for legacy) to spherical vision on long paths of a Seekur Jr mobile robot in various urban and natural environments with synchronized position and 3D orientation ground truth.

Dedicated website: mis.u-picardie.fr/~panoramis.

For more details on the PanoraMIS dataset (e.g. acquisition setup or paths shape), see article:

Houssem-Eddine Benseddik, Fabio Morbidi, Guillaume Caron, PanoraMIS: An Ultra-wide Field of View Image Dataset for Vision-based Robot-Motion Estimation, SAGE International Journal of Robotics Research, IJRR, vol. 39, n. 9, pp.1037-1051, 2020. PDF

HAUTConveJRL: Convenience store objects dataset of JRL

The ConveJRL dataset has been made at CNRS-AIST JRL (Tsukuba, Japan). This dataset was created for evaluating cost functions of eye-to-hand visual servoing of a robot arm for object manipultion. It contains images of a webcam observing 13 axial-symmetric objects undergoing pure rotations on a turn table, each for 360 degrees at a step of less than 1 degree.

The ConveJRL dataset features 6850 images.

For more details on the ConveJRL dataset please see the following paper (to cite if using the dataset):

Guillaume Caron, Yusuke Yoshiyasu, Direct visual servoing in the non-linear scale space of camera pose, IAPR Int. Conf. on Pattern Recognition, ICPR, August 2022. PDF

The ConveJRL dataset (4 GB) is available for download here: ConveJRL.zip.

HAUTArUcOmni: Panoramic dataset of ArUco markers

The ArUcOmni dataset is a collaboration between L@bISEN, Vision-AD team, Yncréa Ouest, ISEN (Brest, France) and the MIS lab of UPJV. This dataset was created for evaluating the adaptation of the detection and pose estimation of the so called ArUco fiducial markers to panoramic vision. It contains images of a hypercatadioptric camera and a fisheye camera observing a 3-plane rig displaying ArUco markers of which the relative poses serve as ground truth.

  

Target images showing calibration checkerboards are provided to calibrate both hypercatadioptric and fisheye cameras.

The ArUcOmni dataset features 189 hypercatadioptric images and 36 fisheye images. Matlab scripts are provided to compute the pose estimation errors with respect to the ground truth.

For more details on the ArUcOmni dataset please see the following paper:

Jaouad Hajjami, Jordan Caracotte, Guillaume Caron, Thibault Napoléon, ArUcOmni: detection of highly reliable fiducial markers in panoramic images, IEEE Conference on Computer Vision and Pattern Recognition (CVPR) workshop on Omnidirectional Computer Vision, OmniCV, June 2020.

The ArUcOmni dataset (486 MB) is available for download here: ArUcOmni_dataset.zip.

HAUTLFMIS: Light-Field dataset of the MIS laboratory

The LFMIS dataset, created for the evaluation of planar object visual tracking, was generated with a static Lytro Photo camera pointed toward the end-effector of a Stäubli manipulator holding a planar target.

Light-field camera pointed toward an industrial manipulator

Target images (12) showing the calibration rigs are provided to calibrate the light-field camera.

Two experimental setups with a textured planar object led to the following datasets:

  • Sequence 1: pure translation, rectangular shape (9 light-fields)
  • Sequence 2: free combination of translation and rotation (9 light-fields)

For each sequence, the ground truth poses of the planar object measured by the Stäubli manipulator are provided.

For more details on the LFMIS dataset (e.g. acquisition setup or paths shape), see Section VI in the paper:

Nathan Crombez, Guillaume Caron, Takuya Funatomi, Yasuhiro Mukaigawa, Reliable planar object pose estimation in light-fields from best sub-aperture camera pairs, IEEE Robotics and Automation Letters, RA-L, vol. 3, no. 8, pp. 3561 - 3568, October 2018. PDF

The LFMIS dataset (2.24 GB) is available for download here: LFMIS_dataset.zip.

HAUTAFAMIS: Image templates dataset for registration evaluation

The AFAMIS dataset, created for the evaluation of image region tracking, under translation and projective motion model, was generated from the MS-COCO dataset and the Yale face database. We built a set of 110000 images to evaluate our AFA-LK tracker (Adaptive Forward Additive Lucas-Kanade tracker) with respect to state of the art methods in:

Yassine Ahmine, Guillaume Caron, El Mustapha Mouaddib, Fatima Chouireb, Adaptive Lucas-Kanade tracking, Elsevier Image and Vision Computing, IMAVIS, vol. 88, pp. 1 - 8, August 2019. PDF

The AFAMIS dataset (520 MB) is available for download here: AFAMIS_dataset.zip.

HAUTSVMIS: Spherical Vision dataset of the MIS laboratory

The SVMIS dataset, created for the evaluation of visual gyroscopes in structured and unstructured environments, was generated with a dual-fisheye camera mounted on the end-effector of a Stäubli manipulator (TX-60) and on a Parrot fixed-wing drone (Disco FPV).

Dual fisheye camera mounted on an industrial manipulator     Disco drone embedding a dual fisheye camera

Multi-plane target images are provided to calibrate the dual-fisheye camera.

Two image datasets acquired with an industrial manipulator are provided (the ground truth is given by robot's odometry):

  • OneDOF: 720 images obtained by rotating the camera with a 2.5° step size about its "vertical" axis leading to 144 images at 5 collection points.
  • ThreeDOFs: 94 images taken at a unique collection point.

An experimental scenario with the Disco drone is also considered leading to an outdoor image sequence in which the camera is rotated and translated (6'41'' video at 30 FPS for a 4.2 km trajectory). The drone flight data (including the IMU measurements) are also provided but not synchronized with the dual-fisheye camera.

For more details about the organization of the SVMIS dataset, see Section V in the paper:

Guillaume Caron and Fabio Morbidi, Spherical Visual Gyroscope for Autonomous Robots using the Mixture of Photometric Potentials, in Proc. of IEEE Int. Conf. on Robotics and Automation, ICRA, pp. 820-827, May 2018. PDF

The SVMIS dataset (1.6 GB) is available for download here: SVMIS_dataset.zip.

Due to requests, we slightly extended the SVMIS dataset. Two videos (810 MB each) that are the equirectangular conversion results of the dual-fisheye video acquired with the Disco drone are now available. The conversion is done with the Ricoh Theta desktop software with or without the top/down correction:

HAUTOVMIS: Omnidirectional Vision dataset of the MIS laboratory

The OVMIS dataset, created for the evaluation of visual compasses in structured and unstructured environments, was generated with a hypercatadioptric camera mounted on the end-effector of a Stäubli manipulator and on a Pioneer mobile robot.

Hypercatadioptric camera mounted on an industrial manipulator     Pioneer 3AT mobile robot embedding a hypercatadoptric camera

Target images showing the calibration rigs are provided to calibrate the hypercatadioptric camera.

3600 images sampling a horizontal disk of the manipulator workspace with the camera frame poses measured by the Stäubli manipulator are provided. These 3600 images are obtained from a pure rotation of the camera with a 2.5° step around its optical axis leading to 144 images at 25 collection positions.

Furthermore, three experimental scenarios with the Pioneer mobile robot led to the following datasets:

  • Scenario 1: indoor, pure rotation (160 images acquired during 49.69 s)
  • Scenario 2: outdoor, pure rotation (156 images acquired during 50.22 s)
  • Scenario 3: outdoor, rotation and translation (318 images acquired during 160.99 s for a 25.27 m length trajectory: mean speed 0.187 m/s)

For each scenario, the ground truth (Pioneer's odometry corrected with the gyroscopic measurements) is provided.

For more details on the organization of the OVMIS dataset, see Section IVA in the paper:

Fabio Morbidi and Guillaume Caron, Phase Correlation for Dense Visual Compass from Omnidirectional Camera-Robot Images, IEEE Robotics and Automation Letters, RA-L, vol. 2, no. 2, pp. 688 - 695, Avril 2017. PDF

The OVMIS dataset (2.72 GB) is available for download here: OVMIS_dataset.zip.