The Data61/2D3D Dataset is made freely available to the scientific community via this web address www.nicta.com.au/computer_vision_datasets. Any publications resulting from the use of this dataset, and derivative forms, should cite the following paper:
S. Taghavi Namin, M. Najafi, M. Salzmann and L. Petersson, A Multi-Modal Graphical Model for Scene Analysis, WACV, 2015.
Data61/2D3D dataset has been prepared for outdoor scene understanding which consists of a series of 2D panoramic images with corresponding 3D LIDAR point clouds. It contains 10 outdoor scenes, each of which includes a block of 3D point cloud together with several panoramic images. The number of 3D points in the scenes varies from 1 to 2 millions, and each scene contains between 11 and 21 panoramic images.
The dataset was manually annotated in the 3D domain and the ground truth labeling of the panoramic images were obtained via 3D-2D projection of the 3D labels. The 2D ground truth images were later checked and retouched to produce a more precise 2D ground truth. This step accounts for projection errors due to misalignments or parallax and also deals with moving and/or reflective objects whose point cloud data is very sparse. Additionally, the label Sky, which does not exist in the 3D data, was included as a new label in the 2D images. The point cloud data is seen from multiple viewpoints thanks to the 360 degree panoramic images, and its FOV covers the entire image (both vertically and horizontally) instead of just a portion in other datasets.
The second advantage of the Data61/2D3D data over the aforementioned datasets is that the panoramic images provide an opportunity to capture each object several times in different frames and from different viewpoints. Therefore it not only provides the corresponding 2D information for each 3D segment, but also does it several times from different views.
Click here to download the Data61/2D3D Dataset. (Please note that some file sizes are quite large and may take time to download).
For further information on the Pedestrian Dataset, please contact Sarah.Namin@nicta.com.au or Lars.Petersson@nicta.com.au
The Data61 Pedestrian Dataset is made freely available to the scientific community via this web address www.nicta.com.au/computer_vision_datasets. Any publications resulting from the use of this dataset, and derivative forms, should cite the following paper:
G. Overett, L. Petersson, N. Brewer, L. Andersson and N. Pettersson, A New Pedestrian Dataset for Supervised Learning, In IEEE Intelligent Vehicles Symposium, 2008.
The final dataset contains 25551 unique pedestrians, allowing for a dataset of over 50K images with mirroring. Additionally, TDB allows the generation of several permutations per source image in order to further bolster the training set. Large negative datasets will also be provided, although researchers training cascaded classifiers may require their own bootstrapped negative datasets. Apart from the datasets linked here the authors are willing, within reason, to produce further pedestrian datasets for the scientific community.
Figure 1. shows a selection of pedestrians from the dataset. Most images were captured using normal digital camera hardware in normal urban environments, in multiple cities and in different countries.
The negative set is drawn from a set of 5207 high resolution pedestrian free images in varied environments. Both the negative and positive sets are divided into unique folds for tasks such as validation.
Click here to download the Pedestrian Dataset. (Please note that some file sizes are quite large and may take time to download).
For further information on the Pedestrian Dataset, please contact Gary.Overett@nicta.com.au or Lars.Petersson@nicta.com.au
This dataset (‘Licensed Material’) is made available to the scientific community for non-commercial research purposes such as academic research, teaching, scientific publications or personal experimentation. Permission is granted by National ICT Australia Limited (Data61) to you (the ‘Licensee’) to use, copy and distribute the Licensed Material in accordance with the following terms and conditions:
The code for Support Vector Registration (SVR) is made freely available to the scientific community. Any publications resulting from the use of this code should cite the following paper:
Campbell, Dylan, and Petersson, Lars, “An Adaptive Data Representations for Robust Point-Set Registration and Merging”, International Conference on Computer Vision (ICCV), Santiago, Chile, IEEE, Dec. 2015
This paper presents a framework for rigid point-set registration and merging using a robust continuous data representation. Our point-set representation is constructed by training a one-class support vector machine with a Gaussian radial basis function kernel and subsequently approximating the output function with a Gaussian mixture model. We leverage the representation’s sparse parametrisation and robustness to noise, outliers and occlusions in an efficient registration algorithm that minimises the L2 distance between our support vector–parametrised Gaussian mixtures. In contrast, existing techniques, such as Iterative Closest Point and Gaussian mixture approaches, manifest a narrower region of convergence and are less robust to occlusions and missing data, as demonstrated in the evaluation on a range of 2D and 3D datasets. Finally, we present a novel algorithm, GMMerge, that parsimoniously and equitably merges aligned mixture models, allowing the framework to be used for reconstruction and mapping.
Click here to download the paper.
Click here to download the code for Support Vector Registration. MATLAB and C++ versions of the code are included. The C++ version is significantly faster, but the MATLAB optimisation package is more powerful and so tends to be more stable and accurate. The svr.zip file contains README files in the distribution/MATLAB/demo and distribution/C++/build folders. The license can be viewed in the license.txt file in the root directory. For further information on the code, please contact email@example.com or firstname.lastname@example.org.
The code for Globally-Optimal Gaussian Mixture Alignment (GOGMA) is made freely available to the scientific community. Any publications resulting from the use of this code should cite the following papers:
Campbell, Dylan, and Petersson, Lars, “GOGMA: Globally-Optimal Gaussian Mixture Alignment”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, USA, IEEE, Jun. 2016.
Campbell, Dylan, and Petersson, Lars, “An Adaptive Data Representation for Robust Point-Set Registration and Merging”, IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, IEEE, Dec. 2015, pp. 4292-4300.
Gaussian mixture alignment is a family of approaches that are frequently used for robustly solving the point-set registration problem. However, since they use local optimisation, they are susceptible to local minima and can only guarantee local optimality. Consequently, their accuracy is strongly dependent on the quality of the initialisation. This paper presents the first globally-optimal solution to the 3D rigid Gaussian mixture alignment problem under the L2 distance between mixtures. The algorithm, named GOGMA, employs a branch-and-bound approach to search the space of 3D rigid motions SE(3), guaranteeing global optimality regardless of the initialisation. The geometry of SE(3) was used to find novel upper and lower bounds for the objective function and local optimisation was integrated into the scheme to accelerate convergence without voiding the optimality guarantee. The evaluation empirically supported the optimality proof and showed that the method performed much more robustly on two challenging datasets than an existing globally-optimal registration solution.
Click here to download the code for Globally-Optimal Gaussian Mixture Alignment. C++ and CUDA code is included. The gogma zip file contains a README.txt file in the root folder. The license can be viewed in the license.txt file in the root directory. For further information on the code, please contact email@example.com or firstname.lastname@example.org.