GitHub - s-gupta/rgbd at dev (original) (raw)
This is dev branch and will include bug fixes and latest code.
This is code to generate the bottom up segmentation, amodal completion, semantic segmentation and scene classification results for our CVPR 13 paper, Perceptual Organization and Recognition of Indoor Scenes from RGB-D Images. Saurabh Gupta, Pablo Arbelaez, Jitendra Malik, In CVPR 2013
Notes
- External Dependencies: are now available from the git under the folder external. You can use the scripts external/build_external.sh to build them. These libraries include: a. gPb-UCM from (http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/BSR/BSR_full.tgz)
b. VLFeat from http://www.vlfeat.org/
c. liblinear from http://www.csie.ntu.edu.tw/~cjlin/liblinear/
d. liblinear dense from http://ttic.uchicago.edu/~smaji/projects/digits/
e. SIFT color desprictors from http://www.colordescriptors.com, We used v2.1.
f. Image Stack Library https://code.google.com/p/imagestack/, for computing the local deph boundary cues.
- You can run the system on a new pair of RGB-D image by using the runAll.m function.
- The code is intended for use with the data available here. This contains the color images, preprocessed point clouds and the preprocessed ground truth.http://www.cs.berkeley.edu/~sgupta/cvpr13/data.tgz . Please modify COM/getPaths.m and allBenchmarks/benchmarkPaths.m to point to where you put the data folder. You can use the default paths and put the extracted folders appropriately in the directory data.
- We also provide the all precomputed results, http://www.cs.berkeley.edu/~sgupta/cvpr13/output.tgz, and the pretrained models, http://www.cs.berkeley.edu/~sgupta/cvpr13/model.tgz.
Data Preprocessing
- The data that we provide is different from the one provided by Silberman etal in the following ways:
a. We use I(46:470, 41:600, :) part of the image. This is different from what they use by 1 pixel on each side, I think.
b. We use the original point clouds from the Kinect from over herehttp://horatio.cs.nyu.edu/mit/silberman/nyu_depth_v2/orig_images.mat, and project them to the RGB frame but instead of just keeping the Z coordinate also keep the X and the Y coordinate (keeping just the Z and projecting back rays to get X and Y introduces a little bit of error, but should still work well, we did not try using that...)
c. Finally, we preprocessed the ground truth to get rid of the double contours.
Precomputed Results
- http://www.cs.berkeley.edu/~sgupta/cvpr13/output.tgz: includes the folder cachedir/release/output which contains the UCM, amodal completion and the semantic segmentation outputs. Their are also .mat files with the performance of our system.
Pre trained Models
- http://www.cs.berkeley.edu/~sgupta/cvpr13/model.tgz, includes the folder cachedir/release/model which has the pretrained ucm models, various semantic segmentation and scene classification models.
Questions and Comments
Please feel free to email, sgupta at eecs dot berkeley dot edu for any questions and comments.