CMVS + PMVS

Posted by Tong on April 19, 2020

Yasutaka Furukawa

CMVS

PMVS

OpenMVG to PMVS

Application

Installation

git clone https://github.com/pmoulon/CMVS-PMVS.git
cd program
mkdir build
cd build
cmake ..
make -j  

Execution

cd ~/Application/openMVG/build/

# Structure from Motion (Note: Dataset needs to be downloaded previously)
python software/SfM/SfM_SequentialPipeline.py software/SfM/ImageDataset_SceauxCastle/images/ software/SfM/ImageDataset_SceauxCastle/Castle_Incremental_Reconstruction/

# Data Conversion 
./Linux-x86_64-Release/openMVG_main_openMVG2PMVS -i software/SfM/ImageDataset_SceauxCastle/Castle_Incremental_Reconstruction/reconstruction_sequential/sfm_data.bin -o software/SfM/ImageDataset_SceauxCastle/Castle_Incremental_Reconstruction/

cd ~/Application/openMVG/build/software/SfM/ImageDataset_SceauxCastle/Castle_Incremental_Reconstruction/PMVS

# PMVS
/home/tong/Application/CMVS-PMVS/program/build/main/pmvs2 ./ pmvs_options.txt

PMVS 1

This paper proposes a novel algorithm for multiview stereopsis that outputs a dense set of small rectangular patches covering the surfaces visible in the images. Stereopsis is implemented as a match, expand, and filter procedure, starting from a sparse set of matched keypoints, and repeatedly expanding these before using visibility constraints to filter away false matches. The keys to the performance of the proposed algorithm are effective techniques for enforcing local photometric consistency and global visibility constraints. Simple but effective methods are also proposed to turn the resulting patch model into a mesh which can be further refined by an algorithm that enforces both photometric consistency and regularization constraints. The proposed approach automatically detects and discards outliers and obstacles and does not require any initialization in the form of a visual hull, a bounding box, or valid depth ranges. We have tested our algorithm on various data sets including objects with fine surface details, deep concavities, and thin structures, outdoor scenes observed from a restricted set of viewpoints, and “crowded” scenes where moving obstacles appear in front of a static structure of interest. A quantitative evaluation on the Middlebury benchmark 2 shows that the proposed method outperforms all others submitted so far for four out of the six data sets.

CMVS 3

Literature

  1. Yasutaka Furukawa and Jean Ponce. Accurate, dense, and robust multi-view stereopsis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 32(8):1362–1376, August 2010. 

  2. S. M. Seitz, B. Curless, J. Diebel, D. Scharstein, and R. Szeliski, “Multiview stereo evaluation.” [Online]. Available: http://vision.middlebury.edu/mview/ 

  3. Yasutaka Furukawa, Brian Curless, Steven M. Seitz, and Richard Szeliski. Towards Internet-scale multiview stereo. In IEEE Conference on Computer Vision and Pattern Recognition, 2010.