Please, note that the main part of the code has been released, though we are still testing it to fix possible glitches. Thank you.
Python and C++ code for realistic 3D face modeling from single image using our shape and detail regression networks
This page contains end-to-end demo code that estimates the 3D facial shape with realistic details directly from an unconstrained 2D face image. For a given input image, it produces standard ply files of the 3D face shape. It accompanies the deep networks described in our paper [1] and [2]. The occlusion recovery code, however, will be published in a future release. We also include demo code and data presented in [1].
- Dlib Python and C++ library
- OpenCV Python and C++ library
- Caffe (version 1.0.0-rc3 or above required)
- Numpy
- Python2.7
- PyTorch
The code has been tested on Linux only. On Linux you can rely on the default version of python, installing all the packages needed from the package manager or on Anaconda Python and install required packages through conda
. A bit more effort is required to install caffé, dlib, and libhdf5.
Before running the code, please, make sure to have all the required data in the following specific folder:
- Download our Bump-CNN and move the CNN model (1 file:
ckpt_109_grad.pth.tar
) into theCNN
folder - Download our CNN and move the CNN model (3 files:
3dmm_cnn_resnet_101.caffemodel
,deploy_network.prototxt
,mean.binaryproto
) into theCNN
folder - Download the Basel Face Model and move
01_MorphableModel.mat
into the3DMM_model
folder - Acquire 3DDFA Expression Model, run its code to generate
Model_Expression.mat
and move this file the3DMM_model
folder - Go into
3DMM_model
folder. Run the scriptpython trimBaselFace.py
. This should output 2 filesBaselFaceModel_mod.mat
andBaselFaceModel_mod.h5
. - Download dlib face prediction model and move the
.dat
file into thedlib_model
folder.
Note that we modified the model files from the 3DMM-CNN paper. Therefore, if you generated these files before, you need to re-create them for this code.
- Install cmake:
apt-get install cmake
- Install opencv (2.4.6 or higher is recommended):
(http://docs.opencv.org/doc/tutorials/introduction/linux_install/linux_install.html)
- Install libboost (1.5 or higher is recommended):
apt-get install libboost-all-dev
- Install OpenGL, freeglut, and glew
sudo apt-get install freeglut3-dev
sudo apt-get install libglew-dev
- Install libhdf5-dev library
sudo apt-get install libhdf5-dev
- Install Dlib C++ library
(http://dlib.net/)
- Update Dlib directory paths (
DLIB_INCLUDE_DIR
andDLIB_LIB_DIR
) inCMakeLists.txt
- Make build directory (temporary). Make & install to bin folder
mkdir build
cd build
cmake -D CMAKE_BUILD_TYPE=RELEASE -D CMAKE_INSTALL_PREFIX=../bin ..
make
make install
This code should generate TestBump
in bin
folder
- Go into
demoCode
folder. The demo script can be used from the command line with the following syntax:
$ Usage: python testBatchModel.py <inputList> <outputDir>
where the parameters are the following:
<inputList>
is a text file containing the paths to each of the input images, one in each line.<outputDir>
is the path to the output directory, where ply files are stored.
An example for <inputList>
is demoCode/testImages.txt
../data/test/03f245cb652c103e1928b1b27028fadd--smith-glasses-too-faced.jpg ../data/test/20140420_011855_News1-Apr-25.jpg ....
The output 3D models will be <outputDir>/<imageName>_<postfix>.ply
with <postfix>
= <modelType>_<poseType>
. <modelType>
can be "foundation"
, "withBump"
(before soft-symmetry),"sparseFull"
(soft-symmetry on the sparse mesh), and "final"
. <poseType>
can be "frontal"
or "aligned"
(based on the estimated pose).
The final 3D shape has <postfix>
as "final_frontal"
.
The PLY files can be displayed using standard off-the-shelf 3D (ply file) visualization software such as MeshLab.
Note that our occlusion recovery code is not included in this release.
- Go into
demoCode
folder. The demo script can be used from the command line with the following syntax:
$ Usage: ./testPaperResults.sh
If you find this work useful, please cite our paper [1] with the following bibtex:
@inproceedings{tran2017extreme
title={Extreme 3D Face Reconstruction: Looking Past Occlusions},
author={Tran, Anh Tuan and Hassner, Tal and Masi, Iacopo and Paz, Eran and Nirkin, Yuval and Medioni, G\'{e}rard},
booktitle={arxiv preprint},
year={2017}
}
[1] A. Tran, T. Hassner, I. Masi, E. Paz, Y. Nirkin, G. Medioni, "Extreme 3D Face Reconstruction: Looking Past Occlusions", arxiv pre-print 2017
[2] A. Tran, T. Hassner, I. Masi, G. Medioni, "Regressing Robust and Discriminative 3D Morphable Models with a very Deep Neural Network", CVPR 2017
- Dec. 2017, First Release
The SOFTWARE PACKAGE provided in this page is provided "as is", without any guarantee made as to its suitability or fitness for any particular use. It may contain bugs, so use of this tool is at your own risk. We take no responsibility for any damage of any sort that may unintentionally be caused through its use.
If you have any questions, drop an email to [email protected] , [email protected] and [email protected] or leave a message below with GitHub (log-in is needed).