We are releasing the code and data corresponding to our ArXiv pre-print on weakly-supervised 3D shape completion — a follow-up work on our earlier CVPR’18 paper. The article provides links to the GitHub repositories and data downloads as well as detailed descriptions. It also highlights the differences between the two papers.
In this follow-up on our CVPR’18 work, we extend our weakly-supervised 3D shape completion approach to obtain high-quality shape predictions, and also present updated, synthetic benchmarks on ShapeNet and ModelNet. The paper is now available as pre-print on ArXiv. Abstract, some experimental results and a comparison to our CVPR’18 work can be found in this article.
Finally, we are able to release the code and the data corresponding to our CVPR’18 paper on “Learning 3D Shape Completion from Laser Scan Data with Weak Supervision”. In this article, I want to briefly outline the released code and data.
Currently, both OpenCV 2 and OpenCV 3 seem to have some minor issues with CUDA 9. However, CUDA 9 is required for the latest generation of NVidia graphics cards. In this article, based on this StackOverflow question, I want to discuss a very simple patch to get OpenCV 2 running with CUDA 9.
In this CVPR’18 paper, based on my master thesis, we propose a weakly-supervised and learning-based approach to 3D shape completion of sparse and noisy point clouds. We show that, using a learned shape prior, shape completion can be learned without access to ground truth shapes — only by knowing the object category at hand. This article provides the paper and its supplementary material.
My master thesis, written at the Autonomous Vision Group of Max Planck Institute for Intelligent Systems under the supervision of Prof. Andreas Geiger, addresses the problem of 3D shape completion of sparse point clouds under weak supervision. Specifically, based on a learned shape prior it is possible to learn 3D shape completion without access to ground truth shapes, as shown on KITTI. This article briefly introduces the problem and the main contributions and offers the thesis as download.
As part of the online course Creative Applications of Deep Learning with TensorFlow, and to get started with TensorFlow, I implemented some experiments on MNIST. Specifically, I tested different architectures, activation functions and initialization schemes. While these experiments are not systematic enough for reliable results, they can be useful as an introduction to TensorFlow. In this article, I want to share the code and the corresponding presentation.
In the last few months, I started to pursue a PhD. Although I did not have many options, also because I decided not to apply to many programs, I found choosing the right PhD incredible difficult. In this article, I want to share some of my insights.