This dataset mainly challenges the monocular video-based 3D reconstruction methods, such as LASR. The dataset includes rigid objects from ShapeNet and non-rigid human animations built from quaternius.
Rendered datasets available at gs://kubric-public/data/video_based_reconstruction
To generate a car dataset, run the following:
docker run --rm --interactive \
--user $(id -u):$(id -g) \
--volume "$(pwd):/kubric" \
kubricdockerhub/kubruntu \
/usr/bin/python3 challenges/video_based_reconstruction/worker.py \
--object=car \
Script parameters include:
rotate_camera: bool, whether to rotate camera during simulation. If enabled, camera will rotate vertically around world center.camera_rot_range: radius angle for which the camera will rotate.object: one of [cube, torus, car, airplane, chair, table, pillow]extra_obj_texture: bool, whether to apply external texture on the objectobj_texture_path: path to the external textureno_texture: bool, whether to remove all texture, including original texture for ShapeNet objects
The script is configured to directly output data in format of LASR input. A folder with name object is created in output directory.
<object>/FlowBW,<object>/FlowFW: backward and forward optical flow images<object>/LASR/Annotations/Full-Resolution/(r)<object>: object masks<object>/LASR/Camera/Full-Resolution/(r)<object>: camera extrinsics in LASR's preferred format. Line 1: focal, line 2-3: x,y translation, line 4-7: WXYZ quaternion, line 8: z translation (depth).<object>/LASR/JPEGImages/Full-Resolution/(r)<object>: object images


