MACVO_Office.mp4
Docker Image See /Docker/DockerfileRoot
to build the container.
Virtual Environment, Need Python 3.10+. See requirements.txt
for environment requirements. Only tested on PyTorch >=2.0
and NumPy < 2.0.0
. Breaking Change: must have PyPose >= 0.6.8
- pretrained model for MACVO, stereo TartanVO and DPVO is in our release
-
Download a demo sequence through drive and the pre-trained model for front-end model and posenet model
-
Remember to change the path in the config file
-
Run MAC-VO on a single sequence
$ python MACVO.py --odom [PATH_TO_ODOM_CONFIG] --data [PATH_TO_DATA_CONFIG] --useRR
see
python MACVO.py --help
for the more flags and configurations.Example Usages (See More)
-
Run MAC-VO (Ours method):
$ python MACVO.py --odom ./Config/Experiment/MACVO/MACVO.yaml --data ./Config/Sequence/TartanAir_abandonfac_001.yaml
-
Run MAC-VO for ablation studies
$ python MACVO.py --odom ./Config/Experiment/MACVO/Ablation_Study/[CHOOSE_ONE_CFG].yaml --data ./Config/Sequence/TartanAir_abandonfac_001.yaml --useRR
-
-
Run MAC-VO on Test Dataset
$ python -m Scripts.Experiment.Experiment_MACVO --odom [PATH_TO_ODOM_CONFIG]
-
Run MAC-VO Mapping Mode
Mapping mode only reproject pixels to 3D space and does not optimize the pose. To run the mapping mode, you need to first run a trajectory through the original mode (MAC-VO), and pass the resulted pose file to MAC-VO mapping mode by modifying the config. (Specifically,
motion > args > pose_file
in config file)$ python MACVO.py --odom ./Config/Experiment/MACVO/MACVO_MappingMode.yaml --data ./Config/Sequence/TartanAir_abandonfac_001.yaml
Every run will produce a Sandbox
(or Space
). A Sandbox
is a storage unit that contains all the results and meta-information of an experiment. The evaluation and plotting script usually requires one or more paths of sandbox(es).
-
Evaluate Sequence(s)
Calculate the absolute translate error (ATE, m); relative translation error (RTE, m/frame); relative orientation error (ROE, deg/frame); relative pose error (per frame on se(3)).
$ python -m Evaluation.EvalSeq --spaces SPACE_0, [SPACE, ...]
-
Plot Sequence(s)
Plot sequences, translation, translation error, rotation and rotation error.
$ python -m Evaluation.PlotSeq --spaces SPACE_0, [SPACE, ...]
See more commands for plotting figures / creating demo.
We used the Rerun visualizer to visualize 3D space including camera pose, point cloud and trajectory.
-
Create Rerun Recording for Run(s)
$ python -m Scripts.AdHoc.DemoCompare --macvo_space [MACVO_RESULT_PATH] --other_spaces [RESULT_PATH, ...] --other_types [{DROID-SLAM, DPVO, TartanVO}, ...]
-
Create Rerun Visualization for Map
Create a
tensor_map_vis.rrd
file in each sandbox that stores the visualization of 3D point cloud map.$ python -m Scripts.AdHoc.DemoCompare --spaces [RESULT_PATH, ...] --recursive?
-
Create Rerun Visualization for a Single Run (Eye-catcher figure for our paper)
$ python -m Scripts.AdHoc.DemoSequence --space [RESULT_PATH] --data [DATA_CONFIG_PATH]
We also integrated two baseline methods (DPVO, TartanVO Stereo) into the codebase for evaluation, visualization and comparison.
Expand All (4 commands)
-
Run DPVO on a single sequence
$ python DPVO.py --odom ./Config/Experiment/Baseline/DPVO/DPVO.yaml --data [PATH_TO_DATA_CONFIG]
-
Run DPVO on Test Dataset
$ python -m Scripts.Experiment.Experiment_DPVO --odom ./Config/Experiment/Baseline/DPVO/DPVO.yaml
-
Run TartanVO (Stereo) on a single sequence
$ python TartanVO.py --odom ./Config/Experiment/Baseline/TartanVO/TartanVOStereo.yaml --data [PATH_TO_DATA_CONFIG]
-
Run TartanVO (Stereo) on Test Dataset
$ python -m Scripts.Experiment.Experiment_TartanVO --odom ./Config/Experiment/Baseline/TartanVO/TartanVOStereo.yaml
-
PyTorch Tensor Data - All images are stored in
BxCxHxW
format following the convention. Batch dimension is always the first dimension of tensor. -
Pixels on Camera Plane - All pixel coordinates are stored in
uv
format following the OpenCV convention, where the direction of uv are "east-down". Note that this requires us to access PyTorch tensor indata[..., v, u]
indexing.(0, 0)----> u | | v
-
World Coordinate -
NED
convention,+x -> North
,+y -> East
,+z -> Down
with the first frame being world origin having identity SE3 pose.
This codebase is designed with modularization in mind so it's easy to modify, replace, and re-configure modules of MAC-VO. One can easily use or replase the provided modules like flow estimator, depth estimator, keypoint selector, etc. to create a new visual odometry.
For details on how to customize your own datasets, please see our custom dataset documentation.
For details on reconfiguring the MAC-VO or further development, please see our configuration documentation.