MRCAL(1) | mrcal: camera projection, calibration toolkit | MRCAL(1) |
mrcal-stereo - Stereo processing
$ mrcal-stereo \ --az-fov-deg 90 \ --el-fov-deg 90 \ --sgbm-block-size 5 \ --sgbm-p1 600 \ --sgbm-p2 2400 \ --sgbm-uniqueness-ratio 5 \ --sgbm-disp12-max-diff 1 \ --sgbm-speckle-window-size 200 \ --sgbm-speckle-range 2 \ --outdir /tmp \ left.cameramodel right.cameramodel \ left.jpg right.jpg Processing left.jpg and right.jpg Wrote '/tmp/rectified0.cameramodel' Wrote '/tmp/rectified1.cameramodel' Wrote '/tmp/left-rectified.png' Wrote '/tmp/right-rectified.png' Wrote '/tmp/left-disparity.png' Wrote '/tmp/left-range.png' Wrote '/tmp/points-cam0.vnl'
Given a pair of calibrated cameras and pairs of images captured by these cameras, this tool runs the whole stereo processing sequence to produce disparity and range images and a point cloud array.
mrcal functions are used to construct the rectified system. Currently only the OpenCV SGBM routine is available to perform stereo matching, but more options will be made available with time.
The commandline arguments to configure the SGBM matcher (--sgbm-...) map to the corresponding OpenCV APIs. Omitting an --sgbm-... argument will result in the defaults being used in the cv2.StereoSGBM_create() call. Usually the cv2.StereoSGBM_create() defaults are terrible, and produce a disparity map that isn't great. The --sgbm-... arguments in the synopsis above are a good start to get usable stereo.
The rectified system is constructed with the axes
- x: from the origin of the first camera to the origin of the
second camera (the
baseline direction)
- y: completes the system from x,z
- z: the mean "forward" direction of the two input
cameras, with the component
parallel to the baseline subtracted off
The active window in this system is specified using a few parameters. These refer to
- the "azimuth" (or "az"): the direction along the baseline: rectified x axis
- the "elevation" (or "el"): the direction across the baseline: rectified y axis
The rectified field of view is given by the arguments --az-fov-deg and --el-fov-deg. At this time there's no auto-detection logic, and these must be given. Changing these is a "zoom" operation.
To pan the stereo system, pass --az0-deg and/or --el0-deg. These specify the center of the rectified images, and are optional.
Finally, the resolution of the rectified images is given with --pixels-per-deg. This is optional, and defaults to the resolution of the first input image. If we want to scale the input resolution, pass a value <0. For instance, to generate rectified images at half the first-input-image resolution, pass --pixels-per-deg=-0.5. Note that the Python argparse has a problem with negative numbers, so "--pixels-per-deg -0.5" does not work.
The input images are specified by a pair of globs, so we can process many images with a single call. Each glob is expanded, and the filenames are sorted. The resulting lists of files are assumed to match up.
There are several modes of operation:
- No images given: we compute the rectified system only, writing
the models to
disk
- No --viz argument given: we compute the rectified system and the
disparity,
and we write all output as images on disk
- --viz geometry: we compute the rectified system, and display its
geometry as a
plot. No rectification is computed, and the images aren't used, and don't
need
to be passed in
- --viz stereo: compute the rectified system and the disparity. We
don't write
anything to disk initially, but we invoke an interactive visualization tool
to
display the results. Requires pyFLTK (homepage:
<https://pyfltk.sourceforge.io)>
and GL_image_display (homepage:
<https://github.com/dkogan/GL_image_display)>
models Camera models representing cameras used to capture the images. Both intrinsics and extrinsics are used images The image globs to use for the stereo. If omitted, we only write out the rectified models. If given, exactly two image globs must be given
-h, --help show this help message and exit --az-fov-deg AZ_FOV_DEG The field of view in the azimuth direction, in degrees. There's no auto-detection at this time, so this argument is required (unless --already-rectified) --el-fov-deg EL_FOV_DEG The field of view in the elevation direction, in degrees. There's no auto-detection at this time, so this argument is required (unless --already-rectified) --az0-deg AZ0_DEG The azimuth center of the rectified images. "0" means "the horizontal center of the rectified system is the mean forward direction of the two cameras projected to lie perpendicular to the baseline". If omitted, we align the center of the rectified system with the center of the two cameras' views --el0-deg EL0_DEG The elevation center of the rectified system. "0" means "the vertical center of the rectified system lies along the mean forward direction of the two cameras" Defaults to 0. --pixels-per-deg PIXELS_PER_DEG The resolution of the rectified images. This is either a whitespace-less, comma-separated list of two values (az,el) or a single value to be applied to both axes. If a resolution of >0 is requested, the value is used as is. If a resolution of <0 is requested, we use this as a scale factor on the resolution of the first input image. For instance, to downsample by a factor of 2, pass -0.5. By default, we use -1 for both axes: the resolution of the input image at the center of the rectified system. --rectification {LENSMODEL_PINHOLE,LENSMODEL_LATLON} The lens model to use for rectification. Currently two models are supported: LENSMODEL_LATLON (the default) and LENSMODEL_PINHOLE. Pinhole stereo works badly for wide lenses and suffers from varying angular resolution across the image. LENSMODEL_LATLON rectification uses a transverse equirectangular projection, and does not suffer from these effects. It is thus the recommended model --already-rectified If given, assume the given models and images already represent a rectified system. This will be checked, and the models will be used as-is if the checks pass --clahe If given, apply CLAHE equalization to the images prior to the stereo matching. If --already-rectified, we still apply this equalization, if requested. Requires --force-grayscale --force-grayscale If given, convert the images to grayscale prior to doing anything else with them. By default, read the images in their default format, and pass those posibly-color images to all the processing steps. Required if --clahe --viz {geometry,stereo} If given, we visualize either the rectified geometry or the stereo results. If --viz geometry: we construct the rectified stereo system, but instead of continuing with the stereo processing, we render the geometry of the stereo world; the images are ignored in this mode. If --viz stereo: we launch an interactive graphical tool to examine the rectification and stereo matching results; the Fl_Gl_Image_Widget Python library must be available --axis-scale AXIS_SCALE Used if --viz geometry. Scale for the camera axes. By default a reasonable default is chosen (see mrcal.show_geometry() for the logic) --title TITLE Used if --viz geometry. Title string for the plot --hardcopy HARDCOPY Used if --viz geometry. Write the output to disk, instead of making an interactive plot. The output filename is given in the option --terminal TERMINAL Used if --viz geometry. The gnuplotlib terminal. The default is almost always right, so most people don't need this option --set SET Used if --viz geometry. Extra 'set' directives to pass to gnuplotlib. May be given multiple times --unset UNSET Used if --viz geometry. Extra 'unset' directives to pass to gnuplotlib. May be given multiple times --force, -f By default existing files are not overwritten. Pass --force to overwrite them without complaint --outdir OUTDIR Directory to write the output into. If omitted, we user the current directory --tag TAG String to use in the output filenames. Non-specific output filenames if omitted --disparity-range DISPARITY_RANGE DISPARITY_RANGE The disparity limits to use in the search, in pixels. Two integers are expected: MIN_DISPARITY MAX_DISPARITY. Completely arbitrarily, we default to MIN_DISPARITY=0 and MAX_DISPARITY=100 --valid-intrinsics-region If given, annotate the image with its valid-intrinsics region. This will end up in the rectified images, and make it clear where successful matching shouldn't be expected --range-image-limits RANGE_IMAGE_LIMITS RANGE_IMAGE_LIMITS The nearest,furthest range to encode in the range image. Defaults to 1,1000, arbitrarily --stereo-matcher {SGBM,ELAS} The stereo-matching method. By default we use the "SGBM" method from OpenCV. libelas isn't always available, and must be enabled at compile-time by setting USE_LIBELAS=1 during the build --sgbm-block-size SGBM_BLOCK_SIZE A parameter for the OpenCV SGBM matcher. If omitted, 5 is used --sgbm-p1 SGBM_P1 A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-p2 SGBM_P2 A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-disp12-max-diff SGBM_DISP12_MAX_DIFF A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-pre-filter-cap SGBM_PRE_FILTER_CAP A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-uniqueness-ratio SGBM_UNIQUENESS_RATIO A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-speckle-window-size SGBM_SPECKLE_WINDOW_SIZE A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-speckle-range SGBM_SPECKLE_RANGE A parameter for the OpenCV SGBM matcher. If omitted, the OpenCV default is used --sgbm-mode {SGBM,HH,HH4,SGBM_3WAY} A parameter for the OpenCV SGBM matcher. Must be one of ('SGBM','HH','HH4','SGBM_3WAY'). If omitted, the OpenCV default (SGBM) is used --write-point-cloud If given, we write out the point cloud as a .ply file. Each point is reported in the reference coordinate system, colored with the nearest-neighbor color of the camera0 image. This is disabled by default because this is potentially a very large file
<https://www.github.com/dkogan/mrcal>
Dima Kogan, "<dima@secretsauce.net>"
Copyright (c) 2017-2021 California Institute of Technology ("Caltech"). U.S. Government sponsorship acknowledged. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
2023-01-30 | mrcal 2.2 |