The math for converting the mismatch (as a percentage of pixels or the width of the image) to the actual distance is pretty well documented (and not very complicated), but I will also write it down here.
The following is an example of a mismatch image (in pixels) and an input image width of 2K (2048 pixels across) the image:
The convergence distance is determined by the rotation between the camera lenses. In this example, it will be 5 meters. A convergence distance of 5 (meters) means that the mismatch of objects at 5 meters is 0.
CD = 5 (meters)
Inverse Convergence Distance: 1 / CD
IZ = 1/5 = 0.2M
Camera sensor size in meters
SS = 0.035 (meters) //35mm camera sensor
The width of the pixel on the sensor in meters
PW = SS/image resolution = 0.035 / 2048(image width) = 0.00001708984
The focal length of your cameras in meters
FL = 0.07 //70mm lens
InterAxial distance: distance from the center of the left lens to the center of the right lens
IA = 0.0025 //2.5mm
A combination of the physical settings of your camera
A = FL * IA / PW
Adjusted with the camera: (for left viewing, only in the right view will use positive [mismatch value])
AD = 2 * (-[disparity value] / A)
From here you can calculate the actual distance using the following equation:
realDistance = 1 / (IZ β AD)
This equation only works for βtoeβ camera systems, parallel cameras will use a slightly different equation to avoid infinity values, but for now I will leave it to that. If you need parallel material, just let me know.