From 04:00 PM CDT – 08:00 PM CDT (09:00 PM UTC – 01:00 AM UTC) Tuesday, April 16, ni.com will undergo system upgrades that may result in temporary service interruption.

We appreciate your patience as we improve our online experience.

LabVIEW

cancel
Showing results for 
Search instead for 
Did you mean: 

Pixel coordinates to “real world” coordinates using depth (Kinect)

Hello everyone,

 

I've been try to code a Global Localization problem using a Kinect and simulating a Laser Range Finder. However, I cannot perform a good Scan Match between what the Kinect sees and a Blueprint of an office if I don't make a translation from Pixel coordinates to Real World coordinates.

When I try to find a solution, most of what I found is for C# users and they simply use "Skeletal Map" wich gives you the real coordinates of the pixels in real world. Using trigonometry you could find that give a FOV theta for the Kinect and a given depth values of the object you are trying to measure. Then, the with of the frame you got would be            

W = tan(theta / 2) * h * 2

 

Where:

  • W = Field of view Width
  • theta = Horizontal field of view Angle (60 degrees)
  • h = Depth Value

Sounds plausible though, I would like to know if anyone has dealt with this problem before and got any solution implemented in Labview because I don't like to reinvent the wheel.

 

Thanks in advance.

0 Kudos
Message 1 of 1
(3,040 Views)