Combining Two Kinects To Make Better 3D Video 106
suraj.sun sends this quote from Engadget about improving the Kinect 3D video recordings we discussed recently:
"[Oliver Kreylos is] blowing minds and demonstrating that two Kinects can be paired and their output meshed — one basically filling in the gaps of the other. He found that the two do create some interference, the dotted IR pattern of one causing some holes and blotches in the other, but when the two are combined they basically help each other out and the results are quite impressive."
Well taht is (Score:2, Funny)
Two eyes are better than one (Score:1)
Re:Two eyes are better than one (Score:5, Funny)
What feat would that be that one stationary ear could do as well as kinect?
Recognize your voice from the kitchen
Re: (Score:1, Interesting)
Ear's can't do that, that's the brain. And for the kinect that would be the program in the device it connects to.
Re: (Score:1, Funny)
Because that's not a plural. What do you think I am, an idiot?
Re: (Score:3, Funny)
Re: (Score:2, Insightful)
Well, your wrong. "Connects" is a verb, and everybody knows that even plural verbs do not get apostrophe's. Sheesh man, do some research.
Re: (Score:1, Insightful)
Thi's is a brilliant thread.
Re: (Score:2)
Re: (Score:3, Interesting)
I wonder what the class of "optical illusions" for the Kinect's vision system and algorithms is... Off the top of my head, I'd imagine that retroreflective materials might kind of freak it out; but I'd be curious to know if there are any stimuli that cause it to wig out in weird ways, the way that optical illusions do the human visual
Re: (Score:2)
I wonder what the class of "optical illusions" for the Kinect's vision system and algorithms is
I'm guessing kinect makes assumptions based on common human bone structure, e.g. something like a dog might freak it out and make it explode.
Re: (Score:2)
Re: (Score:3, Interesting)
Re: (Score:1)
CMOS sensors light gathering capabilities fall off over increasing wavelength.
Silicon's quantum efficiency at NIR is much lower than visible. There's not a
huge range of NIR to play in without QE falling off.
IR diodes don't emit light over a single wavelength. Not only do they shift long with
temperature, but the rated wavelength is really an average of the range the wavelength
drifts over.
Very tight bandpass filters tend to drift shorter in wavelength off axis.
Re: (Score:2)
Re: (Score:1)
This could be solved by replacing the IR source in the 'projector' (actually a point source and a pinhole grid) with one of a different wavelength, and adding appropriate filters to the IR cameras in each Kinect.
Or maybe timing the grid light to be off while the other camera is on and vis-versa. Alternating back and forth quickly like the 3D LCD 'shutter' lenses. This way the grids would not interfere with each other. Don't have to turn off the cameras, just don't use the 3D grid data from those frames where the opposite grid is being used.
Just my .01 worth. :)
Re: (Score:2)
If it takes 30-60 frames to do so, that is only a 1-2 second delay, which is nearly irrelevant from the perspective of the standard use case. Just have the menu do some slightly sci-fi transition during that time and they will never even notice.
If, however, you are trying to use two or more Kinects with sh
Re: (Score:1)
We don't know how to use jelly yet, so we settle with plastic and metal.
Still a crazy task.
Re:Two eyes are better than one (Score:5, Insightful)
The "good ol' brain" does a fairly crappy job, actually. 3D vision systems like these tend to perform quite a bit better than we do. And we only do as well as we do because we can use a lot of indirect clues based on our long experience with a 3D-world - we know how big stuff normally is, for instance, so we can judge distance from size. Mess up those clues and we completely lose it.
And even with good clues we don't actually measure distance well. Have somebody place items on a parking lot or some place like that, then try to guess the distances. Not going to be very accurate. Try to estimate distance vertically rather than horizontally and you'll do even worse; you have fewer clues and less experience to fall back on.
Re: (Score:2)
- A car is approx 3m long 2m wide. A parking space is about same
- The lanes between spaces are 2 cars wide, to allow for idiots who can't follow the arrows.
- Basic trig can give you any distance in a parking lot.
The same applies to buildings. The average person is 6' tall, with 18" spare to the roof. The floor space is approx 6", making each floor approx 7'. Multiply $fl
Re: (Score:2)
Re: (Score:2)
You can get training if you really care.
Re: (Score:2)
Plain buildings a la MiniPeace. however, would throw me completely.
Re: (Score:1)
Yep, just look at the quarterback for the Carolina Panthers.
Re: (Score:2)
Re: (Score:2)
That's typically a product of training. We don't have much experience with it, because we don't need it.
But take an aborigine, and ask him to estimate how far something is, and you'll get a good accurate answer, even if it's not in feet and inches.
Re: (Score:2)
And even with good clues we don't actually measure distance well. Have somebody place items on a parking lot or some place like that, then try to guess the distances. Not going to be very accurate.
And yet we are able to navigate and interact with our environment with a high degree of precision. When I'm driving a car, for instance, without looking at how fast I'm going, knowing distances, the weight of the car, my acceleration and deceleration capabilities, I'm able to stop at a line painted on the road to within half a meter. Just with my eyes!
I work with robots, and even knowing all this information to a high accuracy, there is so much work that needs to be done with localization, navigation, plann
Re: (Score:2)
Yes, we are. Our vision system is pretty successful when you look at how we actually use it in the real world. We don't actually need to know the precise distance to things; what we want to know is rather direction and time to impact and similar and we're really, really good at that (look up tau-margin estimation for instance). Though note that with a human-level vision system you would still need a lot of thos
Re: (Score:2)
But I wrote this in reply to a poster that seemed to believe we humans are actually better than Kinect at the specific vision tasks it's built to do.
But we are better. Kinect is built to recognize faces and body postures, it’s not built to estimate the distance from you to the TV even if it can do that more accurately than we can.
Re: (Score:2)
But we are better. Kinect is built to recognize faces and body postures, it’s not built to estimate the distance from you to the TV even if it can do that more accurately than we can.
That is a ridiculous statement. Kinect builds heightmaps. If that's not estimating the distance from you to the TV then I don't know what is. Kinect in fact does the other cool things it can do specifically because it is built to estimate the distance from you to the TV, when other camera systems are not. If this was ALL it would do you could still do the same stuff on the 360 in software, but it would take away from the available processing power which is why embedding it as a complete solution was the sma
Re: (Score:2)
Re: (Score:2)
The Kinect wasn't created/built for the purpose of measuring objects, even if it's better at this than us humans.
That's a big fail of a response. The statement that prompted your original comment was "But I wrote this in reply to a poster that seemed to believe we humans are actually better than Kinect at the specific vision tasks it's built to do." and you said "But we are better. Kinect is built to recognize faces and body postures, it’s not built to estimate the distance from you to the TV even if it can do that more accurately than we can." But that is plainly false. Kinect is built to measure the distance f
Re: (Score:2)
I simply disagree with everything you said
So you believe that Kinect is superior to humans at recognizing faces and body postures?
Re: (Score:1)
Re: (Score:1)
this-is-what-happens-when-yo (Score:2)
Purpose of Headings (Score:3, Insightful)
Headings are for brief topic summaries (a few words.) Not content.
Anybody in optics? (Score:5, Interesting)
Is that A)Wholly impractical, because of some sort of effect the reflecting materials would have on the IR wavelengths, B)Sure, it's possible; but have you checked the supplier's price list for narrowband IR filters recently, or C)Just a bit of ebay and some steady hands?
Perhaps more practically, I wonder if the Kinects could(with some mixture of hardware shutters and firmware or driver mods) be made to trade off sample rate for coverage(ie. if the kinects are ordinarily taking 60 frames/second, could two kinects be made to take 30 frames/second each, turning off their IR source when it isn't their turn, and turning it on when it is) or does their mechanism of operation require too much time to calibrate itself on startup?
Re:Anybody in optics? (Score:4, Informative)
Re: (Score:3, Informative)
It is definitely possible to use some narrow bandpass filters. In the infrared region there are various filters for available that have a wavelength window of 10 nm at 1000 nm. These filters are not available at Walmart, but they are not too costly either. Depending on size, quality, wavelength and other parameters you should be able to buy some for $50 (Thorlabs).
To actually hack the Kinect you have to test, whether there are other infrared filters used and if the camera is sensitive enough at different wa
Re: (Score:2)
Wouldn't polarized filters do the trick?
Re: (Score:2)
Wouldn't polarized filters do the trick?
As someone in another thread points out, polarity is lost when light is scattered as it reflects (3D cinemas have special screens).
Also, polarizing gives you two channels. Bandwidth selection gives you many.
Re: (Score:1)
Not really as the surface absorbing the light has preserve the polarisation - and anyone who's setup a dual-projector 3D rig with polarised light can attest - you need a special surface coating to get good preservation of polarisation.
Paint with silver particles in it is typically used for painting 3D screens, for example.
Re: (Score:1)
Shuttering might work, but as you said, you'd reduce the overall framerate, meaning worse motion capture. Also you'd need to synchronize the shutters somehow, and that'd be a pain.
Filtering would change the sensitivity of the camera, but it won't do much to
Re: (Score:1)
I.e. (Normal/Today's world):
Diode 1 emits light across the entire IR-A spectrum (700 - 1400nm)
Camera 1 detects light across entire IR-A spectrum (700 - 1400nm)
Diode 2 emits light across the entire IR-A spectrum (700 - 1400nm)
Camera 2 detects light across entire IR-A spectrum (700 - 1400nm)
Apply filters to both emitter and detector, on both Kinect 1 and 2:
Diode 1 emits light across the entire IR-A spectrum (700 - 1400nm), filter is applied so
Re: (Score:1)
So wont 3 Kinects make 3D video? (Score:1)
Re: (Score:1)
Re: (Score:2)
I'd expect that 3 Kinects would make 4D video.
Re: (Score:2)
What about four Kinects? Oh man, 5D... they'd like create a new dimension!!!
Re: (Score:2)
This is the dawning of the Age of Aquarius!
Re: (Score:1)
Re: (Score:2)
So wont 3 Kinects make 3D video?
I get what you're saying. With 3 of these you should be able to get x,y,z coordinates. However, each of these is capable of getting the x,y,z for surfaces facing the camera, the problem is you need to hit all the surfaces. With 6 Kinects to cover front, back, left, right, top and bottom you could probably have the best coverage, but I expect four of them, one in each corner of the room like security cameras, would provide similar results.
This is immense... (Score:2)
This is revolutionary for entertainment. Not stereoscopy.
Re: (Score:3, Interesting)
This is actual 3D on the screen, like a 3D game. You can't zoom in, or even focus, on the background in Avatar. In fact, attempting it gave me a massive headache. With this true 3D rendering of an object, you can zoom, focus, and more importantly pan around objects in the scene, in real time. That is the breakthrough this hack has brought about.
Re: (Score:3, Informative)
With this true 3D rendering of an object, you can zoom, focus, and more importantly pan around objects in the scene, in real time.
Er, if neither of the Kinect cameras is focused on the background, then it's going to be blurry no matter what.
Assuming we're talking about a recording, you'd be able to move the virtual camera, but you wouldn't be able to bring things into focus that were not in focus in the recording.
What this gives you is a 3D model, with an many textures mapped onto it as there are cameras.
Re: (Score:2)
Re: (Score:2)
You miss on how much more expensive the non-CG movies would have to be to allow this. In many sets, if you'd move the camera just a bit outside of what it views, you'd see all of the production equipment, other people, etc. In classical 2D and stereoscopic (market-speak 3D) filming on a set, you only build enough of an expensive set to let you film what's in the screenplay. Anything more is a waste.
Same goes for 3D CG movies: no point in making the character and scene models any more detailed/extensive than
Re: (Score:2)
I thought that the Kinect camera was fixed-focus with infinite DoF? In that case, you can emulate focus by blurring everything not in focus in the final scene. Of course, everything will be equally out of focus...
Re: (Score:2)
Er, if neither of the Kinect cameras is focused on the background, then it's going to be blurry no matter what.
Unless you want to do extreme close ups, focus isn't much of an issue, as the depth-of-field of any webcam and things like Kinect is rather large. The blurring of the background you have in cinema, doesn't happen by accident, but by design, if you just point your regular webcam at the scene you wouldn't get that. The bigger problem would be resolution, as anything further away would naturally have an ever lower resolution then stuff in the foreground. So you couldn't freely float around a room without thing
Re: (Score:2)
Unless you want to do extreme close ups, focus isn't much of an issue, as the depth-of-field of any webcam and things like Kinect is rather large.
I'll take your word for it.
You might expect shallow depth-of-field in low light conditions, since one easy way to get more light is to open the aperture wider.
Re: (Score:2)
Only for a very broad kind of "similar".
The live action parts of Avatar would have been filmed using traditional stereoscopic techniques; two cameras imitating two eyes.
The CG elements would have been traditional CG; models created by a combination of artists and 3D scans.
The CG animation would have been motion capture as used by Peter Jackson and countless video games: multiple cameras tracking reflective points attached to an actor's body.
That's nothing... (Score:2, Funny)
Can you imagine a beowulf cluster of kinects??
Re: (Score:2)
or Natalie Portman with hot kinects?
Re: (Score:3, Funny)
He's not the only one. My depth-first recursive post counter has found hundreds of such posts.
Home Survivelance (Score:1)
Re: (Score:2)
Re: (Score:1)
Would be cool to insert avatars from several people there though. A 3d video version of the old MOO concept, or a local second life with a live video background to use a more modern analogy.
Bandwidth will be an issue however.
X-Ray machines (Score:2)
With all this stuff in the news recently about backscatter machines and the need for improved x-ray machines, this sort of system would be fantastic for improving the quality of screening, being able to look in and see depth in luggage.
2 Kinects, 1 Box... (Score:5, Funny)
... is good, but I'm holding out for 4 Girls, 3 Kinects, 2 Boxes, 1 Cup :)
Re: (Score:2, Funny)
...and a partridge in a pear tree?
Re: (Score:3, Funny)
... is good, but I'm holding out for 4 Girls, 3 Kinects, 2 Boxes, 1 Cup :)
Correct me if I'm wrong, but if there's 4 girls then wouldn't there also be 4 boxes?
Just sayin'...
Re: (Score:2)
Re: (Score:2)
Polarizing filters! (Score:4, Insightful)
When I first saw the video of one Kinect, I immediately wondered how you could get multiple units working together.
It wasn't until I watched the video again later that day that it hit me. I had just explained to someone how 3D theater projection works, and so I had an epiphany: The most sensible course is to use polarizing filters.
With filters on the IR emitters and cameras, the units should be able to only see their own IR illumination. Of course, it would only work for two Kinects with maximum effectiveness, but considering how well this turned out with the units at right-angles from each other, I don't see why you couldn't combine the two ideas for 3-4 units and get sufficient quality.
I wish I had the money to get a couple Kinects and test my idea, but I'm no good with coding anyway.
It'd be awesome to see the Blender Foundation put out a bounty for a Kinect-based open source motion capture and 3D scanning suite though. :D
Re:Polarizing filters! (Score:4, Informative)
Unfortunately, this wouldn't work very well. Light tends to lose its polarization somewhat when it bounces off of things. In a theater that's OK because you can use a special screen that maintains the polarization. Band limiting each kinect would be more effective than polarization (and would also scale better - polarization only allows for 2 kinects; the bandpass idea would only be limited by how good your filters are).
Re: (Score:2)
Light is re-polarized when bouncing off of things, that's why people wear polarized sunglasses; it eliminates glare.
Unfortunately, you wouldn't be able to predict the resulting polarization with great confidence off of curved surfaces at strange angles like bodies have.
Re: (Score:2)
I'd do it in a different way that may well be lower cost and more scalable than any wavelength- or polarization-based selectivity.
1. Run the Kinects off a common reference frequency. The onboard circuitry probably uses one crystal oscillator and PLL-controlled VCOs to generate various derivative frequencies to time everything. A common reference will keep all Kinects phase-synchronized, while the phase itself may well be random.
2. Figure out how to discover the phase angle when the IR camera shutter is open
3D Scanner (Score:2)
Oblig. (Score:1)
In other words... (Score:1)
2. Subscribe
3. Share new videos on Slashdot
4. ????
5. PROFIT!
Basic Webcam (Score:2, Informative)
Be great for homemade pr0n (Score:1)
Re: (Score:1)
this+HUD (Score:2)
Given a quality enough image, bandwidth, and some motion-sensing gear (ahem), any immersion-style display (HUD, dome, etc) could allow for real-time panning of a distant location.
Examples:
- shooting a net of these at an operating table would let remote viewers move around the room and view the procedure without crowding the room or limited to the perspective of the single camera.
- a web site could point this setup at anything interesting (lab experiment, box of
Wow... it's just like a camera... (Score:1)