DSpider :
"Light field technology doesn’t really record video. It records real-world light ray trajectories and then calculates what the scene would look like based on your viewpoint. Lytro then combines that data with live-action video and with 3D renderings to reproduce true to life imagery."
So... it's not going to be an actual 1:1 representation of what's in front of you.
Reminds me of an older Will Smith movie called
"Enemy of the State":
https://www.youtube.com/watch?v=3EwZQddc3kY That's NOT how cameras work, lol. You can't "rotate" if nothing was there (like
another camera) to capture it. It even says "it can hypothesize, Chris". Yeah, right...
I wouldn't be so quick to dismiss it. From what I gather from the article this is pretty hardcore, or the start of something ~very~ hardcore... I really thought this was years away, not 2016.
Basically what we have in the next several years will be an interplay between light-field (eg. Lytro), laser-scanning (eg. Euclideon) and image-scanning (see Blendswap for examples) to recreate an entire 3D capture of a scene.
If it is just one camera then light-field interpolation will be difficult and you'd generally have a fixed perspective based on where the camera is. "Enemy Of The State" and "Deja Vu" will not quite be possible (unless Lytro has some big tricks up their sleeve and can actually "backtrace" light bounces to such an incredible degree).
Regardless, imagine setting up 3 - 4 cameras and capturing a scene. That's essentially full VR cinema right there. Given the right software, that means you can have full 3D x-y-z positioning ie. viewing the video capture from any angle.
Also, since that creates an entire 3D capture of whatever is being filmed (let's assume they start with live action movies) you could go to a cinema, strap on a StarVR or something like that, then all you'd have to do is stream the movie and the user can choose whatever angle to view it from (it might be limited to 3-5 angles at the start for practical and aesthetic purposes)... applications for home cinema would also follow accordingly.
Over time I imagine live 3D video/light-field capture can be integrated with offline or realtime rendered 3D.
Wow. I thought 2K each eye was impressive for VR already (with StarVR) but all this "additional" stuff happening. Just didn't expect it.