Site icon ProVideo Coalition

Do Away with the Array: RE:Lens for After Effects brings clarity to extreme fisheye scenes

Do Away with the Array: RE:Lens for After Effects brings clarity to extreme fisheye scenes 1

RE:Vision Effects, best known for its Academy Sci-Tech award-winning motion-estimation-based effects including Twixtor and Reel Smart Motion Blur, demonstrated RE:Lens for After Effects at NAB 2016. I was their invited guest artist, demonstrating the technology in their booth, and this article goes beyond press-release coverage to give a first-hand account of what I learned working with this technology. RE:Lens goes far beyond After Effects’ built-in (and now rather ancient) Optics Compensation effect to work with accurate FOV data and covert extreme-wide-angle, or “superfish” images commonly up to 280 degrees. The result can be used to produce a single undistorted image in virtually any format—16:9 all the way to ludicrous formats like ultra-wide 9:1—and there are also tools to round-trip to and from the 2:1 equirectangular format that is the standard for virtual reality on platforms such as YouTube 360.

Even if you’ve simply encountered frustration just working with GoPro footage using Optics Compensation (whose FOV controls don’t relate to real-world numbers, requiring instead that you measure straight lines to derive an accurate setting, as I have shown you how to do), you may find these plug-ins helpful, but it is also possible to work with footage from exotic lenses that have been out of bounds for the built-in effects that ship with After Effects. The demo video below includes shots derived from the 280 degree hemispheric Entapano Entanaya Fisheye lens that is mounted on a Go Pro via a RIBCAGE setup from Back-Bone Gear Inc. The entire setup including camera and plug-ins costs less than $1500 and was promoted in the RE:Vision booth.

https://www.youtube.com/watch?v=sLX5bzzYVxY&feature=youtu.be&list=PLJZE0COAfWUXCJtLai9uKkF_nCm97ZaSB

Those are the basics, and you can stop reading here if just looking for industry news. What follows is what I learned from working on documentation, development and demos with RE:Vision about why handling of extreme lens distortion could have far-reaching benefits even for that most mundane of video productions, the talking-head round-table discussion.

Before we get into applications of these lenses, let’s take a look at camera arrays, the current standard for capture of ultra-wide images all the way to a full 360 degree view. Keep in mind that arrays are useful for more than just VR: they can be used to cover a scene that a moving camera can’t, and they provide the means for effects such as Matrix-style bullet-time effects.

Arrays are also a bit of a pain, for two main reasons. One is that they don’t provide a finished image; you have to stitch the results, which means averaging together overlapping areas of frames (which also happen to be the most lens-distorted areas of those frames, the edges). Many of the rigs are custom-built and don’t provide a means to sync color or frame capture, as can be learned the hard way with custom GoPro arrays. Short summary: arrays have to be fixed in post, sometimes expensively so.

The bigger problem with the array? In some ways it has thus far doomed virtual reality to be forever stuck in the wide shot, holding action 15 feet or further from camera. Sure, you can have the talent make sure to hit and hold a mark that is perfectly framed by one of the cameras and move in closer, but they’d better stay on-axis or risk being ripped apart in the resulting image. For an example of this, take a look at the fun Diamond brothers interview posted to this very site earlier this week. Use your mouse to turn the virtual camera around and take a look at interviewer Neil Smith standing below the rig, at this point in the video, right after the 2:00 mark—or rather, take a look at the fraction of his image that is displayed. Welcome to the future.

https://youtu.be/_4GxYt5ju7A?t=2m4s

What if you could capture a scene, including close-ups, with a single camera, without the need for stitching? What I found most fascinating among our RE:Lens demos was not the VR capture and display process, but high-definition scene capture: covering a scene with a superfish lens with the intention of pulling standard, undistorted looking HD images out of it. Specifically, there are a couple of intriguing setups:

The resulting master image in either case contains no gaps, seams, no need for stitching or managing color or frame sync. In the example dance scene that can be viewed starting at 1:51 in the product overview video above, the 180 degree image was captured with an 8k sensor, so that even the softness that is apparent with 4k images is not apparent.

The After Effects virtual camera, animated in post. Covering a scene like this makes effective use of an 8k camera that might not otherwise make sense in an HD video production.

Not deciding is a decision, and leaving camera moves to be solved in post is not the future of cinema, unless the future is pans and zooms from a static location (since moving a VR rig is problematic enough that I’m not even considering it here).

And at 4k, extreme wide shots will appear soft, as can be predicted with the simplest of math. 4k divided by 1920 rounds to 2; even dividing by 1280 only makes the field of view 3 shots wide. With a 360 degree master view, there are 9 shots to be extracted to a standard 40 degree view, and for all 9 to be HD, the sensor would need to be, let’s see, 17,280 pixels square.

But that math allows the dance scene to look pretty great cropped to undistorted HD pans and even zooms, since 8k via an 180 degree lens is, in fact, enough.

If you’re starting to feel sold on the hemispheric fisheye/single sensor setup, be aware that there are currently a few associated annoyances and minor limitations that I haven’t mentioned:

Nevertheless, ultra wide-angle lenses are a huge part of post-production today, and the set of plug-ins that make up RE:Lens make working with these images possible in After Effects. For the time being, this toolset is After Effects-only; for Premiere Pro VR production tools you can also check out SkyBox Studio from Mettle, which was featured in their booth in the VR pavilion area at NAB 2016.

Exit mobile version