Snapchat today confirms that it will be among the first apps to put the new technology to use in its iOS app and take the advantage of the new iPhone 12 Pro’s LiDAR Scanner for AR (Augmented Reality). As explained during the Apple’s new iPhone launch event, the LiDAR (Light Detection And Ranging) Scanner measures how prolonged it takes for the light to reach an object and reflect back.
Moreover, the Light Detection and Ranging sensor were first launched on the iPad Pro line-up that brings a suite of new AR (augmented reality) and photographic capabilities to the new iPhone 12 Pro and iPhone Pro Max. In Addition to, Snapchat declares that it plans to reveal a lens specifically made for iPhone 12 Pro devices.
Apple’s new iPhone 12 Pro Max and iPhone 12 Pro models with the Light Detection And Ranging (LiDAR Scanner) sensor start at $1,099 and $999 respectively. However, the company has accommodated this technology for its new iPhone 12 Pro models to improve its low-light photography and the ability to ‘see in the dark’.
The iPhone’s LiDAR sensor itself is a time-of-flight system that can specifically create a depth map of an environment using lasers. The outcome is quicker and much more accurate and precise augmented reality (AR), as well as new opportunities to use Apple’s new ARKit in creative and innovative ways.
Besides, the LiDAR Scanner technology can also be used by app inaugurators to manufacture an accurate depth map of the scene and that can help speed up AR (augmented reality) so it feels more expeditious while authorizing new app experiences that use AR.
Also, the technology can delegate photo and video effects and more exact placement of AR objects, as the iPhone is actually able to ‘see’ a depth map of the room.