Researchers in SFU’s Computational Photography Lab hope to give computers a visual advantage that we humans take for granted—the ability to see depth in photograph. While humans naturally can determine how close or far objects are from a single point of view, like a photograph or a painting, it’s a challenge for computers—but one they may soon overcome.
Researchers recently published their work improving a process called monocular depth estimation, a technique that teaches computers how to see depth using machine learning.
“When we look at a picture, we can tell the relative distance of objects by looking at their size, position, and relation to each other,” says Mahdi Miangoleh, an MSc student working in the lab. “This requires recognizing the objects in a scene and knowing what size the objects are in real life. This task alone is an active research topic for neural networks.”
Despite progress in recent years, existing efforts to provide high resolution results that can transform an image into a 3-dimensional (3D) space have failed.
To counter this, the lab recognized the untapped potential of existing neural network models in the literature. The proposed research explains the lack of high-resolution results in current methods through the limitations of convolutional neural networks. Despite major advancements in recent years, the neural networks still have a relatively small capacity to generate many details at once.
Another limitation is how much of the scene these networks can ‘look at’ at once, which determines how much information the neural network can make use of to understand complex scenes. Bu working to increase the resolution of their visual estimations, the researchers are now making it possible to create detailed 3D renderings that look realistic to a human eye. These so-called “depth maps” are used to create 3D renderings of scenes and simulate camera motion in computer graphics.
“Our method analyzes an image and optimizes the process by looking at the image content according to the limitations of current architectures,” explains Ph.D. student Sebastian Dille. “We give our input image to our neural network in many different forms, to create as many details as the model allows while preserving a realistic geometry.”
“With the high-resolution depth maps that the team is able to develop for real-world photographs, artists and content creators can now immediately transfer their photograph or artwork into a rich 3D world,” says computing science professor and lab director, Yağız Aksoy, whose team collaborated with researchers Sylvain Paris and Long Mai, from Adobe Research.
Tools enable artists to turn 2D art into 3D worlds
Global artists are already utilizing the applications enabled by Aksoy’s lab’s research. Akira Saito, a visual artist based in Japan, is creating videos that take viewers into fantastic 3D worlds dreamed up in 2D artwork. To do this he combines tools such as Houdini, a computer animation software, with the depth map generated by Aksoy and his team.
“It’s a great pleasure to see independent artists make use of our technology in their own way,” says Aksoy, whose lab has plans to extend this work to videos and develop new tools that will make depth maps more useful for artists.
“We have made great leaps in computer vision and computer graphics in recent years, but the adoption of these new AI technologies by the artist community needs to be an organic process, and that takes time.”
The post Teaching AI to see Depth in Photograph and Painting appeared first on ELE Times.
No comments:
Post a Comment