Analyzing the depth structure implied in two-dimensional images is one of the most active research areas in computer vision. Here, we propose a method of utilizing texture within an image to derive its depth structure. Though most approaches for deriving depth from a single still image utilize luminance edges and shading to estimate scene structure, relatively little work has been done to utilize the abundant texture information in images. Our new approach begins by analyzing the two cues of local spatial frequency and orientation distributions of the textures within an image, which are used to compute the local slant information across the image. The slant and frequency information are merged to create a unified depth map, providing an important channel for image structure information that can be combined with other available cues. The capabilities of the algorithm are illustrated for a variety of images of planar and curved surfaces under perspective projection, in most of which the depth structure is effortlessly perceived by human observers. Since these operations are readily implementable in neural hardware in early visual cortex, they therefore represent a model of the human perception of the depth structure of images from texture gradient cues.
Christopher W. Tyler, Ajay Gopi, "Computational estimation of scene structure through texture gradient cues" in Proc. IS&T Int’l. Symp. on Electronic Imaging: Human Vision and Electronic Imaging, 2017, pp 167 - 176, https://doi.org/10.2352/ISSN.2470-1173.2017.14.HVEI-138