Thomas W. Mitchel1, Benedict Brown2, David Koller3, Tim Weyrich4, Szymon Rusinkiewicz3, Michael Kazhdan1
1 Johns Hopkins University
2 Yale University
4 University College London
Fast methods for convolution and correlation underlie a variety of applications in computer vision and graphics, including efficient filtering, analysis, and simulation. However, standard convolution and correlation are inherently limited to fixed filters: spatial adaptation is impossible without sacrificing efficient computation. In early work, Freeman and Adelson (1991) have shown how steerable filters can address this limitation, providing a way for rotating the filter as it is passed over the signal. In this work, we provide a general, representation-theoretic, framework that allows for spatially varying linear transformations to be applied to the filter. This framework allows for efficient implementation of extended convolution and correlation for transformation groups such as rotation (in 2D and 3D) and scale, and provides a new interpretation for previous methods including steerable filters and the generalized Hough transform. We present applications to pattern matching, image feature description, vector field visualization, and adaptive image filtering.
Thomas W. Mitchel, Benedict Brown, David Koller, Tim Weyrich, Szymon Rusinkiewicz, Michael Kazhdan.
arXiv, 2006.13188, 15 pages, June 2020.Thomas W. Mitchel, Benedict Brown, David Koller, Tim Weyrich, Szymon Rusinkiewicz, and Michael Kazhdan. Efficient spatially adaptive convolution and correlation. arXiv, July 2020.Mitchel, T. W., Brown, B., Koller, D., Weyrich, T., Rusinkiewicz, S., and Kazhdan, M., 2020. Efficient spatially adaptive convolution and correlation. arXiv, July.T. W. Mitchel, B. Brown, D. Koller, T. Weyrich, S. Rusinkiewicz, and M. Kazhdan, “Efficient spatially adaptive convolution and correlation,” arXiv, Jul. 2020.