Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Rotation Equivariant Operators for Machine Learning on Scalar and Vector Fields (2108.09541v3)

Published 21 Aug 2021 in cs.LG

Abstract: We develop theory and software for rotation equivariant operators on scalar and vector fields, with diverse applications in simulation, optimization and machine learning. Rotation equivariance (covariance) means all fields in the system rotate together, implying spatially invariant dynamics that preserve symmetry. Extending the convolution theorems of linear time invariant systems, we theorize that linear equivariant operators are characterized by tensor field convolutions using an appropriate product between the input field and a radially symmetric kernel field. Most Green's functions and differential operators are in fact equivariant operators, which can also fit unknown symmetry preserving dynamics by parameterizing the radial function. We implement the Julia package EquivariantOperators.jl for fully differentiable finite difference equivariant operators on scalar, vector and higher order tensor fields in 2d/3d. It can run forwards for simulation or image processing, or be back propagated for computer vision, inverse problems and optimal control. Code at https://aced-differentiate.github.io/EquivariantOperators.jl/

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Paul Shen (2 papers)
  2. Michael Herbst (2 papers)
  3. Venkat Viswanathan (4 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com