Papers
arxiv:2301.10152

How Jellyfish Characterise Alternating Group Equivariant Neural Networks

Published on Jan 24, 2023

Abstract

The study characterizes all possible alternating group ($A_n$) equivariant neural networks using tensor powers of $\mathbb{R}^{n}$ and provides a basis for their learnable, linear layer functions, which also generalize to local symmetries.

AI-generated summary

We provide a full characterisation of all of the possible alternating group (A_n) equivariant neural networks whose layers are some tensor power of R^{n}. In particular, we find a basis of matrices for the learnable, linear, A_n-equivariant layer functions between such tensor power spaces in the standard basis of R^{n}. We also describe how our approach generalises to the construction of neural networks that are equivariant to local symmetries.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2301.10152
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2301.10152 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2301.10152 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2301.10152 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.