Hennick, Max2023-03-012023-03-012021https://unbscholar.lib.unb.ca/handle/1882/14359In recent years, two distinct paradigms have been emerging in machine learning. The first is the “data driven” paradigm which makes minimal assumptions about the underlying structure of the data, and instead relies on huge swathes of data and very general algorithms to achieve state-of-the-art performance. The other can be referred to as the “inductive prior” paradigm, which makes prior assumptions about the structure of data, and attempts to design algorithms which exploit this structure, i.e they have a “prior belief” about their input data. Of interest in this work are deep learning systems belonging to the second paradigm which are built to exploit data with a group-like structure. In this work, we provide an algebraic foundation for such “group equivariant neural networks”, and use this foundation to examine their properties. Furthermore, we provide a theorem which gives an explicit method for creating such networks from the underlying structure of the group.text/xmlx, 125 pageselectronicen-CAhttp://purl.org/coar/access_right/c_abf2An algebraic approach to group equivariant neural networksmaster thesis2023-03-01Touikan, NicholasMathematics and Statistics