Paper ID | MLR-APPL-MDSP.2 | ||
Paper Title | REMOVING DIMENSIONAL RESTRICTIONS ON COMPLEX/HYPER-COMPLEX NEURAL NETWORKS | ||
Authors | Chase Gaudet, Anthony Maida, University of Louisiana at Lafayette, United States | ||
Session | MLR-APPL-MDSP: Machine learning for multidimensional signal processing | ||
Location | Area F | ||
Session Time: | Monday, 20 September, 13:30 - 15:00 | ||
Presentation Time: | Monday, 20 September, 13:30 - 15:00 | ||
Presentation | Poster | ||
Topic | Applications of Machine Learning: Machine learning for multidimensional signal processing | ||
IEEE Xplore Open Preview | Click here to view in IEEE Xplore | ||
Abstract | It has been shown that the core reasons that complex and hypercomplex valued neural networks offer improvements over their real-valued counterparts is the fact that aspects of their algebra forces treating multi-dimensional data as a single entity. However, both are constrained to a set number of dimensions, two for complex and four for quaternions. These observations motivate us to introduce novel vector map convolutions which capture this property, while dropping the unnatural dimensionality constraints their algebra imposes. This is achieved by introducing a system that mimics the unique linear combination of input dimensions via the Hamilton product using a permutation function, as well as batch normalization and weight initialization for the system. We perform two experiments to show that these novel vector map convolutions seem to capture all the benefits of complex and hyper-complex networks, while avoiding the dimensionality restriction. |