6533b82afe1ef96bd128b613
RESEARCH PRODUCT
Matrix Shuffle- Exchange Networks for Hard 2D Tasks
Kārlis FreivaldsEmils OzolinsAgris Sosrakssubject
Matrix (mathematics)Dependency (UML)ExploitComputer scienceReceptive fieldBinary logarithmConvolutional neural networkAlgorithmData matrix (multivariate statistics)Data modelingdescription
Convolutional neural networks have become the main tools for processing two-dimensional data. They work well for images, yet convolutions have a limited receptive field that prevents its applications to more complex 2D tasks. We propose a new neural model, called Matrix Shuffle-Exchange network, that can efficiently exploit long-range dependencies in 2D data and has comparable speed to a convolutional neural network. It is derived from Neural Shuffle-Exchange network and has O(log N) layers and O(N ^ 2 log N) total time and O(N^2) space complexity for processing a NxN data matrix. We show that the Matrix Shuffle-Exchange network is well-suited for algorithmic and logical reasoning tasks on matrices and dense graphs, exceeding convolutional and graph neural network baselines. Its distinct advantage is the capability of retaining full long-range dependency modelling when generalizing to larger instances - much larger than could be processed with models equipped with a dense attention mechanism.
year | journal | country | edition | language |
---|---|---|---|---|
2021-07-18 | 2021 International Joint Conference on Neural Networks (IJCNN) |