SPN: A novel neural network architecture to improve the performance of MLPs
Abstract
The automated design of neural network architectures has emerged as a key frontier in modern machine learning, driven by the growing complexity of tasks and the scale of available data. Neural Architecture Search (NAS) [7] has enabled researchers to systematically explore vast design spaces, moving beyond manual trial-and-error to discover architectures that strike an optimal balance between performance and efficiency. Despite this progress, many neural network architectures, such as MultiLayer Perceptrons (MLPs) [23], remain limited by conventional connectivity patterns that restrict information flow to simple, hierarchical pathways. This thesis aims to challenge and expand this architectural paradigm. It introduces Sarosh’s Perceptron Networks (SPNs), a novel approach that breaks free from the rigid layer-by-layer connectivity of traditional MLPs and allows neurons more freedom in forming cross-layer connection that lead to more complex architectures. By allowing for more flexible and expressive patterns of neuron connectivity, SPNs seek to unlock new levels of model capability and generalization. This work investigates whether such architectural freedom can yield meaningful improvements in performance and efficiency, and examines the implications for the future of neural network architecture design and the field of NAS- Diplomarbete
- Multi-Layer Perceptron
- Sarosh’s Perceptron Network
- Residual Networks
- Convolutional Neural Networks
- Recurrent Neural Networks
- Vision Transformers
- Neural Architecture Search
- Lottery Ticket Hypothesis
- computer science
- machine learning
- deep learning
- computer vision
- natural language processing
- IT-vetenskaper
- maskininlärning
- djupinlärning
- optisk läsning
- bearbetning av naturligt språk
- tietojenkäsittelytieteet
- koneoppiminen
- syväoppiminen
- konenäkö
- luonnollisen kielen käsittely
- 113 Data- och informationsvetenskap