Academia.eduAcademia.edu

Neural Network Architecture

description34 papers
group3 followers
lightbulbAbout this topic
Neural Network Architecture refers to the structured design of artificial neural networks, encompassing the arrangement of layers, types of neurons, and connections between them. It determines how data is processed and learned, influencing the network's performance in tasks such as classification, regression, and pattern recognition.
lightbulbAbout this topic
Neural Network Architecture refers to the structured design of artificial neural networks, encompassing the arrangement of layers, types of neurons, and connections between them. It determines how data is processed and learned, influencing the network's performance in tasks such as classification, regression, and pattern recognition.

Key research themes

1. How can neural network architecture be optimized for computational efficiency without sacrificing accuracy?

This research area focuses on designing and scaling neural network architectures to achieve high accuracy on specified tasks while minimizing computational complexity and hardware resource usage. It is critical for deploying neural networks on resource-limited devices and speeding up inference by reducing operations and hardware area.

Key finding: Proposed a two-step paradigm integrating compound model scaling (a lightweight NAS approach) and Entropy-Constrained Trained Ternarization (EC2T), a simultaneous pruning and ternary quantization algorithm, which compresses... Read more
Key finding: Introduced a neural network hardware design that reduces the number of physical hidden layers by half (from N to N/2) through multiplexing input and output layers while maintaining the same accuracy as traditional N-layer... Read more
Key finding: Designed a neural network architecture sharing multipliers and adders between two hidden layers, cutting the number of these critical hardware components by half and reducing hardware cost by 63%. The method maintained... Read more
Key finding: Derived analytical formulas to estimate upper bounds on the number of hidden layers and neurons in networks trained via algorithms using the Jacobi matrix (e.g., Levenberg-Marquardt). These bounds aid in selecting compact yet... Read more
Key finding: Developed a heuristic architecture search method leveraging network morphism combined with hill-climbing and functional saving, achieving competitive chest X-ray classification accuracy (73.2% validation accuracy, 84.5% AUC)... Read more

2. What methodologies and algorithms enable automated search and optimization of neural network architectures to improve performance and reduce manual design efforts?

This research theme investigates algorithmic frameworks and search strategies such as genetic algorithms, evolutionary methods, modular search spaces, and heuristics to automate the process of neural network architecture design. Automating architecture search accelerates model development, improves generalization, and allows discovering architectures difficult to design manually, helping in diverse tasks from image classification to medical imaging.

Key finding: Proposed a systematic framework that categorizes and benchmarks NAS methods by summarizing architecture search decisions and strategies, applying quantitative and qualitative metrics for prototyping and comparison. This... Read more
Key finding: Presented a novel heuristic architecture search using enforced hill-climbing and network morphism to efficiently explore architectures. The method found high-performing architectures within 28 GPU hours on medical image... Read more
Key finding: Developed hybrid training algorithms combining genetic algorithms (GA) and backpropagation (BP) that leverage GA’s global search to escape local minima and BP’s efficiency in fine-tuning. The GA-BP hybrids achieved faster... Read more
Key finding: Proposed a modularized neural architecture search space composed of parameterized building blocks derived from NAS-Bench-201 benchmark, represented as multisectoral networks described unambiguously by vectors. Applied to a... Read more
Key finding: Explored an ANN design leveraging massive parallelism with many interconnected processing elements distributed over parallel processors, achieving effective optimization for nonlinear resource allocation problems.... Read more

3. How do architectural elements and training hyperparameters influence neural network learning dynamics and generalization?

This theme examines the role of architectural design choices, such as the number of layers, neurons, and activation functions, as well as learning hyperparameters like learning rate and regularization, on convergence, error minimization, and avoidance of local minima. Understanding these influences is vital to achieve stable and efficient learning with good generalization while preventing issues like overfitting or chaotic training behavior.

Key finding: Derived novel gradient-based algorithms for estimating regularization parameters and optimizing neural net architectures using a validation set. Proposed iterative schemes jointly optimizing weights and hyperparameters that... Read more
Key finding: Analyzed the effect of learning rate (η) on multilayer neural network training, observing bifurcation and chaotic behavior when η exceeds a critical threshold (~0.62 for a 3-layer network with 4 neurons per layer). Found that... Read more
Key finding: Identified limitations of backpropagation training related to sensitivity to learning rate and momentum and susceptibility to local minima. Showed that integrating GA with BP alleviates these issues by global exploration with... Read more

All papers in Neural Network Architecture

This paper aims to present a comprehensive review of advanced techniques and models with a specific focus on deep neural network (DNN) for resource-constrained environments (RCE). The paper contributes by highlighting the RCE devices,... more
Air surveillance radar tracking systems present a variety of known problems related to uncertainty and lack of accurately in radar measurements used as source in these systems. In this work, we feature the theoretical aspects of a... more
This paper aims to investigate whether applying the Singular Value Decomposition (SVD) technique can reduce the workload of Convolutional Neural Networks (CNNs) without compromising image classification results. Usual methods for reducing... more
The scaling hypothesis, the long-standing paradigm that posited predictable performance gains in artificial intelligence from increasing model size, data, and compute, is encountering significant and compounding limitations. Empirical... more
Convolutional neural networks (CNNs) have shown remarkable performance in various real-world applications. Unfortunately, the promising performance of CNNs can be achieved only when their architectures are optimally constructed. The... more
The performance of Convolutional Neural Networks (CNNs) highly relies on their architectures. In order to design a CNN with promising performance, extensive expertise in both CNNs and the investigated problem domain is required, which is... more
We propose a novel convolutional neural networks (CNNs) training procedure to allow dynamically trade-offs between different resource and performance requirements. Our approach prioritizes the channels to enable structured sparsity and... more
We propose a novel convolutional neural networks (CNNs) training procedure to allow dynamically trade-offs between different resource and performance requirements. Our approach prioritizes the channels to enable structured sparsity and... more
Tiny deep learning has attracted increasing attention driven by the substantial demand for deploying deep learning on numerous intelligent Internet-of-Things devices. However, it is still challenging to unleash tiny deep learning's full... more
Artificial neural network play an important role in VLSI circuit to find and diagnosis multiple fault in digital circuit. In this paper, the example of single layer and multilayer neural network had been discussed secondly implement those... more
This study explores the integration of neural network emulators into Computer-Aided Design (CAD) environments, aiming to enhance neural architecture modeling. The objective is to merge the computational power of neural networks with the... more
Diabetic retinopathy (DR) is a major cause of vision loss globally, making early detection vital for effective intervention. However, manual screening is prone to errors and inefficiency. Automated solutions using deep learning models... more
This research paper presents an advanced exploration of the Φπε framework, an integrative model that unifies quantum mechanics, consciousness studies, and metaphysical principles under the auspices of harmony (Φ), quantum entanglement... more
We consider network sparsification as an L0-norm regularized binary optimization problem, where each unit of a neural network (e.g., weight, neuron, or channel, etc.) is attached with a stochastic binary gate, whose parameters are jointly... more
Model compression is an essential technique for deploying deep neural networks (DNNs) on power and memory-constrained resources. However, existing model-compression methods often rely on human expertise and focus on parameters' local... more
Network quantization is an effective solution to compress deep neural networks for practical usage. Existing network quantization methods cannot sufficiently exploit the depth information to generate low-bit compressed network. In this... more
Recent research has focused on weight sparsity in neural network training to reduce FLOPs, aiming for improved efficiency (test accuracy w.r.t training FLOPs). However, sparse weight training often sacrifices accuracy, requiring extended... more
In this paper, a new pruning strategy based on the neuroplasticity of biological neural networks is presented. The novel pruning algorithm proposed is inspired by the knowledge remapping ability after injuries in the cerebral cortex.... more
Network Morphism based Neural Architecture Search (NAS) is one of the most efficient methods, however, knowing where and when to add new neurons or remove dis-functional ones is generally left to black-box Reinforcement Learning models.... more
Recent research has focused on weight sparsity in neural network training to reduce FLOPs, aiming for improved efficiency (test accuracy w.r.t training FLOPs). However, sparse weight training often sacrifices accuracy, requiring extended... more
Neural Architecture Search (NAS) methods have been growing in popularity. These techniques have been fundamental to automate and speed up the time consuming and error-prone process of synthesizing novel Deep Learning (DL) architectures.... more
Neural architecture search (NAS) has emerged as a promising avenue for automatically designing task-specific neural networks. Most existing NAS approaches require one complete search for each deployment specification of hardware or... more
In recent developments, the significance of Edge AI has come to the forefront. Edge devices, which encompass a wide array of IoT devices and embedded systems, benefit from the deployment of efficient and compact neural network models.... more
Most deep architectures for image classification-even those that are trained to classify a large number of diverse categories-learn shared image representations with a single model. Intuitively, however, categories that are more similar... more
Convolutional neural networks (CNNs) introduce state-ofthe-art results for various tasks with the price of high computational demands. Inspired by the observation that spatial correlation exists in CNN output feature maps (ofms), we... more
The performance of Convolutional Neural Networks (CNNs) highly relies on their architectures. In order to design a CNN with promising performance, extensive expertise in both CNNs and the investigated problem domain is required, which is... more
Animals display rich and coordinated motor patterns during walking and running. Previous modelling as well as experimental results suggest that the balance between excitation and inhibition in neural networks may be critical for... more
Although radar and communications signal classification are usually treated separately, they share similar characteristics, and methods applied in one domain can be potentially applied in the other. We propose a simple and unified scheme... more
State-of-the-art computer vision models are rapidly increasing in capacity, where the number of parameters far exceeds the number required to fit the training set. This results in better optimization and generalization performance.... more
When deploying deep learning models to a device, it is traditionally assumed that available computational resources (compute, memory, and power) remain static. However, real-world computing systems do not always provide stable resource... more
Previous neural architecture search (NAS) approaches for mobile platforms have achieved great success in designing a slim-but-accurate neural network that is generally wellmatched to a single computing unit such as a CPU or GPU. However,... more
Humans and other animals navigate different landscapes and environments with ease, a feat that requires the brain’s ability to rapidly and accurately adapt to different visual domains, generalizing across contexts/backgrounds. Despite... more
The detection of novel stimuli is critical to learn and survive in a dynamic environment. Though novel stimuli powerfully affect brain activity, their impact on specific cell types and circuits is not well understood. Disinhibition is one... more
Implicit Neural Representation (INR) is an innovative approach for representing complex shapes or objects without explicitly defining their geometry or surface structure. Instead, INR represents objects as continuous functions. Previous... more
Efficient continual learning in humans is enabled by a rich set of neurophysiological mechanisms and interactions between multiple memory systems. The brain efficiently encodes information in non-overlapping sparse codes, which... more
In this paper, we present the implementation of a system that identifies the modulation of complex radio signals. This is realized using an artificial intelligence model developed, trained, and integrated with Microsoft Azure cloud. We... more
The majority of synaptic inputs to the primary visual cortex (V1) are non-feedforward, instead originating from local and anatomical feedback connections. Animal electrophysiology experiments show that feedback signals originating from... more
The majority of synaptic inputs to the primary visual cortex (V1) are non-feedforward, instead originating from local and anatomical feedback connections. Animal electrophysiology experiments show that feedback signals originating from... more
This paper aims to find an automatic solution for the modulation’s classification of different types of radio signals by relying on Artificial Intelligence. This project is part of a long process of Communications Intelligence looking for... more
Keyword Spotting (KWS) has been the subject of research in recent years given the increase of embedded systems for command recognition such as Alexa, Google Home, and Siri. Performance, model size, processing time, and robustness to noise... more
Spatial attention enhances sensory processing of goal-relevant information and improves perceptual sensitivity. Yet, the specific neural mechanisms underlying the effects of spatial attention on performance are still contested. Here, we... more
In Convolutional Neural Networks (CNNs) information flows across a small neighbourhood of each pixel of an image, preventing long-range integration of features before reaching deep layers in the network. We propose a novel architecture... more
Turning the weights to zero when training a neural network helps in reducing the computational complexity at inference. To progressively increase the sparsity ratio in the network without causing sharp weight discontinuities during... more
Turning the weights to zero when training a neural network helps in reducing the computational complexity at inference. To progressively increase the sparsity ratio in the network without causing sharp weight discontinuities during... more
Designing Deep Neural Networks (DNNs) running on edge hardware remains a challenge. Standard designs have been adopted by the community to facilitate the deployment of Neural Network models. However, not much emphasis is put on adapting... more
Motivated by the recent trend towards the usage of larger receptive fields for more context-aware neural networks in vision applications, we aim to investigate how large these receptive fields really need to be. To facilitate such study,... more
Download research papers for free!