Academia.eduAcademia.edu

Neural Network Architecture

description34 papers
group3 followers
lightbulbAbout this topic
Neural Network Architecture refers to the structured design of artificial neural networks, encompassing the arrangement of layers, types of neurons, and connections between them. It determines how data is processed and learned, influencing the network's performance in tasks such as classification, regression, and pattern recognition.
lightbulbAbout this topic
Neural Network Architecture refers to the structured design of artificial neural networks, encompassing the arrangement of layers, types of neurons, and connections between them. It determines how data is processed and learned, influencing the network's performance in tasks such as classification, regression, and pattern recognition.

Key research themes

1. How can neural network architecture be optimized for computational efficiency without sacrificing accuracy?

This research area focuses on designing and scaling neural network architectures to achieve high accuracy on specified tasks while minimizing computational complexity and hardware resource usage. It is critical for deploying neural networks on resource-limited devices and speeding up inference by reducing operations and hardware area.

Key finding: Proposed a two-step paradigm integrating compound model scaling (a lightweight NAS approach) and Entropy-Constrained Trained Ternarization (EC2T), a simultaneous pruning and ternary quantization algorithm, which compresses... Read more
Key finding: Introduced a neural network hardware design that reduces the number of physical hidden layers by half (from N to N/2) through multiplexing input and output layers while maintaining the same accuracy as traditional N-layer... Read more
Key finding: Designed a neural network architecture sharing multipliers and adders between two hidden layers, cutting the number of these critical hardware components by half and reducing hardware cost by 63%. The method maintained... Read more
Key finding: Derived analytical formulas to estimate upper bounds on the number of hidden layers and neurons in networks trained via algorithms using the Jacobi matrix (e.g., Levenberg-Marquardt). These bounds aid in selecting compact yet... Read more
Key finding: Developed a heuristic architecture search method leveraging network morphism combined with hill-climbing and functional saving, achieving competitive chest X-ray classification accuracy (73.2% validation accuracy, 84.5% AUC)... Read more

2. What methodologies and algorithms enable automated search and optimization of neural network architectures to improve performance and reduce manual design efforts?

This research theme investigates algorithmic frameworks and search strategies such as genetic algorithms, evolutionary methods, modular search spaces, and heuristics to automate the process of neural network architecture design. Automating architecture search accelerates model development, improves generalization, and allows discovering architectures difficult to design manually, helping in diverse tasks from image classification to medical imaging.

Key finding: Proposed a systematic framework that categorizes and benchmarks NAS methods by summarizing architecture search decisions and strategies, applying quantitative and qualitative metrics for prototyping and comparison. This... Read more
Key finding: Presented a novel heuristic architecture search using enforced hill-climbing and network morphism to efficiently explore architectures. The method found high-performing architectures within 28 GPU hours on medical image... Read more
Key finding: Developed hybrid training algorithms combining genetic algorithms (GA) and backpropagation (BP) that leverage GA’s global search to escape local minima and BP’s efficiency in fine-tuning. The GA-BP hybrids achieved faster... Read more
Key finding: Proposed a modularized neural architecture search space composed of parameterized building blocks derived from NAS-Bench-201 benchmark, represented as multisectoral networks described unambiguously by vectors. Applied to a... Read more
Key finding: Explored an ANN design leveraging massive parallelism with many interconnected processing elements distributed over parallel processors, achieving effective optimization for nonlinear resource allocation problems.... Read more

3. How do architectural elements and training hyperparameters influence neural network learning dynamics and generalization?

This theme examines the role of architectural design choices, such as the number of layers, neurons, and activation functions, as well as learning hyperparameters like learning rate and regularization, on convergence, error minimization, and avoidance of local minima. Understanding these influences is vital to achieve stable and efficient learning with good generalization while preventing issues like overfitting or chaotic training behavior.

Key finding: Derived novel gradient-based algorithms for estimating regularization parameters and optimizing neural net architectures using a validation set. Proposed iterative schemes jointly optimizing weights and hyperparameters that... Read more
Key finding: Analyzed the effect of learning rate (η) on multilayer neural network training, observing bifurcation and chaotic behavior when η exceeds a critical threshold (~0.62 for a 3-layer network with 4 neurons per layer). Found that... Read more
Key finding: Identified limitations of backpropagation training related to sensitivity to learning rate and momentum and susceptibility to local minima. Showed that integrating GA with BP alleviates these issues by global exploration with... Read more

All papers in Neural Network Architecture

For the past years, many researchers and engineers have been developing and optimising deep neural networks (DNN). The process of neural architecture design and tuning its hyperparameters remains monotonous, timeconsuming, and do not... more
A problem of improving the performance of convolutional neural networks is considered. A parameter of the training set is investigated. The parameter is the batch size. The goal is to find an impact of training set batch size on the... more
The past years of research have shown that automated machine learning and neural architecture search are an inevitable future for image recognition tasks. In addition, a crucial aspect of any automated search is the predefined search... more
for generic signal processing applications. In the proposed paper analog components like Gilbert Cell Multiplier (GCM), Neuron activation Function (NAF) are used to implement artificial NNA. The analog components used are comprises of... more
Convolutional neural networks (CNNs) have taken the spotlight in a variety of machine learning applications. To reach the desired performance, CNNs have become increasingly deeper and larger which goes along with a tremendous amount of... more
In the past decade, a new way in neural networks research called Network architectures search has demonstrated noticeable results in the design of architectures for image segmentation and classification. Despite the considerable success... more
With the advent of new technologies and advancement in medical science we are trying to process the information artificially as our biological system performs inside our body. Artificial intelligence through a biological word is realized... more
As animals adapt to their environments, their brains are tasked with processing stimuli in different sensory contexts. Whether these computations are context dependent or independent, they are all implemented in the same neural tissue. A... more
This piece of research introduces a purely data-driven, directly reconfigurable, divide-and-conquer on-line monitoring (OLM) methodology for automatically selecting the minimum number of neutron detectors (NDs)-and corresponding neutron... more
Download research papers for free!