Academia.eduAcademia.edu

Support Vector Machines for regression

description7 papers
group1 follower
lightbulbAbout this topic
Support Vector Machines for regression (SVR) is a supervised learning algorithm that extends Support Vector Machines to predict continuous outcomes. It identifies a function that approximates the relationship between input features and target values by minimizing prediction error while maintaining a margin of tolerance, thus ensuring robustness against overfitting.
lightbulbAbout this topic
Support Vector Machines for regression (SVR) is a supervised learning algorithm that extends Support Vector Machines to predict continuous outcomes. It identifies a function that approximates the relationship between input features and target values by minimizing prediction error while maintaining a margin of tolerance, thus ensuring robustness against overfitting.

Key research themes

1. How can Support Vector Machines be adapted and optimized for effective regression tasks in time series forecasting and high-dimensional data?

This area focuses on extending SVM methodologies, particularly Support Vector Regression (SVR), to handle complex regression problems such as time series prediction and high-dimensional feature spaces. It examines adaptations including different loss functions, online and incremental learning algorithms, kernel selections, and integration with optimization and dimensionality reduction techniques. These adaptations aim to improve generalization, computational efficiency, and applicability in real-world scenarios involving nonlinear and large-volume data.

Key finding: Introduced an exact incremental algorithm for ε-insensitive SVR that allows incremental addition, removal, and updating of training points, enabling online regression learning, particularly useful for streaming data or... Read more
Key finding: Proposed a novel convex ϵ-penalty loss function generalizing and encompassing ϵ-insensitive and Laplace losses, leading to two new SVR variants with either L2 or L1 norm regularization. These models utilize penalization... Read more
Key finding: Reviewed and integrated principal component methods with SVR to handle high-dimensional data where p (features) > n (samples), overcoming matrix inversion issues in least squares regression. Demonstrated that combining... Read more
Key finding: Developed a hybrid prediction framework combining SVR with recurrent neural networks (RNN) optimized via metaheuristic parameter tuning, addressing limitations of traditional RNN training and improving prediction robustness... Read more
Key finding: Applied SVR with comprehensive kernel parameter tuning and innovative windowing data preprocessing for forecasting rainfall time series, demonstrating that careful parameter selection and input preprocessing significantly... Read more

2. What methods effectively enhance computational efficiency and scalability of Support Vector Regression models, particularly with nonlinear kernels?

Given the computational and memory costs associated with large-scale SVR models, especially those using nonlinear kernels like RBF, this theme explores algorithmic and approximation techniques to speed up prediction and training times without significant accuracy loss. It examines methods such as kernel approximation, online/incremental learning, and parallel/distributed computation, focusing on making SVR applicable to real-time and big data contexts.

Key finding: Presented a second-order Maclaurin series approximation for RBF kernel evaluation in SVR that reduces prediction complexity from dependence on the number of support vectors to quadratic in input dimensionality, enabling... Read more
Key finding: Introduced an incremental online learning algorithm for ε-insensitive SVR allowing dynamic modification of training points and immediate model updates without retraining from scratch, thereby increasing computational... Read more
Key finding: Developed PSOGS, a hybrid hyperparameter optimization algorithm combining Particle Swarm Optimization and Grid Search to efficiently tune SVR hyperparameters (C, ε, γ), demonstrating improved prediction accuracy and... Read more
Key finding: Conducted an empirical comparison of explicit parallelization (e.g., SMO with hand-parallelized components) versus implicit parallelization approaches (expressing SVM algorithms with large dense linear algebra operations to... Read more

3. How do Support Vector Regression models perform in practical predictive applications such as stock market, rainfall prediction, student performance, and hydrological drought forecasting?

This theme investigates the application of SVR models tailored through domain-specific preprocessing, feature selection, and hybrid modeling strategies to improve prediction accuracy and reliability in real-world regression problems involving complex nonlinear and time-dependent data. It emphasizes practical insights from parameter tuning, integration with windowing or dimensionality reduction, and comparison with alternative regression methods in various domains.

Key finding: Demonstrated that incorporating different windowing operators as data preprocessing inputs significantly enhances SVR's prediction accuracy for rainfall time series on a four-year Dhaka Stock Exchange dataset, with evaluation... Read more
Key finding: Applied SVR alongside Gene Expression Programming and M5 model trees to model drought indices including Standardized Streamflow Index (SSI), finding SVR competitive for predicting monthly precipitation and recommending M5... Read more
Key finding: Built machine learning models using SVR and logistic regression to predict diabetes risk and disease state with high accuracy (90-92%), emphasizing model scalability, usability, and reliability in handling patient datasets,... Read more
Key finding: Developed an improved nonlinear regression prediction model using a hybrid of SVR and recurrent neural networks (RNN), applying metaheuristic optimization for parameter tuning, leading to superior performance compared to... Read more

All papers in Support Vector Machines for regression

The ability of Minkowski Functionals to characterize local structure in different biological tissue types has been demonstrated in a variety of medical image processing tasks. We introduce anisotropic Minkowski Functionals (AMFs) as a... more
Hydrological droughts are characterized based on their duration, severity, and magnitude. Among the most critical factors, precipitation, evapotranspiration, and runoff are essential in modeling the droughts. In this study, three indices... more
Download research papers for free!