SVMlight

Description (homepage): SVMlight is an implementation of Vapnik’s Support Vector Machine [Vapnik, 1995] for the problem of pattern recognition, for the problem of regression, and for the problem of learning a ranking function. The optimization algorithms used in SVMlight are described in [Joachims, 2002a ]. [Joachims, 1999a]. The algorithm has scalable memory requirements and can handle problems with many thousands of support vectors efficiently. The software also provides methods for assessing the generalization performance efficiently. It includes two efficient estimation methods for both error rate and precision/recall. XiAlpha-estimates [Joachims, 2002a, Joachims, 2000b] can be computed at essentially no computational expense, but they are conservatively biased. Almost unbiased estimates provides leave-one-out testing. SVMlight exploits that the results of most leave-one-outs (often more than 99%) are predetermined and need not be computed [Joachims, 2002a]. New in this version is an algorithm for learning ranking functions [Joachims, 2002c]. The goal is to learn a function from preference examples, so that it orders a new set of objects as accurately as possible. Such ranking problems naturally occur in applications like search engines and recommender systems. Futhermore, this version includes an algorithm for training large-scale transductive SVMs. The algorithm proceeds by solving a sequence of optimization problems lower-bounding the solution using a form of local search. A detailed description of the algorithm can be found in [Joachims, 1999c]. A similar transductive learner, which can be thought of as a transductive version of k-Nearest Neighbor is the Spectral Graph Transducer. SVMlight can also train SVMs with cost models (see [Morik et al., 1999]). The code has been used on a large range of problems, including text classification [Joachims, 1999c][Joachims, 1998a], image recognition tasks, bioinformatics and medical applications. Many tasks have the property of sparse instance vectors. This implementation makes use of this property which leads to a very compact and efficient representation.


References in zbMATH (referenced in 237 articles )

Showing results 1 to 20 of 237.
Sorted by year (citations)

1 2 3 ... 10 11 12 next

  1. Muthu Krishnan, S.: Using Chou’s general pseaac to analyze the evolutionary relationship of receptor associated proteins (RAP) with various folding patterns of protein domains (2018)
  2. Bacciu, Davide; Carta, Antonio; Gnesi, Stefania; Semini, Laura: An experience in using machine learning for short-term predictions in smart transportation systems (2017)
  3. Duarte Silva, A. Pedro: Optimization approaches to supervised classification (2017)
  4. Farooq, Muhammad; Steinwart, Ingo: An SVM-like approach for expectile regression (2017)
  5. Ingo Steinwart, Philipp Thomann: liquidSVM: A Fast and Versatile SVM package (2017) arXiv
  6. Bai, Yan-Qin; Shen, Kai-Ji: Alternating direction method of multipliers for $\ell_1$-$\ell_2$-regularized logistic regression model (2016)
  7. Bloom, Veronica; Griva, Igor; Quijada, Fabio: Fast projected gradient method for support vector machines (2016)
  8. Doğan, Ürün; Glasmachers, Tobias; Igel, Christian: A unified view on multi-class support vector classification (2016)
  9. Muthu Krishnan, S.: Classify vertebrate hemoglobin proteins by incorporating the evolutionary information into the general PseAAC with the hybrid approach (2016)
  10. Souillard-Mandar, William; Davis, Randall; Rudin, Cynthia; Au, Rhoda; Libon, David J.; Swenson, Rodney; Price, Catherine C.; Lamar, Melissa; Penney, Dana L.: Learning classification models of cognitive conditions from subtle behaviors in the digital clock drawing test (2016)
  11. Zheng, Songfeng: Smoothly approximated support vector domain description (2016)
  12. Niu, Lingfeng; Zhou, Ruizhi; Zhao, Xi; Shi, Yong: Two new decomposition algorithms for training bound-constrained support vector machines (2015)
  13. Steidl, Gabriele: Supervised learning by support vector machines (2015)
  14. Veelaert, Peter: Combinatorial properties of support vectors of separating hyperplanes (2015)
  15. Beck, Amir: The 2-coordinate descent method for solving double-sided simplex constrained minimization problems (2014)
  16. Bridge, James P.; Holden, Sean B.; Paulson, Lawrence C.: Machine learning for first-order theorem proving (2014)
  17. Brooks, J. Paul; Lee, Eva K.: Solving a multigroup mixed-integer programming-based constrained discrimination model (2014)
  18. Carrizosa, Emilio; Martín-Barragán, Belén; Morales, Dolores Romero: A nested heuristic for parameter tuning in support vector machines (2014)
  19. Chen, Xiaobo; Yang, Jian; Chen, Long: An improved robust and sparse twin support vector regression via linear programming (2014)
  20. Huang, Zongyan; England, Matthew; Wilson, David; Davenport, James H.; Paulson, Lawrence C.; Bridge, James: Applying machine learning to the problem of choosing a heuristic to select the variable ordering for cylindrical algebraic decomposition (2014)

1 2 3 ... 10 11 12 next