SVMlight

Description (homepage): SVMlight is an implementation of Vapnik’s Support Vector Machine [Vapnik, 1995] for the problem of pattern recognition, for the problem of regression, and for the problem of learning a ranking function. The optimization algorithms used in SVMlight are described in [Joachims, 2002a ]. [Joachims, 1999a]. The algorithm has scalable memory requirements and can handle problems with many thousands of support vectors efficiently. The software also provides methods for assessing the generalization performance efficiently. It includes two efficient estimation methods for both error rate and precision/recall. XiAlpha-estimates [Joachims, 2002a, Joachims, 2000b] can be computed at essentially no computational expense, but they are conservatively biased. Almost unbiased estimates provides leave-one-out testing. SVMlight exploits that the results of most leave-one-outs (often more than 99%) are predetermined and need not be computed [Joachims, 2002a]. New in this version is an algorithm for learning ranking functions [Joachims, 2002c]. The goal is to learn a function from preference examples, so that it orders a new set of objects as accurately as possible. Such ranking problems naturally occur in applications like search engines and recommender systems. Futhermore, this version includes an algorithm for training large-scale transductive SVMs. The algorithm proceeds by solving a sequence of optimization problems lower-bounding the solution using a form of local search. A detailed description of the algorithm can be found in [Joachims, 1999c]. A similar transductive learner, which can be thought of as a transductive version of k-Nearest Neighbor is the Spectral Graph Transducer. SVMlight can also train SVMs with cost models (see [Morik et al., 1999]). The code has been used on a large range of problems, including text classification [Joachims, 1999c][Joachims, 1998a], image recognition tasks, bioinformatics and medical applications. Many tasks have the property of sparse instance vectors. This implementation makes use of this property which leads to a very compact and efficient representation.


References in zbMATH (referenced in 257 articles )

Showing results 1 to 20 of 257.
Sorted by year (citations)

1 2 3 ... 11 12 13 next

  1. Kantavat, Pittipol; Kijsirikul, Boonserm; Songsiri, Patoomsiri; Fukui, Ken-Ichi; Numao, Masayuki: Efficient decision trees for multi-class support vector machines using entropy and generalization error estimation (2019)
  2. Horn, Daniel; Demircioğlu, Aydın; Bischl, Bernd; Glasmachers, Tobias; Weihs, Claus: A comparative study on large scale kernelized support vector machines (2018)
  3. Ramaswamy, Harish G.; Tewari, Ambuj; Agarwal, Shivani: Consistent algorithms for multiclass classification with an abstain option (2018)
  4. Wang, Yuchuang; Shi, Guoyou; Sun, Xiaotong: A forecast model of the number of containers for containership voyage (2018)
  5. Bacciu, Davide; Carta, Antonio; Gnesi, Stefania; Semini, Laura: An experience in using machine learning for short-term predictions in smart transportation systems (2017)
  6. Duarte Silva, A. Pedro: Optimization approaches to supervised classification (2017)
  7. Farooq, Muhammad; Steinwart, Ingo: An SVM-like approach for expectile regression (2017)
  8. Ingo Steinwart, Philipp Thomann: liquidSVM: A Fast and Versatile SVM package (2017) arXiv
  9. Bai, Yan-Qin; Shen, Kai-Ji: Alternating direction method of multipliers for (\ell_1)-(\ell_2)-regularized logistic regression model (2016)
  10. Bloom, Veronica; Griva, Igor; Quijada, Fabio: Fast projected gradient method for support vector machines (2016)
  11. Doğan, Ürün; Glasmachers, Tobias; Igel, Christian: A unified view on multi-class support vector classification (2016)
  12. Marrocco, C.; Tortorella, F.: Exploiting coding theory for classification: an LDPC-based strategy for multiclass-to-binary decomposition (2016)
  13. Muthu Krishnan, S.: Classify vertebrate hemoglobin proteins by incorporating the evolutionary information into the general PseAAC with the hybrid approach (2016)
  14. Pan, Binbin; Chen, Wen-Sheng; Chen, Bo; Xu, Chen: Efficient learning of supervised kernels with a graph-based loss function (2016)
  15. Zheng, Songfeng: Smoothly approximated support vector domain description (2016)
  16. Mozharovskyi, Pavlo; Mosler, Karl; Lange, Tatjana: Classifying real-world data with the (DD\alpha)-procedure (2015)
  17. Niu, Lingfeng; Zhou, Ruizhi; Zhao, Xi; Shi, Yong: Two new decomposition algorithms for training bound-constrained support vector machines (2015)
  18. Steidl, Gabriele: Supervised learning by support vector machines (2015)
  19. Veelaert, Peter: Combinatorial properties of support vectors of separating hyperplanes (2015)
  20. Beck, Amir: The 2-coordinate descent method for solving double-sided simplex constrained minimization problems (2014)

1 2 3 ... 11 12 13 next