Classification algorithms have been widely used in many application domains. Most of these domains deal with massive collection of data and hence demand classification algorithms that scale well with the size of the data sets involved. A classification algorithm is said to be scalable if there is no significant increase in time and space requirements for the algorithm (without compromising the generalization performance) when dealing with an increase in the training set size. Support Vector Machine (SVM) is one of the most celebrated kernel based classification methods used in Machine Learning. An SVM capable of handling large scale classification problems will definitely be an ideal candidate in many real world applications. The training process involved in SVM classifier is usually formulated as a Quadratic Programing (QP) problem. The existing solution strategies for this problem have an associated time and space complexity that is (at least) quadratic in the number of training points. It makes SVM training very expensive. This thesis addresses the scalability of the training algorithms involved in SVM to make it feasible with large training data sets.
"Sinopsis" puede pertenecer a otra edición de este libro.
Classification algorithms have been widely used in many application domains. Most of these domains deal with massive collection of data and hence demand classification algorithms that scale well with the size of the data sets involved. A classification algorithm is said to be scalable if there is no significant increase in time and space requirements for the algorithm (without compromising the generalization performance) when dealing with an increase in the training set size. Support Vector Machine (SVM) is one of the most celebrated kernel based classification methods used in Machine Learning. An SVM capable of handling large scale classification problems will definitely be an ideal candidate in many real world applications. The training process involved in SVM classifier is usually formulated as a Quadratic Programing (QP) problem. The existing solution strategies for this problem have an associated time and space complexity that is (at least) quadratic in the number of training points. It makes SVM training very expensive. This thesis addresses the scalability of the training algorithms involved in SVM to make it feasible with large training data sets.
Asharaf S is a faculty member in the IT&Systems area in IIM, Kozhikode, India. He received his PhD and Master of Engineering degrees from Indian Institute of Science, Bangalore. Prior to joining IIMK, he has been with America Online as a Research Scientist. He is a recipient of IBM Outstanding PhD student award.
"Sobre este título" puede pertenecer a otra edición de este libro.
Librería: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Alemania
Taschenbuch. Condición: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -Classification algorithms have been widely used in many application domains. Most of these domains deal with massive collection of data and hence demand classification algorithms that scale well with the size of the data sets involved. A classification algorithm is said to be scalable if there is no significant increase in time and space requirements for the algorithm (without compromising the generalization performance) when dealing with an increase in the training set size. Support Vector Machine (SVM) is one of the most celebrated kernel based classification methods used in Machine Learning. An SVM capable of handling large scale classification problems will definitely be an ideal candidate in many real world applications. The training process involved in SVM classifier is usually formulated as a Quadratic Programing (QP) problem. The existing solution strategies for this problem have an associated time and space complexity that is (at least) quadratic in the number of training points. It makes SVM training very expensive. This thesis addresses the scalability of the training algorithms involved in SVM to make it feasible with large training data sets. 132 pp. Englisch. Nº de ref. del artículo: 9783846541463
Cantidad disponible: 2 disponibles
Librería: moluna, Greven, Alemania
Condición: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. Autor/Autorin: S AsharafAsharaf S is a faculty member in the IT&Systems area in IIM, Kozhikode, India. He received his PhD and Master of Engineering degrees from Indian Institute of Science, Bangalore. Prior to joining IIMK, he has been with Americ. Nº de ref. del artículo: 5497818
Cantidad disponible: Más de 20 disponibles
Librería: buchversandmimpf2000, Emtmannsberg, BAYE, Alemania
Taschenbuch. Condición: Neu. This item is printed on demand - Print on Demand Titel. Neuware -Classification algorithms have been widely used in many application domains. Most of these domains deal with massive collection of data and hence demand classification algorithms that scale well with the size of the data sets involved. A classification algorithm is said to be scalable if there is no significant increase in time and space requirements for the algorithm (without compromising the generalization performance) when dealing with an increase in the training set size. Support Vector Machine (SVM) is one of the most celebrated kernel based classification methods used in Machine Learning. An SVM capable of handling large scale classification problems will definitely be an ideal candidate in many real world applications. The training process involved in SVM classifier is usually formulated as a Quadratic Programing (QP) problem. The existing solution strategies for this problem have an associated time and space complexity that is (at least) quadratic in the number of training points. It makes SVM training very expensive. This thesis addresses the scalability of the training algorithms involved in SVM to make it feasible with large training data sets.VDM Verlag, Dudweiler Landstraße 99, 66123 Saarbrücken 132 pp. Englisch. Nº de ref. del artículo: 9783846541463
Cantidad disponible: 1 disponibles
Librería: AHA-BUCH GmbH, Einbeck, Alemania
Taschenbuch. Condición: Neu. nach der Bestellung gedruckt Neuware - Printed after ordering - Classification algorithms have been widely used in many application domains. Most of these domains deal with massive collection of data and hence demand classification algorithms that scale well with the size of the data sets involved. A classification algorithm is said to be scalable if there is no significant increase in time and space requirements for the algorithm (without compromising the generalization performance) when dealing with an increase in the training set size. Support Vector Machine (SVM) is one of the most celebrated kernel based classification methods used in Machine Learning. An SVM capable of handling large scale classification problems will definitely be an ideal candidate in many real world applications. The training process involved in SVM classifier is usually formulated as a Quadratic Programing (QP) problem. The existing solution strategies for this problem have an associated time and space complexity that is (at least) quadratic in the number of training points. It makes SVM training very expensive. This thesis addresses the scalability of the training algorithms involved in SVM to make it feasible with large training data sets. Nº de ref. del artículo: 9783846541463
Cantidad disponible: 1 disponibles
Librería: preigu, Osnabrück, Alemania
Taschenbuch. Condición: Neu. Efficient Kernel Methods For Large Scale Classification | Scalable methods for training Support Vector Machines | Asharaf S | Taschenbuch | 132 S. | Englisch | 2011 | LAP LAMBERT Academic Publishing | EAN 9783846541463 | Verantwortliche Person für die EU: BoD - Books on Demand, In de Tarpen 42, 22848 Norderstedt, info[at]bod[dot]de | Anbieter: preigu. Nº de ref. del artículo: 106735434
Cantidad disponible: 5 disponibles
Librería: Revaluation Books, Exeter, Reino Unido
Paperback. Condición: Brand New. 132 pages. 8.50x0.39x5.91 inches. In Stock. Nº de ref. del artículo: 384654146X
Cantidad disponible: 1 disponibles