Documente Academic
Documente Profesional
Documente Cultură
data and recognize patterns, used for classification and regression analysis. The original SVM
algorithm was invented by Vladimir Vapnik and the current standard incarnation (soft margin)
was proposed by Corinna Cortes and Vladimir Vapnik[1]. The standard SVM is a non-
probabilistic binary linear classifier, i.e. it predicts, for each given input, which of two possible
classes the input is a member of. Since an SVM is a classifier, then given a set of training
examples, each marked as belonging to one of two categories, an SVM training algorithm builds
a model that predicts whether a new example falls into one category or the other. Intuitively, an
SVM model is a representation of the examples as points in space, mapped so that the examples
of the separate categories are divided by a clear gap that is as wide as possible. New examples
are then mapped into that same space and predicted to belong to a category based on which side
of the gap they fall on.
More formally, a support vector machine constructs a hyperplane or set of hyperplanes in a high
or infinite dimensional space, which can be used for classification, regression or other tasks.
Intuitively, a good separation is achieved by the hyperplane that has the largest distance to the
nearest training data points of any class (so-called functional margin), since in general the larger
the margin the lower the generalization error of the classifier.
Whereas the original problem may be stated in a finite dimensional space, it often happens that
in that space the sets to be discriminated are not linearly separable. For this reason it was
proposed that the original finite dimensional space be mapped into a much higher dimensional
space presumably making the separation easier in that space. SVM schemes use a mapping into a
larger space so that cross products may be computed easily in terms of the variables in the
original space making the computational load reasonable. The cross products in the larger space
are defined in terms of a kernel function K(x,y) which can be selected to suit the problem. The
hyperplanes in the large space are defined as the set of points whose cross product with a vector
in that space is constant. The vectors defining the hyperplanes can be chosen to be linear
combinations with parameters αi of images of feature vectors which occur in the data base. With
this choice of a hyperplane the points x in the feature space which are mapped into the
hyperplane are defined by the relation:
[edit] Motivation
H3 (green) doesn't separate the 2 classes. H1 (blue) does, with a small margin and H2 (red) with
the maximum margin.
Classifying data is a common task in machine learning. Suppose some given data points each
belong to one of two classes, and the goal is to decide which class a new data point will be in. In
the case of support vector machines, a data point is viewed as a p-dimensional vector (a list of p
numbers), and we want to know whether we can separate such points with a p − 1-dimensional
hyperplane. This is called a linear classifier. There are many hyperplanes that might classify the
data. One reasonable choice as the best hyperplane is the one that represents the largest
separation, or margin, between the two classes. So we choose the hyperplane so that the distance
from it to the nearest data point on each side is maximized. If such a hyperplane exists, it is
known as the maximum-margin hyperplane and the linear classifier it defines is known as a
maximum margin classifier.
• Reference(s):
“Support vector machines-based generalized predictive control,” Serdar
Iplikci, INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, Vol.
16, pp. 843-862, 2006
• Reference link(s):
http://ietfec.oxfordjournals.org/cgi/content/abstract/E89-A/10/2787
• Data link(s):
Entered by: Serdar Iplikci <iplikci@pau.edu.tr> - Monday, October 23, 2006 at 18:05:17
(GMT)
Comments:
• Reference(s):
Physica D, Vol. 216, pp. 282-293, 2006
• Reference link(s):
• Data link(s):
Entered by: Serdar Iplikci <iplikci@pau.edu.tr> - Monday, May 29, 2006 at 12:53:56 (GMT)
Comments:
• Reference(s):
"Application of the kernel method to the inverse geosounding problem", Hugo
Hidalgo, Sonia Sosa and E. Gómez-Treviño, Neural Networks, vol. 16, pp. 349-
353, 2003
• Reference link(s):
http://cienciascomp.cicese.mx/recopat/articulos/NeuralNetworks03.pdf
• Data link(s):
Entered by: Hugo Hidalgo <hugo@cicese.mx> - Wednesday, March 22, 2006 at 14:04:25
(MST)
Comments:
• Reference(s):
Goh ATC. Seismic Liquefaction Potential Assessed by Neural Networks. Journal
of Geotechnical Engineering 1994; 120(9): 1467-1480.
Goh ATC. Neural-Network Modeling of CPT Seismic Liquefaction Data. Journal
of Geotechnical Engineering 1996; 122(1): 70-73
• Reference link(s):
Accepted for publication in International Journal for Numerical and Analytical
Methods in Geomechanics.
• Data link(s):
• Reference(s):
1. N. Gilardi, M. Kanevski, M. Maignan and E. Mayoraz. Environmental and
Pollution Spatial Data Classification with Support Vector Machines and
Geostatistics. Workshop W07 “Intelligent techniques for Spatio-Temporal
Data Analysis in Environmental Applications”. ACAI99, Greece, July, 1999. pp.
43-51. www.idiap.ch
2. M Kanevski, N Gilardi, E Mayoraz, M Maignan. Spatial Data Classification
with Support Vector Machines. Geostat 2000 congress. South Africa, April
2000.
3. Kanevski M., Wong P., Canu S. Spatial Data Mapping with Support Vector
Regression and Geostatistics. 7th International Conference on Neural
Information Processing, Taepon, Korea. Nov. 14-18, 2000. Pp. 1307-1311.
4. N GILARDI, Alex GAMMERMAN, Mikhail KANEVSKI, Michel MAIGNAN, Tom
MELLUISH, Craig SAUNDERS, Volodia VOVK. Application des méthodes
d’apprentissage pour l’étude des risques de pollution dans le Lac Léman. 5e
Colloque transfrontalier CLUSE. Risques majeurs: perception, globalisation et
management. Université de Genève, 2000.
5. M. Kanevski. Evaluation of SVM Binary Classification with Nonparametric
Stochastic Simulations. IDIAP Research Report, IDIAP-RR-01-07, 17 p. 2001.
www.idiap.ch
6. M. Kanevski, A. Pozdnukhov, S. Canu, M. Maignan. Advanced Spatial Data
Analysis and Modelling with Support Vector Machines. International Journal on
Fuzzy Systems 2002. p. 606-615.
7. M. Kanevski , A. Pozdnukhov , S. Canu ,M. Maignan , P.M. Wong , S.A.R.
Shibli “Support Vector Machines for Classification and Mapping of Reservoir
Data”. In: “Soft Computing for Reservoir Characterization and Modelling”. P.
Wong, F. Aminzadeh, M. Nikravesh (Eds.). Physica-Verlag, Heidelberg, N.Y.
pp. 531-558, 2002.
8. Kanevski M., Pozdnukhov A., McKenna S., Murray Ch., Maignan M.
Statistical Learning Theory for Spatial Data. In proceedings of GeoENV2002
conference. Barcelona, 2002.
9. M. Kanevski et al. Environmental data mining and modelling based on
machine learning algorithms and geostatistics. Journal of Environmental
Modelling and Software, 2004. vol. 19, pp. 845-855.
10. M. Kanevski, M. Maignan et al. Advanced geostatistical and machine
learning models for spatial data analysis of radioactively contaminated
territories. Journal of Environmental Sciences and Pollution Research, pp.137-
149, 2003.
11. Kanevski M., Maignan M. and Piller G. Advanced analysis and modelling
tools for spatial environmental data. Case study: indoor radon data n
Switzerland. International conference EnviroInfo, 2004.
http://www.enviroinfo2004.org/cdrom/Datas/Kanevski.htm
12. Kanevski M., Maignan M. and Pozdnukhov A. Active Learning of
Environmental Data Using Support Vector Machines. Conference of the
International Association for Mathematical Geology, Toronto 2005.
http://www.iamgconference.com/
13. M. Kanevski, A. Pozdnukhov, M. Tonini, M. Motelica, E. Savelieva, M.
Maignan. Statistical Learning Theory for Geospatial Data. Case study: Aral
Sea. 14th European colloquium on Theoretical and Quantitative Geography.
Portugal, September 2005.
14. Pozdnukhov A., Kanevski M. Monitoring network optimisation using
support vector machines. In: Geostatistics for Environmental applications.
(Renard Ph., Demougeot-Renard H and Froidevaux, Eds.). Springer, 2005. pp.
39-50.
15. Pozdnukhov A. and Kanevski M. Monitoring Network Optimisation for
Spatial Data Classification Using Support Vector Machines. (2006).
International Journal of Environment and Pollution. Vol.28. 20 pp.
• Reference link(s):
www.unil.ch/igar
www.idiap.ch
• Data link(s):
• Reference(s):
Profile based direct kernels for remote homology detection and fold
recognition by Huzefa Rangwala and George Karypis (Bioinformatics 2005)
• Reference link(s):
http://bioinformatics.oxfordjournals.org/cgi/content/abstract/bti687v1
• Data link(s):
http://bioinfo.cs.umn.edu/supplements/remote-homology/
• Reference(s):
Dacheng Tao, Xiaoou Tang, Xuelong Li, and Xindong Wu, Asymmetric
Bagging and Random Subspacing for Support Vector Machines-based
Relevance Feedback in Image Retrieval, IEEE Transactions on Pattern
Analysis and Machine Intelligence, accepted, to appear.
• Reference link(s):
• Data link(s):
Entered by: Dacheng Tao <Dacheng Tao> - Tuesday, October 11, 2005 at 19:03:18 (GMT)
Comments:
DATA Classification ursing SSVM
Smoothing methods, extensively used for solving important mathematical
programming problems and applications, are applied here to generate and solve an
unconstrained smooth reformulation of the support vector machine for data or
pattern classification using a completely arbitrary kernel. The basic SVM is
reformulated in to smooth support vector machine (SSVM) which possesses the
mathematical property of strong convexity, which makes the basic SVM
classification , a minimization problem. In this work Newton-Armijo algorithm is used
for solving the SSVM Unconstraint optimization problem. On larger datasets SSVM is
faster than SVM light . SSVM can also generate a highly nonlinear separating
surface such as a checkerboard.
• Reference(s):
• Reference link(s):
• Data link(s):
• Reference(s):
• Reference link(s):
http://www.dtreg.com/svm.htm
• Data link(s):
• Reference(s):
• Reference link(s):
http://www.dtreg.com/svm.htm
• Data link(s):
Entered by: Phil Sherrod <phil.sherrod@sandh.com> - Friday, August 26, 2005 at 20:09:46
(GMT)
Comments: DTREG supports Linear, Polynomial, Sigmoid and Radial Basis kernel functions. It
can handle problems with millions of data rows and hundreds of variables.
• Reference(s):
J. Ghent and J. McDonald, "Facial Expression Classification using a One-
Against-All Support Vector Machine", proceedings of the Irish Machine Vision
and Image Processing Conference, Aug 2005.
• Reference link(s):
• Data link(s):
Entered by: John Ghent <jghent@cs.may.ie> - Tuesday, August 09, 2005 at 10:14:08 (GMT)
Comments:
• Reference(s):
C. Cortes and V.N. Vapnik, Support vector networks, Machine Learning 20
(1995), pp. 273–297.
S. Dey, Free over fall in circular channels with flat base: a method of open
channel flow measurement, Flow Meas. Instrum. 13 (2002), pp. 209–221.
S. Dey, Free over fall in open channels: state-of-the-art review, Flow Meas.
Instrum. 13 (2002), pp. 247–264.
Y.B. Dibike, S. Velickov, D.P. Solomatine and M.B. Abbott, Model induction
with support vector machines: Introduction and applications, J. Comput. Civil
Eng. 15 (2001), pp. 208–216.
D. Leunberger, Linear and Nonlinear Programming, Addison-Wesley (1984).
H. Rouse, Discharge characteristics of the free overfall, Civil Engineering,
ASCE 6 (1936) (4), pp. 257–260.
R.V. Raikar, D. Nagesh Kumar and S. Dey, End depth computation in inverted
semi circular channels using ANNs, Flow Meas. Instrum. 15 (2004), pp. 285–
293.
A.J. Smola, Regression estimation with support vector learning machines,
Master’s Thesis, Technische Universität München, Germany, 1996.
M. Sterling and D.W. Knight, The free overfall as a flow measuring device in a
circular channel, Water and Maritime Engineering Proceedings of Institution
of Civil Engineers London 148 (December) (2001), pp. 235–243.
V.N. Vapnik, Statistical Learning Theory, John Wiley and Sons, New York
(1998).
• Reference link(s):
http://www.sciencedirect.com/science/journal/09555986
• Data link(s):
Entered by: mahesh pal <mpce_pal@yahoo.co.uk> - Monday, August 01, 2005 at 10:20:34
(GMT)
Comments:
• Reference(s):
Dror G., Sorek R. and Shamir S.
Accurate identification of alternatively spliced exons using Support Vector
Machine
Bioinformatics. 2005 Apr 1;21(7):897-901.
Epub 2004 Nov 5.
• Reference link(s):
http://www2.mta.ac.il/~gideon/nns_pub.html
• Data link(s):
Entered by: Gideon Dror <gideon@mta.ac.il> - Monday, June 20, 2005 at 11:55:09 (GMT)
Comments: 2 class, 243 positive , 1753 negative instances. total 228 features gaussian kernel.
Baseline systems: neural networks and Naive Bayes. SVM outperformed them in terms of area
under ROC curve, but most inportantly, in its ability to get very high true positives rate (50%)
for very low false positives rate (0.5%). This performance would enable effective scan of exon
databases in search for novel alternatively spliced exons, in the human or other genomes.
• Reference(s):
1.Support Vector Machines for Texture Classification
Kwang In Kim, Keechul Jung, Se Hyun Park, and
Hang Joon Kim,IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE
INTELLIGENCE, VOL. 24, NO. 11, NOVEMBER 2002,
2.“An introduction to Support Vector Machines and other kernel-based
learning
methods” by Nello Cristianini & John Shawe-Taylor. (http://www.support
http://www.supportvector.
net/)
• Reference link(s):
• Data link(s):
• Reference(s):
• Reference link(s):
• Data link(s):
• Reference(s):
• Reference link(s):
• Data link(s):
Entered by: Duong DInh DUng <dungngtq8@yahoo.com> - Thursday, March 24, 2005 at
06:03:04 (GMT)
Comments:
• Reference(s):
• Reference link(s):
• Data link(s):
Entered by: Suhaimi Abd Latif <suhaimie@iiu.edu.my> - Wednesday, January 19, 2005 at
06:02:27 (GMT)
Comments:
Image Clustering
Clustering is an important task for image compression. This clustering can be done
by SVM efficiently.
• Reference(s):
• Reference link(s):
• Data link(s):
Entered by: Ahmed Yousuf Saber <saber_uap@yahoo.com> - Wednesday, January 19, 2005 at
02:16:09 (GMT)
Comments:
• Reference(s):
Bomhardt, C. (2004): NewsRec, a SVM-driven Personal Recommendation
System for News Websites
In: Web Intelligence, IEEE/WIC/ACM International Conference on (WI'04)
Keywords: Personal Recommendation, Support-Vector-Machine,
Personalization, Text Classification
• Reference link(s):
http://csdl.computer.org/comp/proceedings/wi/2004/2100/00/2100toc.htm
• Data link(s):
Equbits Foresight
Equbits Foresight is a SVM based predictive modeling application designed for HTS
and ADME-Tox Chemists.
• Reference(s):
• Reference link(s):
www.equbits.com
• Data link(s):
Entered by: Ravi Mallela <ravi@equbits.com> - Saturday, October 09, 2004 at 15:37:20
(GMT)
Comments:
• Reference(s):
X.DONG......
ELECTRONICS LETTERS ,VOL.37,PP.527-529,(2001)
C.MA.RANDOLPH ....
IEEE INT.CONFERENCE ON ACOUSTICS,SPEECH,AND SIGNAL PROCESSING
VOL.1,PP.381-384,(2001)
V.WAN...
IEEE WORKSHOP ON NEURAL NETWORK FOR SIGNAL PROCESSING X,VOL.2,
(2000)
• Reference link(s):
• Data link(s):
STUDENT IN AI
SPEAKER /SPEECH RECOGNITION UTTERANCE VERIFICATION FOR SPEECH
RECOGNITION SVM ARE USED TO ACCEPT KEYWORD OR REJECT NON-KEYWORD
FOR SPEECH RECOGNITION SPEAKER VERIFICATION /RCOGNITION POLYVAR
TELEPHONE DATABASE IS USED NEW METHOD FOR NORMALIZING POLYNORMIAL
KERNEL TO USE WITH SVM YOHO DATABASE ,TEXT INDEPENDENT , BEST
EER=0.34% COMBINED GAUSSIAN MIXTURE MODEL IN SVM OUTPUTS TEXT
INDEPENDENT SPEAKER VERIFICATION BEST EER=1.56%
• Reference(s):
X.DONG......
ELECTRONICS LETTERS ,VOL.37,PP.527-529,(2001)
C.MA.RANDOLPH ....
IEEE INT.CONFERENCE ON ACOUSTICS,SPEECH,AND SIGNAL PROCESSING
VOL.1,PP.381-384,(2001)
V.WAN...
IEEE WORKSHOP ON NEURAL NETWORK FOR SIGNAL PROCESSING X,VOL.2,
(2000)
• Reference link(s):
• Data link(s):
• Reference(s):
[1] ÂÀ½ð»¢µÈ.»ìãçʱ¼äÐòÁзÖÎö¼°ÆäÓ¦ÓÃ[M]. Î人: Î人´óѧ³ö°æÉç,2001
[2] Stefania Tronci, Massimiliano Giona, Roberto Baratti, ¡°Reconstruction of
chaotic time series by neural models: a case study,¡± Neurocomputing, vol.
55, pp. 581-591, 2003.
[3] ºØÌ«¸Ù,Ö£³çÑ«. »ìãçÐòÁеķÇÏßÐÔÔ¤²â[J].×ÔÈ»ÔÓÖ¾, 19(1): 10-13,
2001.
[4] À÷, ÍõÕýÅ·.
»ùÓÚRBFÍøÂçµÄ»ìãçʱ¼äÐòÁеĽ¨Ä£Óë¶à²½Ô¤²â[J].ϵͳ¹¤³ÌÓëµç×Ó¼¼Êõ,
24(6): 81-83, 2002.
[5] ½ªÌÎ. º½¿Õ·¢¶¯»ú´Õñ/ʧËÙÔ¤¹ÀÄ£Ðͺ͹ÊÕϼì²âÑо¿[D], ²©Ê¿Ñ§Î»ÂÛÎÄ,
Î÷°²:¿Õ¾ü¹¤³Ì´óѧ¹¤³ÌѧԺ,2002.
[6] Íõº£Ñà, Ê¢ÕÑå«. »ìãçʱ¼äÐòÁÐÏà¿Õ¼äÖع¹²ÎÊýµÄÑ¡È¡·½·¨[J].¶«ÄÏ
´óѧѧ±¨, 30(5):113-117, 2000.
[7] L.-Y. Cao, ¡°Practical method for determining minimum embedding
dimension of a scalar time series,¡± Physica D, vol. 110, pp. 43-52, 1997.
[8] Eckmann J.P, Kamphorst S.O, ¡°Lyapunov exponent from time series,¡±
Phys. Rev. A, vol. 34, no. 6, pp.4971~4979, Dec. 1986.
[9] Oiwa N.N, Fiedler-Ferrara N, ¡°A fast algorithm for estimating lyapunov
exponents from time series,¡± Physics Letter A, vol.246, pp.117-121, Sep.
1998.
[10] Fabio Sattin, ¡°Lyap: A FORTRAN 90 program to compute the Lyapunov
exponents of a dynamical system from a time series,¡± Computer Physics
Communications, vol.107, pp.253-257. 1997.
[11] K.R.M¨¹ler, A.J. Smola, G.Rätsch, ¡°Predicting time series with support
vector machines,¡± in Proceeding of ICANN 97¡¯, Berlin: Springer LNCS, vol.
1327, pp. 999-1004. 1997.
[12] B.-J. Chen, ¡°Load forecasting using Support vector machines: A study on
EUNITE Competition 2001,¡±unpublished.
[13] L.-J.Cao, Q.-M. Gu, ¡°Dynamic support vector machines for non-stationary
time series forecasting,¡± Intelligent Data Analysis. vol. 6, no. 1, pp. 67-83,
2002.
[14] F.E.H.Tay, L.-J. Cao, ¡°Applications of support vector machines in
financial forecasting,¡± Omega, vol. 9, no. 4, pp.309-317, Aug. 2001.
[15] K.W.Lau, Q.-H. Wu, ¡°Local prediction of chaotic time series based on
Gaussian processes,¡± in Proceeding of the 2002 IEEE International
Conference on Control Applications, Glasgow, Scotland, U.K, pp. 1309-1313,
Sep. 18-20 2002.
[16] Sayan Mukherjee, Edgar Osuna, Frederico Girosi, ¡° Nonlinear prediction
of chaotic time series using support vector machines,¡± in Proc.of IEEE NNSP
97, Amelia Island, FL, Sep. 1997.
• Reference link(s):
In press
Proceeding of WCICA 2004
• Data link(s):
• Reference(s):
V. Kecman, Learning And Soft Computing: Support Vector Machines, Neural
Networks, And Fuzzy Logic Models, The MIT press, Cambridge,
Massachusetts, London, England.
• Reference link(s):
• Data link(s):
Entered by: Lelitha Vanajakshi <lelitha@yahoo.com> - Friday, January 30, 2004 at 17:39:08
(GMT)
Comments: When the training data was less SVM outperformed ANN, when enough data was
available both performed more or less same.
none
none
• Reference(s):
• Reference link(s):
• Data link(s):
Entered by: leechs <leechs@sohu.com> - Sunday, January 25, 2004 at 13:44:16 (GMT)
Comments:
svm learning
svm face learning
• Reference(s):
• Reference link(s):
• Data link(s):
• Reference(s):
1. Kim, H. and H. Park, "Prediction of protein relative solvent accessibility with
support vector machines and long-range interaction 3D local descriptor",
Proteins:structure, function, and genetics, to appear. (pdf download)
2. Kim, H. and H. Park, "Protein secondary structure prediction by support
vector machines and position-specific scoring matrices",
Protein Engineering, to appear. (pdf download)
• Reference link(s):
http://www.cs.umn.edu/~hpark/papers/surface.pdf
http://www.cs.umn.edu/~hpark/papers/protein2.pdf
• Data link(s):
Entered by: Dr. Haesun Park <hpark@cs.umn.edu> - Friday, July 11, 2003 at 18:59:18 (GMT)
Comments:
Support vector classifiers for land cover classification
SVM wasused to classifiy different land covers using remote sensing data. Results
from this study suggests that Multi-class support vector machine perform well in
comparison with neural network and decision tree classifiers.
• Reference(s):
Mahesh Pal recently finished his PhD form the university of Nottingham, UK
and presently working as a lecturer in department of civil enginnering NIT
kurukshetra, haryana, India.
• Reference link(s):
http://www.gisdevelopment.net/technology/rs/pdf/23.pdf
• Data link(s):
Entered by: Mahesh Pal <mpce_pal@yahoo.co.uk> - Wednesday, May 21, 2003 at 07:17:46
(GMT)
Comments:
Intrusion Detection
Intrusion Detection Systems (IDSs) have become more widely used to protect
computer networks. However, it is difficult to build highly effective IDS since some
of the pattern recognition problems involved are intractable. In this paper, we
propose the use of Support Vector Machines (SVMs) for intrusion detection and
analyze its performance. We conduct experiments using a large set of DARPA-
provided intrusion data. Two groups of SVMs are built to perform, respectively,
binary classifications (normal pattern vs. attack pattern) and five-class
classifications (normal pattern, and four classes of attack patterns). Detailed
analysis is provided on the classification accuracy and time performance (regarding
both learning time and running time). Performance comparison between SVMs and
neural networks (that are built using the same training and testing data) is also
given. Based on the simulation results, we argue that SVMs are superior to neural
networks in several important aspects of IDS. Our results therefore indicate that
SVMs are superior candidates to be used as the learning machines for IDSs.
• Reference(s):
Srinivas Mukkamala joined the Computer Science graduate program of New
Mexico Tech in 2000 and is currently a Ph.D. student. He received his B.E.
degree in computer science and engineering in 1999 from University of
Madras. His interests are information assurance, information hiding, artificial
intelligence, soft computing techniques for computer security.
• Data link(s):
http://kdd.ics.uci.edu/
During the last years the task of on-line handwriting recognition has gained an immense
importance in all-day applications, mainly due to the increasing popularity of the personal digital
assistant (pda). Currently a next generation of ``smart phones'' and tablet-style PCs, which also
rely on handwriting input, is further targeting the consumer market. However, in the majority of
these devices the handwriting input method is still not satisfying. In current pdas people still use
input methods, which abstract from the natural writing style, e.g. in the widespread Graffiti.
Thus there is demand for a handwriting recognition system which is accurate, efficient and
which can deal with the natural handwriting of a wide range of different writers.
• Reference(s):
Data URL:
UNIPEN Train-R01/V07
• Reference link(s):
• Data link(s):
UNIPEN Train-R01/V07
Usual SVM kernels are designed to deal with data of fixed dimension. However, on-line
handwriting data is not of a fixed dimension, but of a variable-length sequential form. In this
respect SVMs cannot be applied to HWR straightforwardly.
We have addressed this issue by developing an appropriate SVM kernel for sequential data, the
Gaussian dynamic time warping (GDTW) kernel. The basic idea of the GDTW kernel is, that
instead of the squared Euclidean distance in the usual Gaussian kernel it uses the dynamic time
warping distance. In addition to on-line handwriting recognition the GDTW kernel can be
straightforwardly applied to all classification problems, where DTW gives a reasonable distance
measure, e.g. speech recognition or genome processing.
Experiments have shown superior recognition rate in comparison to an HMM-based classifier for
relative small training sets (~ 6000) and comparable rates for larger training sets.
During the last years the task of on-line handwriting recognition has gained an immense
importance in all-day applications, mainly due to the increasing popularity of the personal digital
assistant (pda). Currently a next generation of ``smart phones'' and tablet-style PCs, which also
rely on handwriting input, is further targeting the consumer market. However, in the majority of
these devices the handwriting input method is still not satisfying. In current pdas people still use
input methods, which abstract from the natural writing style, e.g. in the widespread Graffiti.
Thus there is demand for a handwriting recognition system which is accurate, efficient and
which can deal with the natural handwriting of a wide range of different writers.
• Reference(s):
<A
href="http://lmb.informatik.uni-freiburg.de/people/bahlmann/">Claus
Bahlmann, <A
href="http://lmb.informatik.uni-freiburg.de/people/haasdonk/">Bernard
Haasdonk and <A
href="http://lmb.informatik.uni-freiburg.de/people/burkhardt/">Hans
Burkhardt. <A
href="http://lmb.informatik.uni-
freiburg.de/people/bahlmann/science.en.html#Anchor_ba_ha_bu_iwfh02">On-line
Handwriting Recognition using Support Vector Machines - A kernel
approach. In Int. Workshop on Frontiers in Handwriting
Recognition (IWFHR) 2002, Niagara-on-the-Lake, August 2002.
<A
href="http://lmb.informatik.uni-freiburg.de/people/bahlmann/frog.en.html">Description
of the frog on hand recognition system
• Reference link(s):
• Data link(s):
UNIPEN Train-R01/V07
Usual SVM kernels are designed to deal with data of fixed dimension. However, on-line
handwriting data is not of a fixed dimension, but of a variable-length sequential form. In this
respect SVMs cannot be applied to HWR straightforwardly.
We have addressed this issue by developing an appropriate SVM kernel for sequential data, the
Gaussian dynamic time warping (GDTW) kernel. The basic idea of the GDTW kernel is, that
instead of the squared Euclidean distance in the usual Gaussian kernel it uses the dynamic time
warping distance. In addition to on-line handwriting recognition the GDTW kernel can be
straightforwardly applied to all classification problems, where DTW gives a reasonable distance
measure, e.g. speech recognition or genome processing.
Experiments have shown superior recognition rate in comparison to an HMM-based classifier for
relative small training sets (~ 6000) and comparable rates for larger training sets.
forecast
forecast stock
• Reference(s):
• Reference link(s):
• Data link(s):
• Reference(s):
Detecting Hidden Messages Using Higher-Order Statistics and Support Vector
Machines
• Reference link(s):
http://www.cs.dartmouth.edu/~farid/publications/ih02.html
• Data link(s):
http://www.cs.dartmouth.edu/~farid/publications/ih02.html
Entered by: Siwei Lyu <lsw@cs.dartmouth.edu> - Thursday, August 22, 2002 at 15:58:54
(GMT)
Comments: 2 classes 3600 training examples, over 18,000 testing samples 1100 SVs RBF
kernel LibSVM
• Reference(s):
• Reference link(s):
http://www.cs.dartmouth.edu/~farid/publications/ih02.html
• Data link(s):
http://www.cs.dartmouth.edu/~farid/publications/ih02.html
Entered by: Siwei Lyu <lsw@cs.dartmouth.edu> - Thursday, August 22, 2002 at 15:57:24
(GMT)
Comments: 2 classes 3600 training examples, over 18,000 testing samples 1100 SVs RBF
kernel LibSVM
• Reference(s):
• Reference link(s):
members.aol.com/awareai
• Data link(s):
Entered by: Michael Bickel <awareai@aol.com> - Tuesday, July 23, 2002 at 01:11:11 (GMT)
Comments:
• Reference(s):
Yuh-Jye Lee, O. L. Mangasarian and W. H. Wolberg: ¡§Survival-Time
Classification of Breast Cancer Patients, Data Mining Institute Technical
Report 01-03, March 2001.
• Reference link(s):
• Data link(s):
WPBCC: Wisconsin Prognostic Breast Cancer Chemotherapy Database.
ftp://ftp.cs.wisc.edu/math-prog/cpo-dataset/machine-learn/WPBCC/
Entered by: Yuh-Jye Lee <yjlee@cs.ccu.edu.tw> - Wednesday, October 24, 2001 at 19:38:50
(MDT)
Comments:
• Reference(s):
• Reference link(s):
• Data link(s):
Entered by: Robin Willis <rew198@soton.ac.uk> - Friday, May 04, 2001 at 08:31:41 (PDT)
Comments:
Image classification
Classification of natural images from the Corel database using a SVM and a color
histogram as input feature
• Reference(s):
• Reference link(s):
www.ens-lyon.fr/~ochapell/tnn99.ps.gz
• Data link(s):
• Reference(s):
Classifying LEP Data with Support Vector Algorithms
P. Vannerem, K.-R. Müller, B. Schölkopf, A. Smola, S. Söldner-Rembold
submitted to Proceedings of AIHENP'99.
• Reference link(s):
http://wwwrunge.physik.uni-freiburg.de/preprints/EHEP9901.ps
ftp://ftp.physics.uch.gr/aihenp99/Vannerem/
• Data link(s):
Entered by: Philippe Vannerem <philippe.vannerem@cern.ch> - Tuesday, October 19, 1999 at
16:17:56 (PDT)
Comments: We saw only small differences in performance between NNs and SVMs.
Object Detection
The problem of object detection is to differentiate a certain class of objects (the A
class) from all other patterns and objects (the not-A class). This is contrasted with
object recognition where the problem is to be able to differentiate between
elements of the same class.
• Reference(s):
Constantine Papageorgiou, Tomaso Poggio
A Pattern Classification Approach to Dynamical Object Detection
Proceedings of ICCV, 1999, pp. 1223-1228
Anuj Mohan
Object Detection in Images by Components
CBCL Paper #178/AI Memo #1664, Massachusetts Institute of Technology
• Reference link(s):
http://www.ai.mit.edu/projects/cbcl/publications/ps/dyn-obj-iccv99.ps.gz
ftp://publications.ai.mit.edu/ai-publications/1500-199/AIM-1664.ps
www.ai.mit.edu/projects/cbcl/publications/ps/ped-det-iv98.ps.gz
www.ai.mit.edu/projects/cbcl/publications/ps/gen-obj-det-iccv98.ps.gz
• Data link(s):
• Reference(s):
M.Rychetsky, S.Ortmann, M.Glesner: Construction of a Support Vector
Machine with Local Experts. Workshop on Support Vector Machines at the
International Joint Conference on Artificial Intelligence (IJCAI 99), August,
1999, Stockholm, Sweden
M.Rychetsky, S.Ortmann, M.Glesner: Support Vector Approaches for Engine
Knock Detection. International Joint Conference on Neural Networks (IJCNN
99), July, 1999, Washington, USA
• Reference link(s):
• Data link(s):
• Reference(s):
A. Zien and G. Rätsch and S. Mika and B. Schölkopf and C. Lemmen and A.
Smola and T. Lengauer and K.-R. Müller
Engineering Support Vector Machine Kernels That Recognize Translation
Initiation Sites
German Conference on Bioinformatics 1999
• Reference link(s):
http://www.bioinfo.de/isb/gcb99/talks/zien/
• Data link(s):
The data sets we used were kindly supplied by Pedersen and Nielsen
(Center for Biological Sequence Analysis, Denmark; http://www.cbs.dtu.dk/).
• Reference(s):
A discriminative framework for detecting remote protein homologies.
Tommi Jaakkola, Mark Diekhans, and David Haussler
• Reference link(s):
http://www.cse.ucsc.edu/research/compbio/discriminative/Jaakola2-1998.ps
• Data link(s):
Dataset (12Mb compressed)
• Reference(s):
o Support Vector Regression Machines.
Drucker, H.; Burges, C.; Kaufman, L.; Smola, A.; Vapnik, V. 1997.
In: M. Mozer, M. Jordan, and T. Petsche (eds.):
Neural Information Processing Systems, Vol. 9. MIT Press, Cambridge,
MA,
1997.
o Support Vector Regression with ANOVA Decomposition Kernels.
Mark O. Stitson, Alex Gammerman, Vladimir Vapnik, Volodya Vovk,
Chris Watkins, and Jason Weston
in Advances in Kernel Methods, B. Schölkopf, C.J.C. Burges, and A.J.
Smola Eds.
Pages 285-291, MIT Press, 1999. ISBN 0-262-19416-3.
• Reference link(s):
Drucker-97,
• Data link(s):
ftp://ftp.ics.uci.edu/pub/machine-learning-databases/housing
• Reference(s):
• Reference link(s):
Blanz-96, Osuna-97, Schölkopf-thesis-97, Schölkopf-NIPS-97, Pontil-97,
Zhao-98, Friess-98, Roobaert-99a, Roobaert-99b
• Data link(s):
Chair data set
Sonar data
Text Categorization
Text categorization is the assignment of natural language texts to one or more
predefined categories based on their content. Applications include: assigning
subject categories to documents to support text retrieval, routing, and filtering;
email or files sorting into folder hierarchies; web page sorting into search engine
categories.
• Reference(s):
o Text Categorization with Support Vector Machines: Learning with Many
Relevant Features.
T. Joachims,
European Conference on Machine Learning (ECML),
1998.
o Inductive Learning Algorithms and Representations for Text
Categorization,
S. Dumais, J. Platt, D. Heckerman, M. Sahami,
7th International Conference on Information and Knowledge
Management,
1998.
o Support Vector Machines for Spam Categorization. H. Drucker, with D.
Wu and V. Vapnik. IEEE Trans. on Neural Networks , vol 10, number 5,
pp. 1048-1054. 1999.
o Transductive Inference for Text Classification using Support Vector
Machines.
Thorsten Joachims.
International Conference on Machine Learning (ICML),
1999.
• Reference link(s):
Joachims-98 Postcript, Joachims-98 PDF
Dumais et al 98
Drucker et al 98
Joachins-99 Postcript
Joachims-99 PDF
• Data link(s):
Reuters-21578
• Reference(s):
Predicting Time Series with Support Vector Machines.
K.-R. Müller, A. Smola, G. Rätsch, B. Schölkopf, J. Kohlmorgen, V. Vapnik.
Proceedings ICANN'97, p.999.
Springer Lecture Notes in Computer Science, 1997
• Reference link(s):
Müller et al
Mukherjee et al
• Data link(s):
Synthetic data used: Mackey-Glass, Ikewda Map and Lorenz, and
Santa Fe competition Data Set D
Entered by: Isabelle Guyon <isabelle@clopinet.com> - Thursday, September 16,
1999 at 14:54:32 (PDT)
Comments: Müller et al report excellent performance of SVM. They set a new
record on the Santa Fe competition data set D, 37% better than the winning
approach during the competition. Mattera et al report that SVM are effective for
such tasks and that their main advantage is the possibility of trading off the
required accuracy with the number of Support Vectors.
• Reference(s):
Support Vector Machine Classification of Microarray Gene Expression Data
M. Brown, W. Grundy, D. Lin, N. Cristianini C. Sugnet, M. Ares Jr., D. Haussler
University of California, Santa Cruz,
technical report UCSC-CRL-99-09.
• Reference link(s):
http://www.cse.ucsc.edu/research/compbio/genex/genex.tech.html
• Data link(s):
http://www.cse.ucsc.edu/research/compbio/genex/
• Reference(s):
o An training algorithm for optimal margin classifiers.
B. Boser, I. Guyon, and V. Vapnik.
In Fifth Annual Workshop on Computational Learning Theory, pages
144--152, Pittsburgh, ACM.
1992.
o Writer adaptation for on-line handwritten character recognition.
N. Matic , I. Guyon, J. Denker, and V. Vapnik
In Second International Conference on Pattern Recognition and
Document Analysis , pages 187--191, Tsukuba, Japan, IEEE Computer
Society Press,
1993.
o Support Vector Networks.
C. Cortes and V. Vapnik,
Machine Learning, 20:273-297,
1995.
o Learning algorithms for classification: A comparison on handwritten
digit recognition.
Y. Le Cun, L.D. Jackel, L. Bottou, C. Cortes, J. Denker, H. Drucker, I.
Guyon, U.A. Muller, E.
Sackinger, P. Simard, and V. Vapnik.
In J.H. Kwon and S. Cho, editors, Neural Networks: The Statistical
Mechanics Perspective, pages
261--276. World Scientific
1995.
o Discovering informative patterns and data cleaning.
I. Guyon, N. Matic , and V. Vapnik,
In U.M. Fayyad, G. Piatetsky-Shapiro, P. Smyth, and R. Uthurusamy,
editors, Advances in Knowledge Discovery and Data Mining, pages
181--203. MIT Press.
1996.
o Incorporating Invariances in Support Vector Learning Machines.
B. Schölkopf, C. Burges, and V. Vapnik,
In: C. von der Malsburg, W. von Seelen, J. C. Vorbr|ggen, and B.
Sendhoff (eds.): Artificial Neural Networks - ICANN'96. Springer Lecture
Notes in Computer Science Vol. 1112, Berlin,47-52
1996.
o Prior Knowledge in Support Vector Kernels
B. Schölkopf, P. Simard, A. Smola, and V. Vapnik,
NIPS'97
1997.
o Pairwise Classification and Support Vector Machines
Ulrich H.-G. Kressel
in Advances in Kernel Methods, B. Schölkopf , C.J.C. Burges, and A.J.
Smola Eds.
Pages 255-268, MIT Press, 1999. ISBN 0-262-19416-3.
o The kernel adatron algorithm: a fast and simple learning procedure for
support vector machines.
T.-T. Friess, N. Cristianini, C. Campbell.
15th Intl. Conf. Machine Learning, Morgan Kaufman Publishers.
1998.
• Reference link(s):
Boser-92 , Matic-93, Guyon-96, Schölkopf-96, Schölkopf-98
• Data link(s):
http://www.clopinet.com/isabelle/Projects/LITTLE1200/
http://www.research.att.com/~yann/ocr/mnist/
• Reference(s):
O. L. Mangasarian, W. Nick Street and W. H. Wolberg: ``Breast cancer
diagnosis and prognosis via linear programming", Operations Research,
43(4), July-August 1995, 570-577.
• Reference link(s):
ftp://ftp.cs.wisc.edu/math-prog/tech-reports/94-10.ps
ftp://ftp.cs.wisc.edu/math-prog/tech-reports/95-21.ps
ftp://ftp.cs.wisc.edu/math-prog/tech-reports/96-03.ps
http://svm.first.gmd.de/papers/FriCriCam98.ps.gz
• Data link(s):
WDBC: Wisconsin Diagnostic Breast Cancer Database
BC: Wisconsin Prognostic Breast Cancer Database
• Reference(s):
On Support Vector Decision Trees for Database Marketing K.P. Bennett, D.
Wu, and L. Auslender Report No. 98-100, Rensselaer Polytechnic Institute,
Troy, NY, 1998.
• Reference link(s):
http://www.rpi.edu/~bennek/mr98100.ps
Data link(s):
Data is proprietary. The ABCs of TCP/IP
• By Mark Sportack.
• Sample Chapter is provided courtesy of Cisco Press.
• Date: Mar 25, 2005.
Chapter Information
Contents
1. What Is TCP/IP?
2. The Abilities and Limitations of TCP/IP
3. Chapter Summary
4. Chapter Review Questions
Chapter Description
Learn the very basics of TCP/IP in this sample chapter by Mark Sportack. Find out
how TCP/IP interacts with the internet, as well as with other communication
protocols.
TCP/IP First-Step
Figure 1-1 shows a logical model. Logical models can be confusing due to their nature, but the
best way to understand is to think of them as a necessary sequence of events. In this particular
model, if you want to communicate across the Internet, you start with your application. Say you
are sending a file directly to a friend across the Internet. The first thing you do is pick an
application that lets you do that.
Fortunately, TCP/IP contains just such an application: It's called the File Transfer Protocol
(FTP). You supply FTP with the necessary information, which is basically your friend's
computer address and the file you want to send. FTP needs to engage the operating system to
access your computer's physical resources including memory, CPU, and the peripheral device
that connects your computer to the physical network.
The preceding example shows how TCP/IP can blur the traditional distinctions between types of
software. Even though TCP/IP is just a communications protocol, it also contains numerous
utilities or applications, such as FTP. It's part communications protocol and part application
software.
On the receiving end, your friend's computer performs the same process, only in reverse. The file
comes through the network, where your operating system uses whatever physical resources it
must and ensures that TCP/IP gets the file via its own copy of FTP. Figure 1-2 illustrates this
process.
In its role as intermediary between software and the physical network, TCP/IP provides the rules
and mechanisms that enable two or more different machines—even if they are very different
machines—to communicate and share information with each other.
TCP/IP's mechanisms are designed to enable communications between any two or more
machines connected to a network regardless of the device type, who made them, or their physical
proximity to each other. Bridging the gap between dissimilar devices is a phenomenally powerful
concept: You have created a mechanism that brings everybody together regardless of where they
are or what technologies they use.
As shown in the preceding FTP example, the boundaries between operating system, application
software, and communications protocols are blurred quite a bit by TCP/IP. When installed,
TCP/IP functions as an extension of the operating system. So seamless is this integration that
many people aren't even aware that they are using TCP/IP to communicate!
Even more confusion is added when you remember that TCP/IP contains quite a few simple
applications that enable a user to directly do simple tasks like push files through a network or log
on to remote computers. Part IV, "User Services—Making the Most Use of Your Network,"
looks more closely at those utilities. Users who are familiar with such utilities might think of
TCP/IP as an application. They'd have justification for thinking that way! However, these
utilities are just one small piece of the comprehensive set of protocols that you know and love
and call TCP/IP.
Economically, too, TCP/IP has been a boon. The concept of an open mechanism that is not
owned or controlled by any one commercial organization has opened doors few people could
have imagined possible. For example, TCP/IP has become widely accepted around the world
simply because it becomes the means by which different manufacturers' products can
interoperate with each other. More importantly, TCP/IP broke new ground by pioneering a new
method for developing new technologies that anyone could manufacture. Now, entities trying to
develop a new communications-based product or technology can simply start with TCP/IP's
openly published specifications and focus on adding value based on that preexisting set of
capabilities. More importantly, developing products around TCP/IP brings a tremendous base of
users who could easily integrate any new products based on that protocol suite.
By now you're probably anxious to get into what TCP/IP can do for you. Some of the
applications that rely on TCP/IP (such as web surfing and e-mail) have been discussed, but you
have seen how one set of mechanisms can so successfully support a rich and diverse array of
applications.
The answer lies in determining the basic underlying capabilities any application requires to
communicate. At the risk of sounding like a late-night talk show host, TCP/IP's top five critical
functions follow:
TCP/IP can, of course, do many other things. Thus, this list is not comprehensive. Instead, these
five things form the basis for communicating. Other features, including those that are part of
TCP/IP and those commercially developed around TCP/IP, build upon this foundation.
What It Can't Do
TCP/IP is not a proverbial silver bullet. It offers tremendous capabilities, but it does have its
limitations. Those limitations have more to do with the intended scope of the protocol suite than
weaknesses. Please consider all the "can't dos" in this section as being by design rather than
indicative of a flawed implementation.
First of all, TCP/IP is not an operating system (OS). An OS is the software that runs on a
computer and creates the environment in which other applications can run. The various Windows
platforms are examples of operating systems. When installed, TCP/IP becomes an important
extension of your computer's operating system, but it does not replace it.
There was a time when some software vendors billed their operating system software as a
Network Operating System (NOS), but that term is misleading. It indicates a computer OS that
supports a robust ability to communicate. One critical component of that ability was the
inclusion of communications protocols, such as TCP/IP. However, bundling a set of
communications protocols with an operating system doesn't make those protocols the operating
system! As networking ability has become a more integral part of both operating systems and
application software, this distinction has fallen by the wayside; consumers began expecting their
operating systems to offer a rich networking capability.
TCP/IP also isn't a network! Installing and configuring TCP/IP on your workstation or home
computer does not, in and of itself, mean you can use that computer to communicate with other
computers or devices. TCP/IP, by itself, is incapable of transmitting or carrying data. Those
functions remain in the purview of a physical network such as an Ethernet or a wide-area
network (WAN). TCP/IP prepares data received from applications for transmission and provides
all the logical mechanisms needed to send and receive data between two devices connected via a
physical network.
3. Chapter Summary | Next SectionPrevious Section
0 Replies
0 Replies