Documente Academic
Documente Profesional
Documente Cultură
Summary of the paper Dynamic Movement Primitives for Human Movement Recognition
This paper was written by Alp Burak Pehlivan and Erhan Oztop from the Özyeğin University in Istanbul, Turkey. It was
th
released on November 9 , 2015 and presented in the 41st Annual Conference of the IEEE: Industrial Electronics Society,
IECON 2015. Along this work, the authors take the Dynamic Movement Primitives (DMPs) —which is a method originally
used for movement trajectory generation— and uses it for recognition tasks. It also presents a comparison of commonly
used Hidden Markov Model (HMM) based recognition with DMP based recognition using human generated letter
trajectories. At the end, the results indicate that HMM gives better results than DMP, with possible noise robustness
advantage in DMPs for human movement.
I. INTRODUCTION
Dynamic Movement Primitives (DMPs) are proposed as a mathematical formulation of movement primitives. DMPs were
used in several applications including movement segmentation and hierarchical trajectory learning. Dynamic Movement
Primitives (DMPs) encode a desired movement trajectory in terms of the attractor dynamics of nonlinear differential
equations with the help of adjustable weight parameters. The HMM is one of most famous tools as a recognition method
for time-series data, especially in speech recognition.
DMPs and HMMs are two very different methods. They can be employed in movement recognition. In order to understand
how these methods compare in their movement recognition capabilities, in the paper, they designed a human movement
recognition task, and applied tests with both methods.
with the forcing term f representing an arbitrary non-linear function as a sum of weighted exponential basis functions:
and
The above dynamical system, named transformation system, has a global damping term -Dv, and K(g-x) creates an attractor
system towards the goal position. Finally, (g-x0)f(s) acts as the non-linear forcing term scaled by the (g-x0) factor. The
variable s on which the forcing term depends is a phase variable and its evolution is determined by the following decoupled
linear system . This variable evolves exponentially from 1 to 0. It is used to remove the direct time dependency of
the forcing term f(s), and provides the complete system with a time scalability by adjusting the parameter τ. The phase
variable is also used to weigh the forcing term, enabling to continuously shift towards a purely goal-attracted system.
B. HMM
HMMs are ubiquitous tool for modeling time series data. There are used in many applications including speech recognition,
pattern recognition, and molecular biology. HMM is a mathematical model of a stochastic process and includes three
parameters¸ λ=(Π,A,B) where Π represents initial vector, A is the transition matrix and B refers to emission matrix. HMMs
can be used for representing probability distributions over sequences of observations.
Miguel Adad Martínez Genis - 21485951
In HMMs, it is assumed that the observation at time t was generated by some process whose state St is hidden from the
observer. HMM also assumes that the state of the hidden process satisfies the Markov Property, that is given the value of
St-1, the current state St is independent of all states prior to t-1, In other words, the state at some time includes the history
of process inside. So, in principle we can predict the next state given the current state. The output also satisfies the Markov
Property. Taken together these Markov Properties mean that the joint distribution of sequence of states and observations
can be found in the following way:
In HMM one of the main problems is to calculating the probability of an observed sequence. This problem can be solved
more efficiently by forward-backward procedure. We use this probability to compare how well a model fits the given
observation. Thus we decide the class of movement that our test data belongs.
b) HMM
The hidden Markov Model is a probabilistic model. It is a representation of the joint distribution P(x,y). It is defined by two
probability distributions: the transition distribution and the observation distribution. We can get the overall probability of a
test data to be generated by this model using forward-backward algorithm. Then we can compare several model’s
probabilities of generating the test data for classification. In recognition task, we use the models with these calculated
parameters to guess most probable letter by testing each model’s log probability of the given observation sequence. The
observation sequence’s probability is calculated by forward-backward procedure.
IV. RESULTS
Tests were made in several data categories. They applied their tests with the best performance models that are trained
with the parameters chosen. Each test result consists of percentage of successful recognition of total of 100 handwritten
letter movements. In addition to the recognition performance in original data tests, the process time of both algorithms in
computer were measured. The success rates in HMM higher than DMP on the noise-free original data.
V. CONCLUSION
With the series of tests we made, we have observed that DMP gives 78% success compared to 83% success of recognition
with HMM in the test on the non-modified, noiseless human movement data. In general, HMM was much faster than DMP
in same recognition task. Furthermore, in tests with human movement data which we modified to emulate data loss during
recording, DMP gives worse results than HMM.