Margin-maximizing classification of sequential data with infinitely-long temporal dependencies
► We present a method for sequential data modeling. ► Our approach models temporal dependencies of infinite length. ► It employs a margin maximization training scheme. ► We evaluate it in computer vision applications. Generative models for sequential data are usually based on the assumption of tempo...
Saved in:
Published in: | Expert systems with applications Vol. 40; no. 11; pp. 4519 - 4527 |
---|---|
Main Author: | |
Format: | Journal Article |
Language: | English |
Published: |
Amsterdam
Elsevier Ltd
01-09-2013
Elsevier |
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | ► We present a method for sequential data modeling. ► Our approach models temporal dependencies of infinite length. ► It employs a margin maximization training scheme. ► We evaluate it in computer vision applications.
Generative models for sequential data are usually based on the assumption of temporal dependencies described by a first-order Markov chain. To ameliorate this shallow modeling assumption, several authors have proposed models with higher-order dependencies. However, the practical applicability of these approaches is hindered by their prohibitive computational costs in most cases. In addition, most existing approaches give rise to model training algorithms with objective functions that entail multiple spurious local optima, thus requiring application of tedious countermeasures to avoid getting trapped to bad model estimates. In this paper, we devise a novel margin-maximizing model with convex objective function that allows for capturing infinitely-long temporal dependencies in sequential datasets. This is effected by utilizing a recently proposed nonparametric Bayesian model of label sequences with infinitely-long temporal dependencies, namely the sequence memoizer, and training our model using margin maximization and a versatile mean-field-like approximation to allow for increased computational efficiency. As we experimentally demonstrate, the devised margin-maximizing construction of our model, which leads to a convex optimization scheme, without any spurious local optima, combined with the capacity of our model to capture long and complex temporal dependencies, allow for obtaining exceptional pattern recognition performance in several applications. |
---|---|
Bibliography: | ObjectType-Article-2 SourceType-Scholarly Journals-1 ObjectType-Feature-1 content type line 23 ObjectType-Article-1 ObjectType-Feature-2 |
ISSN: | 0957-4174 1873-6793 |
DOI: | 10.1016/j.eswa.2013.01.051 |