Margin-maximizing classification of sequential data with infinitely-long temporal dependencies

► We present a method for sequential data modeling. ► Our approach models temporal dependencies of infinite length. ► It employs a margin maximization training scheme. ► We evaluate it in computer vision applications. Generative models for sequential data are usually based on the assumption of tempo...

Full description

Saved in:
Bibliographic Details
Published in:Expert systems with applications Vol. 40; no. 11; pp. 4519 - 4527
Main Author: Chatzis, Sotirios P.
Format: Journal Article
Language:English
Published: Amsterdam Elsevier Ltd 01-09-2013
Elsevier
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:► We present a method for sequential data modeling. ► Our approach models temporal dependencies of infinite length. ► It employs a margin maximization training scheme. ► We evaluate it in computer vision applications. Generative models for sequential data are usually based on the assumption of temporal dependencies described by a first-order Markov chain. To ameliorate this shallow modeling assumption, several authors have proposed models with higher-order dependencies. However, the practical applicability of these approaches is hindered by their prohibitive computational costs in most cases. In addition, most existing approaches give rise to model training algorithms with objective functions that entail multiple spurious local optima, thus requiring application of tedious countermeasures to avoid getting trapped to bad model estimates. In this paper, we devise a novel margin-maximizing model with convex objective function that allows for capturing infinitely-long temporal dependencies in sequential datasets. This is effected by utilizing a recently proposed nonparametric Bayesian model of label sequences with infinitely-long temporal dependencies, namely the sequence memoizer, and training our model using margin maximization and a versatile mean-field-like approximation to allow for increased computational efficiency. As we experimentally demonstrate, the devised margin-maximizing construction of our model, which leads to a convex optimization scheme, without any spurious local optima, combined with the capacity of our model to capture long and complex temporal dependencies, allow for obtaining exceptional pattern recognition performance in several applications.
Bibliography:ObjectType-Article-2
SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 23
ObjectType-Article-1
ObjectType-Feature-2
ISSN:0957-4174
1873-6793
DOI:10.1016/j.eswa.2013.01.051