Skip to main content

Robust Recognition via Information Theoretic Learning

  • Book
  • © 2014

Overview

Part of the book series: SpringerBriefs in Computer Science (BRIEFSCOMPUTER)

This is a preview of subscription content, log in via an institution to check access.

Access this book

eBook USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access

Licence this eBook for your library

Institutional subscriptions

Table of contents (6 chapters)

Keywords

About this book

This Springer Brief represents a comprehensive review of information theoretic methods for robust recognition. A variety of information theoretic methods have been proffered in the past decade, in a large variety of computer vision applications; this work brings them together, attempts to impart the theory, optimization and usage of information entropy.

The authors resort to a new information theoretic concept, correntropy, as a robust measure and apply it to solve robust face recognition and object recognition problems. For computational efficiency, the brief introduces the additive and multiplicative forms of half-quadratic optimization to efficiently minimize entropy problems and a two-stage sparse presentation framework for large scale recognition problems. It also describes the strengths and deficiencies of different robust measures in solving robust recognition problems.

Authors and Affiliations

  • National Laboratory of Pattern Recognition, Institute of Automation Chinese Academy of Sciences, Beijing, China

    Ran He, Baogang Hu

  • School of Information and Control, Nanjing University of Information Science and Technology, Nanjing, China

    Xiaotong Yuan

  • Institute of Automaton Chinese Academy of Sciences, Beijing, China

    Liang Wang

Bibliographic Information

Publish with us