Skip to main content
Book cover

Boolean Functions and Computation Models

  • Textbook
  • © 2002

Overview

  • A survey of the present state of the art by internationally well-known authors
  • Focus on "fast" parallel computation
  • Includes numerous exercises ranging in difficulty
  • Includes supplementary material: sn.pub/extras

This is a preview of subscription content, log in via an institution to check access.

Access this book

eBook USD 64.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book USD 109.00
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book USD 84.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access

Licence this eBook for your library

Institutional subscriptions

Table of contents (7 chapters)

Keywords

About this book

The foundations of computational complexity theory go back to Alan Thring in the 1930s who was concerned with the existence of automatic procedures deciding the validity of mathematical statements. The first example of such a problem was the undecidability of the Halting Problem which is essentially the question of debugging a computer program: Will a given program eventu­ ally halt? Computational complexity today addresses the quantitative aspects of the solutions obtained: Is the problem to be solved tractable? But how does one measure the intractability of computation? Several ideas were proposed: A. Cobham [Cob65] raised the question of what is the right model in order to measure a "computation step" , M. Rabin [Rab60] proposed the introduction of axioms that a complexity measure should satisfy, and C. Shannon [Sha49] suggested the boolean circuit that computes a boolean function. However, an important question remains: What is the nature of computa­ tion? In 1957, John von Neumann [vN58] wrote in his notes for the Silliman Lectures concerning the nature of computation and the human brain that . . . logics and statistics should be primarily, although not exclusively, viewed as the basic tools of 'information theory'. Also, that body of experience which has grown up around the planning, evaluating, and coding of complicated logical and mathematical automata will be the focus of much of this information theory. The most typical, but not the only, such automata are, of course, the large electronic computing machines.

Reviews

From the reviews:

"The monograph gives the most recent and complete description of lower bounds for depth-restricted circuits, and propositional proof systems. … the authors present a research monograph on important subjects and provide many very recent results. I would recommend it for any university library and also for researchers." (Ingo Wegener, The Computer Journal, Vol. 46 (3), 2003)

Authors and Affiliations

  • Department of Computer Science and Department of Biology, Boston College, Chestnut Hill, USA

    Peter Clote

  • School of Computer Science, Carleton University, Ottawa, Canada

    Evangelos Kranakis

Bibliographic Information

Publish with us