Gør som tusindvis af andre bogelskere
Tilmeld dig nyhedsbrevet og få gode tilbud og inspiration til din næste læsning.
Ved tilmelding accepterer du vores persondatapolitik.Du kan altid afmelde dig igen.
While Java offers an apparently universal graphics library (the abstract window toolkit), the reality is quite different: For high-quality graphics and image display, we must program on the target platform itself (X or one of Microsoft's APIs) rather than rely on Java peer objects.
A comprehensive account of recent algorithms developed in computational number theory and primality testing. Provides a general framework for the theoretical study of public key cryptography and pseudorandom generators. Unique in its approach, the book will be a valuable addition to computer literature.
This is a textbook on program verification. It concentrates on those verification methods that have now become classic such as the inductive assertions method of Floyd, the axiomatic method of Hoare and Scott's fixpoint induction. The aim of the book is to present these different verification methods and to explain their mathemati cal background. The subject is treated with mathematical precision, and many examples are included. Throughout the book the same examples will reappear to illustrate how the different methods are related. The material is self-contained and accessible without prior knowledge of logic or semantics, but elementary knowledge of programming languages, formal languages and the theory of computation is helpful. A main concern has been to present the subject in as simple a setting as possible. For this reason three elementary, representative programming languages are in troduced: a flowchart programming language, a language of while-programs and a language of recursive programs. For these programming languages the operational and denotational semantics are introduced. Each verification method is then illustrated in the most appropriate of these languages and proved correct with the help of the most appropriate of the semantics.
With the advent of portable and autonomous computing systems, power con- sumption has emerged as a focal point in many research projects, commercial systems and DoD platforms. One current research initiative, which drew much attention to this area, is the Power Aware Computing and Communications (PAC/C) program sponsored by DARPA. Many of the chapters in this book include results from work that have been supported by the PACIC program. The performance of computer systems has been tremendously improving while the size and weight of such systems has been constantly shrinking. The capacities of batteries relative to their sizes and weights has been also improv- ing but at a rate which is much slower than the rate of improvement in computer performance and the rate of shrinking in computer sizes. The relation between the power consumption of a computer system and it performance and size is a complex one which is very much dependent on the specific system and the technology used to build that system. We do not need a complex argument, however, to be convinced that energy and power, which is the rate of energy consumption, are becoming critical components in computer systems in gen- eral, and portable and autonomous systems, in particular. Most of the early research on power consumption in computer systems ad- dressed the issue of minimizing power in a given platform, which usually translates into minimizing energy consumption, and thus, longer battery life.
Multiple processor systems are an important class of parallel systems. Over the years, several architectures have been proposed to build such systems to satisfy the requirements of high performance computing. These architectures span a wide variety of system types. At the low end of the spectrum, we can build a small, shared-memory parallel system with tens of processors. These systems typically use a bus to interconnect the processors and memory. Such systems, for example, are becoming commonplace in high-performance graph- ics workstations. These systems are called uniform memory access (UMA) multiprocessors because they provide uniform access of memory to all pro- cessors. These systems provide a single address space, which is preferred by programmers. This architecture, however, cannot be extended even to medium systems with hundreds of processors due to bus bandwidth limitations. To scale systems to medium range i. e. , to hundreds of processors, non-bus interconnection networks have been proposed. These systems, for example, use a multistage dynamic interconnection network. Such systems also provide global, shared memory like the UMA systems. However, they introduce local and remote memories, which lead to non-uniform memory access (NUMA) architecture. Distributed-memory architecture is used for systems with thousands of pro- cessors. These systems differ from the shared-memory architectures in that there is no globally accessible shared memory. Instead, they use message pass- ing to facilitate communication among the processors. As a result, they do not provide single address space.
This book introduces the fundamental concepts and practical simulation te- niques for modeling different aspects of operating systems to study their g- eral behavior and their performance.
Dr. Rocchi's topical and ever-timely volume proposes a novel, exhaustive solution to this vibrant issue.Paolo Rocchi, a versatile IBM scientist, outlines a new philosophical and mathematical approach inspired by well-tested software techniques.
This explosive growth, linked with the expectation that performance will continue its exponential rise with each new generation of hardware and that (in stark contrast to software) computer hardware will function correctly as soon as it comes off the assembly line, has its down side.
Mobile IP: Present State and Future is an up-to-date introduction to the rapidly evolving field of mobile IP.
Dr. Rocchi's topical and ever-timely volume proposes a novel, exhaustive solution to this vibrant issue.Paolo Rocchi, a versatile IBM scientist, outlines a new philosophical and mathematical approach inspired by well-tested software techniques.
Michael Stroebel worked for several years as a software engineer and consultant in the German IT industry before joining IBM Research in Switzerland, where he developed his interest in support for negotiations in electronic markets.
Dynamic Reconfiguration: Architectures and Algorithms offers a comprehensive treatment of dynamically reconfigurable computer architectures and algorithms for them.
Hard real-time systems are very predictable, but not sufficiently flexible to adapt to dynamic situations. Soft real-time systems are built to reduce resource consumption, tolerate overloads and adapt to system changes.
This explosive growth, linked with the expectation that performance will continue its exponential rise with each new generation of hardware and that (in stark contrast to software) computer hardware will function correctly as soon as it comes off the assembly line, has its down side.
Michael Stroebel worked for several years as a software engineer and consultant in the German IT industry before joining IBM Research in Switzerland, where he developed his interest in support for negotiations in electronic markets.
Tilmeld dig nyhedsbrevet og få gode tilbud og inspiration til din næste læsning.
Ved tilmelding accepterer du vores persondatapolitik.