Contains the proceedings from the Data Compression Conference 1998. Subjects covered include: piecewise integer mapping for arithmetic coding; compression of sparse matrices by arithmetic coding; tag based models of English text; practical multi-resolution source coding; and TSVQ."
The aim of this reference work is to make African human rights law accessible to all those involved in or interested in human rights law on the continent in order to strengthen its impact. Primary documents are introduced and reproduced and presented in a coherent framework. The main institutions - public and private - dealing with human rights in Africa are identified and discussed. Comprehensive overviews of the international human rights legal regimes applicable to Africa, as well as country reports are provided. This book tries to contribute towards documenting, systemising and anchoring the African human rights system. The print edition is available as a set of two volumes (9789004138810).
The proliferation of massive data sets brings with it a series of special computational challenges. This "data avalanche" arises in a wide range of scientific and commercial applications. With advances in computer and information technologies, many of these challenges are beginning to be addressed by diverse inter-disciplinary groups, that indude computer scientists, mathematicians, statisticians and engineers, working in dose cooperation with application domain experts. High profile applications indude astrophysics, bio-technology, demographics, finance, geographi cal information systems, government, medicine, telecommunications, the environment and the internet. John R. Tucker of the Board on Mathe matical Seiences has stated: "My interest in this problern (Massive Data Sets) isthat I see it as the rnost irnportant cross-cutting problern for the rnathernatical sciences in practical problern solving for the next decade, because it is so pervasive. " The Handbook of Massive Data Sets is comprised of articles writ ten by experts on selected topics that deal with some major aspect of massive data sets. It contains chapters on information retrieval both in the internet and in the traditional sense, web crawlers, massive graphs, string processing, data compression, dustering methods, wavelets, op timization, external memory algorithms and data structures, the US national duster project, high performance computing, data warehouses, data cubes, semi-structured data, data squashing, data quality, billing in the large, fraud detection, and data processing in astrophysics, air pollution, biomolecular data, earth observation and the environment.
You'll learn not only to choose the optimal compression strategy for your project, but also to apply it in a way that guarantees the best possible results."--BOOK JACKET.
Opening with a detailed review of existing techniques for selective encryption, this text then examines algorithms that combine both encryption and compression. The book also presents a selection of specific examples of the design and implementation of secure embedded multimedia systems. Features: reviews the historical developments and latest techniques in multimedia compression and encryption; discusses an approach to reduce the computational cost of multimedia encryption, while preserving the properties of compressed video; introduces a polymorphic wavelet architecture that can make dynamic resource allocation decisions according to the application requirements; proposes a light-weight multimedia encryption strategy based on a modified discrete wavelet transform; describes a reconfigurable hardware implementation of a chaotic filter bank scheme with enhanced security features; presents an encryption scheme for image and video data based on chaotic arithmetic coding.