Machine Learning under Resource Constraints - Fundamentals

Machine Learning under Resource Constraints - Fundamentals

Author: Katharina Morik

Publisher: Walter de Gruyter GmbH & Co KG

Published: 2022-12-31

Total Pages: 542

ISBN-13: 3110786125

DOWNLOAD EBOOK

Machine Learning under Resource Constraints addresses novel machine learning algorithms that are challenged by high-throughput data, by high dimensions, or by complex structures of the data in three volumes. Resource constraints are given by the relation between the demands for processing the data and the capacity of the computing machinery. The resources are runtime, memory, communication, and energy. Hence, modern computer architectures play a significant role. Novel machine learning algorithms are optimized with regard to minimal resource consumption. Moreover, learned predictions are executed on diverse architectures to save resources. It provides a comprehensive overview of the novel approaches to machine learning research that consider resource constraints, as well as the application of the described methods in various domains of science and engineering. Volume 1 establishes the foundations of this new field. It goes through all the steps from data collection, their summary and clustering, to the different aspects of resource-aware learning, i.e., hardware, memory, energy, and communication awareness. Several machine learning methods are inspected with respect to their resource requirements and how to enhance their scalability on diverse computing architectures ranging from embedded systems to large computing clusters.


Machine Learning under Resource Constraints - Applications

Machine Learning under Resource Constraints - Applications

Author: Katharina Morik

Publisher: Walter de Gruyter GmbH & Co KG

Published: 2022-12-31

Total Pages: 497

ISBN-13: 3110786141

DOWNLOAD EBOOK

Machine Learning under Resource Constraints addresses novel machine learning algorithms that are challenged by high-throughput data, by high dimensions, or by complex structures of the data in three volumes. Resource constraints are given by the relation between the demands for processing the data and the capacity of the computing machinery. The resources are runtime, memory, communication, and energy. Hence, modern computer architectures play a significant role. Novel machine learning algorithms are optimized with regard to minimal resource consumption. Moreover, learned predictions are executed on diverse architectures to save resources. It provides a comprehensive overview of the novel approaches to machine learning research that consider resource constraints, as well as the application of the described methods in various domains of science and engineering. Volume 3 describes how the resource-aware machine learning methods and techniques are used to successfully solve real-world problems. The book provides numerous specific application examples. In the areas of health and medicine, it is demonstrated how machine learning can improve risk modelling, diagnosis, and treatment selection for diseases. Machine learning supported quality control during the manufacturing process in a factory allows to reduce material and energy cost and save testing times is shown by the diverse real-time applications in electronics and steel production as well as milling. Additional application examples show, how machine-learning can make traffic, logistics and smart cities more effi cient and sustainable. Finally, mobile communications can benefi t substantially from machine learning, for example by uncovering hidden characteristics of the wireless channel.


Discovery in Physics

Discovery in Physics

Author: Katharina Morik

Publisher: Walter de Gruyter GmbH & Co KG

Published: 2022-12-31

Total Pages: 364

ISBN-13: 311078596X

DOWNLOAD EBOOK

Machine learning is part of Artificial Intelligence since its beginning. Certainly, not learning would only allow the perfect being to show intelligent behavior. All others, be it humans or machines, need to learn in order to enhance their capabilities. In the eighties of the last century, learning from examples and modeling human learning strategies have been investigated in concert. The formal statistical basis of many learning methods has been put forward later on and is still an integral part of machine learning. Neural networks have always been in the toolbox of methods. Integrating all the pre-processing, exploitation of kernel functions, and transformation steps of a machine learning process into the architecture of a deep neural network increased the performance of this model type considerably. Modern machine learning is challenged on the one hand by the amount of data and on the other hand by the demand of real-time inference. This leads to an interest in computing architectures and modern processors. For a long time, the machine learning research could take the von-Neumann architecture for granted. All algorithms were designed for the classical CPU. Issues of implementation on a particular architecture have been ignored. This is no longer possible. The time for independently investigating machine learning and computational architecture is over. Computing architecture has experienced a similarly rampant development from mainframe or personal computers in the last century to now very large compute clusters on the one hand and ubiquitous computing of embedded systems in the Internet of Things on the other hand. Cyber-physical systems’ sensors produce a huge amount of streaming data which need to be stored and analyzed. Their actuators need to react in real-time. This clearly establishes a close connection with machine learning. Cyber-physical systems and systems in the Internet of Things consist of diverse components, heterogeneous both in hard- and software. Modern multi-core systems, graphic processors, memory technologies and hardware-software codesign offer opportunities for better implementations of machine learning models. Machine learning and embedded systems together now form a field of research which tackles leading edge problems in machine learning, algorithm engineering, and embedded systems. Machine learning today needs to make the resource demands of learning and inference meet the resource constraints of used computer architecture and platforms. A large variety of algorithms for the same learning method and, moreover, diverse implementations of an algorithm for particular computing architectures optimize learning with respect to resource efficiency while keeping some guarantees of accuracy. The trade-off between a decreased energy consumption and an increased error rate, to just give an example, needs to be theoretically shown for training a model and the model inference. Pruning and quantization are ways of reducing the resource requirements by either compressing or approximating the model. In addition to memory and energy consumption, timeliness is an important issue, since many embedded systems are integrated into large products that interact with the physical world. If the results are delivered too late, they may have become useless. As a result, real-time guarantees are needed for such systems. To efficiently utilize the available resources, e.g., processing power, memory, and accelerators, with respect to response time, energy consumption, and power dissipation, different scheduling algorithms and resource management strategies need to be developed. This book series addresses machine learning under resource constraints as well as the application of the described methods in various domains of science and engineering. Turning big data into smart data requires many steps of data analysis: methods for extracting and selecting features, filtering and cleaning the data, joining heterogeneous sources, aggregating the data, and learning predictions need to scale up. The algorithms are challenged on the one hand by high-throughput data, gigantic data sets like in astrophysics, on the other hand by high dimensions like in genetic data. Resource constraints are given by the relation between the demands for processing the data and the capacity of the computing machinery. The resources are runtime, memory, communication, and energy. Novel machine learning algorithms are optimized with regard to minimal resource consumption. Moreover, learned predictions are applied to program executions in order to save resources. The three books will have the following subtopics: Volume 1: Machine Learning under Resource Constraints - Fundamentals Volume 2: Machine Learning and Physics under Resource Constraints - Discovery Volume 3: Machine Learning under Resource Constraints - Applications Volume 2 is about machine learning for knowledge discovery in particle and astroparticle physics. Their instruments, e.g., particle accelerators or telescopes, gather petabytes of data. Here, machine learning is necessary not only to process the vast amounts of data and to detect the relevant examples efficiently, but also as part of the knowledge discovery process itself. The physical knowledge is encoded in simulations that are used to train the machine learning models. At the same time, the interpretation of the learned models serves to expand the physical knowledge. This results in a cycle of theory enhancement supported by machine learning.


Applications

Applications

Author: Katharina Morik

Publisher: Walter de Gruyter GmbH & Co KG

Published: 2022-12-31

Total Pages: 478

ISBN-13: 3110785986

DOWNLOAD EBOOK

Machine learning is part of Artificial Intelligence since its beginning. Certainly, not learning would only allow the perfect being to show intelligent behavior. All others, be it humans or machines, need to learn in order to enhance their capabilities. In the eighties of the last century, learning from examples and modeling human learning strategies have been investigated in concert. The formal statistical basis of many learning methods has been put forward later on and is still an integral part of machine learning. Neural networks have always been in the toolbox of methods. Integrating all the pre-processing, exploitation of kernel functions, and transformation steps of a machine learning process into the architecture of a deep neural network increased the performance of this model type considerably. Modern machine learning is challenged on the one hand by the amount of data and on the other hand by the demand of real-time inference. This leads to an interest in computing architectures and modern processors. For a long time, the machine learning research could take the von-Neumann architecture for granted. All algorithms were designed for the classical CPU. Issues of implementation on a particular architecture have been ignored. This is no longer possible. The time for independently investigating machine learning and computational architecture is over. Computing architecture has experienced a similarly rampant development from mainframe or personal computers in the last century to now very large compute clusters on the one hand and ubiquitous computing of embedded systems in the Internet of Things on the other hand. Cyber-physical systems’ sensors produce a huge amount of streaming data which need to be stored and analyzed. Their actuators need to react in real-time. This clearly establishes a close connection with machine learning. Cyber-physical systems and systems in the Internet of Things consist of diverse components, heterogeneous both in hard- and software. Modern multi-core systems, graphic processors, memory technologies and hardware-software codesign offer opportunities for better implementations of machine learning models. Machine learning and embedded systems together now form a field of research which tackles leading edge problems in machine learning, algorithm engineering, and embedded systems. Machine learning today needs to make the resource demands of learning and inference meet the resource constraints of used computer architecture and platforms. A large variety of algorithms for the same learning method and, moreover, diverse implementations of an algorithm for particular computing architectures optimize learning with respect to resource efficiency while keeping some guarantees of accuracy. The trade-off between a decreased energy consumption and an increased error rate, to just give an example, needs to be theoretically shown for training a model and the model inference. Pruning and quantization are ways of reducing the resource requirements by either compressing or approximating the model. In addition to memory and energy consumption, timeliness is an important issue, since many embedded systems are integrated into large products that interact with the physical world. If the results are delivered too late, they may have become useless. As a result, real-time guarantees are needed for such systems. To efficiently utilize the available resources, e.g., processing power, memory, and accelerators, with respect to response time, energy consumption, and power dissipation, different scheduling algorithms and resource management strategies need to be developed. This book series addresses machine learning under resource constraints as well as the application of the described methods in various domains of science and engineering. Turning big data into smart data requires many steps of data analysis: methods for extracting and selecting features, filtering and cleaning the data, joining heterogeneous sources, aggregating the data, and learning predictions need to scale up. The algorithms are challenged on the one hand by high-throughput data, gigantic data sets like in astrophysics, on the other hand by high dimensions like in genetic data. Resource constraints are given by the relation between the demands for processing the data and the capacity of the computing machinery. The resources are runtime, memory, communication, and energy. Novel machine learning algorithms are optimized with regard to minimal resource consumption. Moreover, learned predictions are applied to program executions in order to save resources. The three books will have the following subtopics: Volume 1: Machine Learning under Resource Constraints - Fundamentals Volume 2: Machine Learning and Physics under Resource Constraints - Discovery Volume 3: Machine Learning under Resource Constraints - Applications Volume 3 describes how the resource-aware machine learning methods and techniques are used to successfully solve real-world problems. The book provides numerous specific application examples. In the areas of health and medicine, it is demonstrated how machine learning can improve risk modelling, diagnosis, and treatment selection for diseases. Machine learning supported quality control during the manufacturing process in a factory allows to reduce material and energy cost and save testing times is shown by the diverse real-time applications in electronics and steel production as well as milling. Additional application examples show, how machine-learning can make traffic, logistics and smart cities more efficient and sustainable. Finally, mobile communications can benefit substantially from machine learning, for example by uncovering hidden characteristics of the wireless channel.


Hybrid Information Systems

Hybrid Information Systems

Author: Ramakant Bhardwaj

Publisher: Walter de Gruyter GmbH & Co KG

Published: 2024-07-22

Total Pages: 520

ISBN-13: 311133113X

DOWNLOAD EBOOK

The book provides comprehensive and cognitive approach to building and deploying sophisticated information systems. The book utilizes non-linear optimization techniques, fuzzy logic, and rough sets to model various real-world use cases for the digital era. The hybrid information system modeling handles both qualitative and quantitative data and can effectively handle uncertainty and imprecision in the data. The combination of non-linear optimization mechanisms, fuzzy logic, and rough sets provides a robust foundation for next-generation information systems that can fulfill the demands of adaptive, aware, and adroit software applications for the knowledge era. The book emphasizes the importance of the hybrid approach, which combines the strengths of both mathematical and AI techniques, to achieve a more comprehensive and effective modeling process. Hybrid information system modeling techniques combine different approaches, such as fuzzy logic, rough sets, and neural networks, to create models that can handle the complexity and uncertainty of real-world problems. These techniques provide a powerful tool for modeling and analyzing complex systems, and the applications of hybrid information system modeling demonstrate their potential for solving real-world problems in various fields.


Machine Learning for Automated Theorem Proving

Machine Learning for Automated Theorem Proving

Author: Sean B. Holden

Publisher:

Published: 2021-11-22

Total Pages: 202

ISBN-13: 9781680838985

DOWNLOAD EBOOK

In this book, the author presents the results of his thorough and systematic review of the research at the intersection of two apparently rather unrelated fields: Automated Theorem Proving (ATP) and Machine Learning (ML).


Column Generation

Column Generation

Author: Guy Desaulniers

Publisher: Springer Science & Business Media

Published: 2006-03-20

Total Pages: 369

ISBN-13: 0387254862

DOWNLOAD EBOOK

Column Generation is an insightful overview of the state of the art in integer programming column generation and its many applications. The volume begins with "A Primer in Column Generation" which outlines the theory and ideas necessary to solve large-scale practical problems, illustrated with a variety of examples. Other chapters follow this introduction on "Shortest Path Problems with Resource Constraints," "Vehicle Routing Problem with Time Window," "Branch-and-Price Heuristics," "Cutting Stock Problems," each dealing with methodological aspects of the field. Three chapters deal with transportation applications: "Large-scale Models in the Airline Industry," "Robust Inventory Ship Routing by Column Generation," and "Ship Scheduling with Recurring Visits and Visit Separation Requirements." Production is the focus of another three chapters: "Combining Column Generation and Lagrangian Relaxation," "Dantzig-Wolfe Decomposition for Job Shop Scheduling," and "Applying Column Generation to Machine Scheduling." The final chapter by François Vanderbeck, "Implementing Mixed Integer Column Generation," reviews how to set-up the Dantzig-Wolfe reformulation, adapt standard MIP techniques to the column generation context (branching, preprocessing, primal heuristics), and deal with specific column generation issues (initialization, stabilization, column management strategies).


Human-Centered Data Science

Human-Centered Data Science

Author: Cecilia Aragon

Publisher: MIT Press

Published: 2022-03-01

Total Pages: 201

ISBN-13: 0262367599

DOWNLOAD EBOOK

Best practices for addressing the bias and inequality that may result from the automated collection, analysis, and distribution of large datasets. Human-centered data science is a new interdisciplinary field that draws from human-computer interaction, social science, statistics, and computational techniques. This book, written by founders of the field, introduces best practices for addressing the bias and inequality that may result from the automated collection, analysis, and distribution of very large datasets. It offers a brief and accessible overview of many common statistical and algorithmic data science techniques, explains human-centered approaches to data science problems, and presents practical guidelines and real-world case studies to help readers apply these methods. The authors explain how data scientists’ choices are involved at every stage of the data science workflow—and show how a human-centered approach can enhance each one, by making the process more transparent, asking questions, and considering the social context of the data. They describe how tools from social science might be incorporated into data science practices, discuss different types of collaboration, and consider data storytelling through visualization. The book shows that data science practitioners can build rigorous and ethical algorithms and design projects that use cutting-edge computational tools and address social concerns.


Resource-Constrained Project Scheduling

Resource-Constrained Project Scheduling

Author: Christian Artigues

Publisher: John Wiley & Sons

Published: 2013-03-01

Total Pages: 235

ISBN-13: 1118623703

DOWNLOAD EBOOK

This title presents a large variety of models and algorithms dedicated to the resource-constrained project scheduling problem (RCPSP), which aims at scheduling at minimal duration a set of activities subject to precedence constraints and limited resource availabilities. In the first part, the standard variant of RCPSP is presented and analyzed as a combinatorial optimization problem. Constraint programming and integer linear programming formulations are given. Relaxations based on these formulations and also on related scheduling problems are presented. Exact methods and heuristics are surveyed. Computational experiments, aiming at providing an empirical insight on the difficulty of the problem, are provided. The second part of the book focuses on several other variants of the RCPSP and on their solution methods. Each variant takes account of real-life characteristics which are not considered in the standard version, such as possible interruptions of activities, production and consumption of resources, cost-based approaches and uncertainty considerations. The last part presents industrial case studies where the RCPSP plays a central part. Applications are presented in various domains such as assembly shop and rolling ingots production scheduling, project management in information technology companies and instruction scheduling for VLIW processor architectures.


TinyML

TinyML

Author: Pete Warden

Publisher: O'Reilly Media

Published: 2019-12-16

Total Pages: 504

ISBN-13: 1492052019

DOWNLOAD EBOOK

Deep learning networks are getting smaller. Much smaller. The Google Assistant team can detect words with a model just 14 kilobytes in size—small enough to run on a microcontroller. With this practical book you’ll enter the field of TinyML, where deep learning and embedded systems combine to make astounding things possible with tiny devices. Pete Warden and Daniel Situnayake explain how you can train models small enough to fit into any environment. Ideal for software and hardware developers who want to build embedded systems using machine learning, this guide walks you through creating a series of TinyML projects, step-by-step. No machine learning or microcontroller experience is necessary. Build a speech recognizer, a camera that detects people, and a magic wand that responds to gestures Work with Arduino and ultra-low-power microcontrollers Learn the essentials of ML and how to train your own models Train models to understand audio, image, and accelerometer data Explore TensorFlow Lite for Microcontrollers, Google’s toolkit for TinyML Debug applications and provide safeguards for privacy and security Optimize latency, energy usage, and model and binary size