/* ---- Google Analytics Code Below */

Sunday, May 05, 2019

Linear Algebra for Large-Scale Machine Learning

Technical paper on machine learning and linear algebra.  Looks to have considerable promise for faster training and retraining systems by using compression techniques and then operating on the compressed data.  I often mention re-training because it is key to maintaining models in changing context. Sometimes also in real time situations.

https://vimeo.com/328181241   Technical Video introduction by Matthias Boehm.

Compressed Linear Algebra for Declarative Large-Scale Machine Learning
A Research Highlights article in the May 2019 CACM.

By Ahmed Elgohary, Matthias Boehm, Peter J. Haas, Frederick R. Reiss, Berthold Reinwald 
Communications of the ACM, May 2019, Vol. 62 No. 5, Pages 83-91
10.1145/3318221

Large-scale Machine Learning (ML) algorithms are often iterative, using repeated read-only data access and I/O-bound matrix-vector multiplications. Hence, it is crucial for performance to fit the data into single-node or distributed main memory to enable fast matrix-vector operations. General-purpose compression struggles to achieve both good compression ratios and fast decompression for block-wise uncompressed operations. Therefore, we introduce Compressed Linear Algebra (CLA) for lossless matrix compression. CLA encodes matrices with lightweight, value-based compression techniques and executes linear algebra operations directly on the compressed representations. We contribute effective column compression schemes, cache-conscious operations, and an efficient sampling-based compression algorithm. Our experiments show good compression ratios and operations performance close to the uncompressed case, which enables fitting larger datasets into available memory. We thereby obtain significant end-to-end performance improvements.

1. Introduction

Large-scale ML leverages large data collections to find interesting patterns or build robust predictive models.7 Applications range from traditional regression, classification, and clustering to user recommendations and deep learning for unstructured data. The labeled data required to train these ML models is now abundant, thanks to feedback loops in data products and weak supervision techniques. Many ML systems exploit data-parallel frameworks such as Spark20 or Flink2 for parallel model training and scoring on commodity hardware. It remains challenging, however, to train ML models on massive labeled data sets in a cost-effective manner. We provide compression-based methods for accelerating the linear algebra operations that are central to training. The key ideas are to perform these operations directly on the compressed data, and to automatically determine the best lossless compression scheme, as required by declarative ML systems.  .....  "

No comments: