This book presents an integrated collection of representative approaches for scaling up machine learning and data mining methods on parallel and distributed computing platforms. Demand for parallelizing learning algorithms is highly task-specific: in some settings it is driven by the enormous dataset sizes, in others by model complexity or by real-time performance requirements. Making task-appropriate algorithm and platform choices for large-scale machine learning requires understanding the benefits, trade-offs and constraints of the available options. Solutions presented in the book cover a range of parallelization platforms from FPGAs and GPUs to multi-core systems and commodity clusters, concurrent programming frameworks including CUDA, MPI, MapReduce and DryadLINQ, and learning settings (supervised, unsupervised, semi-supervised and online learning). Extensive coverage of parallelization of boosted trees, SVMs, spectral clustering, belief propagation and other popular learning algorithms and deep dives into several applications make the book equally useful for researchers, students and practitioners.
asdasda
评分虽然一开始说是面向工程师的,但还是有点过于偏向理论了,而且是由一篇篇独立论文组成的,深浅不一,内容感觉太杂,什么领域的都有
评分asdasda
评分不是我要的distributed learning.
评分虽然一开始说是面向工程师的,但还是有点过于偏向理论了,而且是由一篇篇独立论文组成的,深浅不一,内容感觉太杂,什么领域的都有
本站所有内容均为互联网搜索引擎提供的公开搜索信息,本站不存储任何数据与内容,任何内容与数据均与本站无关,如有需要请联系相关搜索引擎包括但不限于百度,google,bing,sogou 等
© 2025 onlinetoolsland.com All Rights Reserved. 本本书屋 版权所有