pith. machine review for the scientific record. sign in

arxiv: 1610.02273 · v3 · submitted 2016-10-06 · 💻 cs.AR · cs.DC· cs.LG

Recognition: unknown

Near-Data Processing for Differentiable Machine Learning Models

Authors on Pith no claims yet
classification 💻 cs.AR cs.DCcs.LG
keywords processingdataadvantagescomparecomputationdifferentiableisp-mllearning
0
0 comments X
read the original abstract

Near-data processing (NDP) refers to augmenting memory or storage with processing power. Despite its potential for acceleration computing and reducing power requirements, only limited progress has been made in popularizing NDP for various reasons. Recently, two major changes have occurred that have ignited renewed interest and caused a resurgence of NDP. The first is the success of machine learning (ML), which often demands a great deal of computation for training, requiring frequent transfers of big data. The second is the popularity of NAND flash-based solid-state drives (SSDs) containing multicore processors that can accommodate extra computation for data processing. In this paper, we evaluate the potential of NDP for ML using a new SSD platform that allows us to simulate instorage processing (ISP) of ML workloads. Our platform (named ISP-ML) is a full-fledged simulator of a realistic multi-channel SSD that can execute various ML algorithms using data stored in the SSD. To conduct a thorough performance analysis and an in-depth comparison with alternative techniques, we focus on a specific algorithm: stochastic gradient descent (SGD), which is the de facto standard for training differentiable models such as logistic regression and neural networks. We implement and compare three SGD variants (synchronous, Downpour, and elastic averaging) using ISP-ML, exploiting the multiple NAND channels to parallelize SGD. In addition, we compare the performance of ISP and that of conventional in-host processing, revealing the advantages of ISP. Based on the advantages and limitations identified through our experiments, we further discuss directions for future research on ISP for accelerating ML.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. OffloadFS: Leveraging Disaggregated Storage for Computation Offloading

    cs.DC 2026-04 unverdicted novelty 5.0

    OffloadFS enables offloading of IO tasks to disaggregated storage nodes, improving RocksDB by up to 3.36x and ML preprocessing by 1.85x versus OCFS2.