Scalable Strategies for Computing with Massive Data

This paper presents two complementary statistical computing frameworks that address challenges in parallel processing and the analysis of massive data. First, the foreach package allows users of the R programming environment to define parallel loops that may be run sequentially on a single machine,...

Celý popis

Uložené v:
Podrobná bibliografia
Vydané v:Journal of statistical software Ročník 55; číslo 14; s. 1 - 19
Hlavní autori: Kane, Michael J., Emerson, John, Weston, Stephen
Médium: Journal Article
Jazyk:English
Vydavateľské údaje: Foundation for Open Access Statistics 01.11.2013
ISSN:1548-7660, 1548-7660
On-line prístup:Získať plný text
Tagy: Pridať tag
Žiadne tagy, Buďte prvý, kto otaguje tento záznam!
Popis
Shrnutí:This paper presents two complementary statistical computing frameworks that address challenges in parallel processing and the analysis of massive data. First, the foreach package allows users of the R programming environment to define parallel loops that may be run sequentially on a single machine, in parallel on a symmetric multiprocessing (SMP) machine, or in cluster environments without platform-specific code. Second, the bigmemory package implements memory- and file-mapped data structures that provide (a) access to arbitrarily large data while retaining a look and feel that is familiar to R users and (b) data structures that are shared across processor cores in order to support efficient parallel computing techniques. Although these packages may be used independently, this paper shows how they can be used in combination to address challenges that have effectively been beyond the reach of researchers who lack specialized software development skills or expensive hardware.
ISSN:1548-7660
1548-7660
DOI:10.18637/jss.v055.i14