Improving Data Locality of Tasks by Executor Allocation in Spark Computing Environment

The concept of data locality is crucial for distributed systems (e.g., Spark and Hadoop) to process Big Data. Most of the existing research optimized the data locality from the aspect of task scheduling. However, as the execution container of Spark's tasks, the executor launched on different no...

Full description

Saved in:
Bibliographic Details
Published in:IEEE transactions on cloud computing Vol. 12; no. 3; pp. 876 - 888
Main Authors: Fu, Zhongming, He, Mengsi, Yi, Yang, Tang, Zhuo
Format: Journal Article
Language:English
Published: Piscataway IEEE 01.07.2024
The Institute of Electrical and Electronics Engineers, Inc. (IEEE)
Subjects:
ISSN:2168-7161, 2372-0018
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Be the first to leave a comment!
You must be logged in first