Time Estimation and Resource Minimization Scheme for Apache Spark and Hadoop Big Data Systems With Failures
Apache Spark and Hadoop are open source frameworks for big data processing, which have been adopted by many companies. In order to implement a reliable big data system that can satisfy processing target completion times, accurate resource provisioning and job execution time estimations are needed. I...
Main Authors: | Jinbae Lee, Bobae Kim, Jong-Moon Chung |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2019-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/8605312/ |
Similar Items
-
Big Data Analysis Using Apache Spark MLlib and Hadoop HDFS with Scala and Java
by: Hoger Khayrolla Omar, et al.
Published: (2019-05-01) -
Large Scale Implementations for Twitter Sentiment Classification
by: Andreas Kanavos, et al.
Published: (2017-03-01) -
A comprehensive performance analysis of Apache Hadoop and Apache Spark for large scale data sets using HiBench
by: N. Ahmed, et al.
Published: (2020-12-01) -
SANJYOT – WE SAVE LIFE Using Big Data - Apache Spark
by: Nipun Tyagi, et al.
Published: (2020-12-01) -
Comparative Analysis of Skew-Join Strategies for Large-Scale Datasets with MapReduce and Spark
by: Cao, H.-P, et al.
Published: (2022)