The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments

This paper identifies four common misconceptions about the scalability of volunteer computing on big data problems. The misconceptions are then clarified by analyzing the relationship between scalability and the impact factors including the problem size of big data, the heterogeneity and dynamics of...

Full description

Bibliographic Details
Main Authors: Wei Li, Maolin Tang
Format: Article
Language:English
Published: MDPI AG 2021-08-01
Series:Big Data and Cognitive Computing
Subjects:
Online Access:https://www.mdpi.com/2504-2289/5/3/38
id doaj-11e2caf15bdd4ad6bd4584278d5e65b6
record_format Article
spelling doaj-11e2caf15bdd4ad6bd4584278d5e65b62021-09-25T23:45:19ZengMDPI AGBig Data and Cognitive Computing2504-22892021-08-015383810.3390/bdcc5030038The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic EnvironmentsWei Li0Maolin Tang1School of Engineering & Technology, Central Queensland University, Rockhampton, QLD 4702, AustraliaSchool of Computer Science, Queensland University of Technology, Brisbane, QLD 4000, AustraliaThis paper identifies four common misconceptions about the scalability of volunteer computing on big data problems. The misconceptions are then clarified by analyzing the relationship between scalability and the impact factors including the problem size of big data, the heterogeneity and dynamics of volunteers, and the overlay structure. This paper proposes optimization strategies to find the optimal overlay for the given big data problem. This paper forms multiple overlays to optimize the performance of individual steps in terms of MapReduce paradigm. The optimization is to achieve the maximum overall performance by using a minimum number of volunteers, not overusing resources. This paper has demonstrated that the simulations on the concerned factors can fast find the optimization points. This paper concludes that always welcoming more volunteers is an overuse of available resources because they do not always bring benefit to the overall performance. Finding optimal use of volunteers are possible for the given big data problems even on the dynamics and opportunism of volunteers.https://www.mdpi.com/2504-2289/5/3/38bigdataoptimizationsimulationvolunteer computing
collection DOAJ
language English
format Article
sources DOAJ
author Wei Li
Maolin Tang
spellingShingle Wei Li
Maolin Tang
The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments
Big Data and Cognitive Computing
bigdata
optimization
simulation
volunteer computing
author_facet Wei Li
Maolin Tang
author_sort Wei Li
title The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments
title_short The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments
title_full The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments
title_fullStr The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments
title_full_unstemmed The Optimization Strategies on Clarification of the Misconceptions of Big Data Processing in Dynamic and Opportunistic Environments
title_sort optimization strategies on clarification of the misconceptions of big data processing in dynamic and opportunistic environments
publisher MDPI AG
series Big Data and Cognitive Computing
issn 2504-2289
publishDate 2021-08-01
description This paper identifies four common misconceptions about the scalability of volunteer computing on big data problems. The misconceptions are then clarified by analyzing the relationship between scalability and the impact factors including the problem size of big data, the heterogeneity and dynamics of volunteers, and the overlay structure. This paper proposes optimization strategies to find the optimal overlay for the given big data problem. This paper forms multiple overlays to optimize the performance of individual steps in terms of MapReduce paradigm. The optimization is to achieve the maximum overall performance by using a minimum number of volunteers, not overusing resources. This paper has demonstrated that the simulations on the concerned factors can fast find the optimization points. This paper concludes that always welcoming more volunteers is an overuse of available resources because they do not always bring benefit to the overall performance. Finding optimal use of volunteers are possible for the given big data problems even on the dynamics and opportunism of volunteers.
topic bigdata
optimization
simulation
volunteer computing
url https://www.mdpi.com/2504-2289/5/3/38
work_keys_str_mv AT weili theoptimizationstrategiesonclarificationofthemisconceptionsofbigdataprocessingindynamicandopportunisticenvironments
AT maolintang theoptimizationstrategiesonclarificationofthemisconceptionsofbigdataprocessingindynamicandopportunisticenvironments
AT weili optimizationstrategiesonclarificationofthemisconceptionsofbigdataprocessingindynamicandopportunisticenvironments
AT maolintang optimizationstrategiesonclarificationofthemisconceptionsofbigdataprocessingindynamicandopportunisticenvironments
_version_ 1717368079467413504