Journal Articles
Permanent URI for this collectionhttps://mro.massey.ac.nz/handle/10179/7915
Browse
Search Results
Item A Machine Learning Approach to Enhance the Performance of D2D-Enabled Clustered Networks(IEEE, 20/01/2021) Aslam S; Alam F; Hasan SF; Rashid MAClustering has been suggested as an effective technique to enhance the performance of multicasting networks. Typically, a cluster head is selected to broadcast the cached content to its cluster members utilizing Device-to-Device (D2D) communication. However, some users can attain better performance by being connected with the Evolved Node B (eNB) rather than being in the clusters. In this article, we apply machine learning algorithms, namely Support Vector Machine, Random Forest, and Deep Neural Network to identify the users that should be serviced by the eNB. We therefore propose a mixed-mode content distribution scheme where the cluster heads and eNB service the two segregated groups of users to improve the performance of existing clustering schemes. A D2D-enabled multicasting scenario has been set up to perform a comprehensive simulation study that demonstrates that by utilizing the mixed-mode scheme, the performance of individual users, as well as the whole network, improve significantly in terms of throughput, energy consumption, and fairness. This study also demonstrates the trade-off between eNB loading and performance improvement for various parameters.Item A comprehensive performance analysis of Apache Hadoop and Apache Spark for large scale data sets using HiBench(BioMed Central Ltd, 14/12/2020) Ahmed N; Barczak ALC; Susnjak T; Rashid MABig Data analytics for storing, processing, and analyzing large-scale datasets has become an essential tool for the industry. The advent of distributed computing frameworks such as Hadoop and Spark offers efficient solutions to analyze vast amounts of data. Due to the application programming interface (API) availability and its performance, Spark becomes very popular, even more popular than the MapReduce framework. Both these frameworks have more than 150 parameters, and the combination of these parameters has a massive impact on cluster performance. The default system parameters help the system administrator deploy their system applications without much effort, and they can measure their specific cluster performance with factory-set parameters. However, an open question remains: can new parameter selection improve cluster performance for large datasets? In this regard, this study investigates the most impacting parameters, under resource utilization, input splits, and shuffle, to compare the performance between Hadoop and Spark, using an implemented cluster in our laboratory. We used a trial-and-error approach for tuning these parameters based on a large number of experiments. In order to evaluate the frameworks of comparative analysis, we select two workloads: WordCount and TeraSort. The performance metrics are carried out based on three criteria: execution time, throughput, and speedup. Our experimental results revealed that both system performances heavily depends on input data size and correct parameter selection. The analysis of the results shows that Spark has better performance as compared to Hadoop when data sets are small, achieving up to two times speedup in WordCount workloads and up to 14 times in TeraSort workloads when default parameter values are reconfigured.Item IoT Big Data provenance scheme using blockchain on Hadoop ecosystem(BioMed Central Ltd, 2021-12) Honar Pajooh H; Rashid MA; Alam F; Demidenko SThe diversity and sheer increase in the number of connected Internet of Things (IoT) devices have brought significant concerns associated with storing and protecting a large volume of IoT data. Storage volume requirements and computational costs are continuously rising in the conventional cloud-centric IoT structures. Besides, dependencies of the centralized server solution impose significant trust issues and make it vulnerable to security risks. In this paper, a layer-based distributed data storage design and implementation of a blockchain-enabled large-scale IoT system are proposed. It has been developed to mitigate the above-mentioned challenges by using the Hyperledger Fabric (HLF) platform for distributed ledger solutions. The need for a centralized server and a third-party auditor was eliminated by leveraging HLF peers performing transaction verifications and records audits in a big data system with the help of blockchain technology. The HLF blockchain facilitates storing the lightweight verification tags on the blockchain ledger. In contrast, the actual metadata are stored in the off-chain big data system to reduce the communication overheads and enhance data integrity. Additionally, a prototype has been implemented on embedded hardware showing the feasibility of deploying the proposed solution in IoT edge computing and big data ecosystems. Finally, experiments have been conducted to evaluate the performance of the proposed scheme in terms of its throughput, latency, communication, and computation costs. The obtained results have indicated the feasibility of the proposed solution to retrieve and store the provenance of large-scale IoT data within the Big Data ecosystem using the HLF blockchain. The experimental results show the throughput of about 600 transactions, 500 ms average response time, about 2–3% of the CPU consumption at the peer process and approximately 10–20% at the client node. The minimum latency remained below 1 s however, there is an increase in the maximum latency when the sending rate reached around 200 transactions per second (TPS).

