HIERARCHICALLY DISTRIBUTED KNOWLEDGE MATRIX: A BIG DATA PROCESSING FRAMEWORK COMPOSABLE

Hierarchically Distributed Knowledge Matrix – Over the past years, frameworks like MapReduce and Spark are introduced to ease the task of developing massive knowledge programs and applications. But, the roles in these frameworks are roughly defined and packaged as executable jars without any functionality being exposed or described. This suggests that that deployed jobs aren’t natively composable and reusable for subsequent development. Besides, it conjointly hampers the power for applying optimizations on the information flow of job sequences and pipelines.

In this Project, we tend to gift the Hierarchically Distributed Knowledge Matrix (HDM) that could be a practical, strongly-typed data representation for writing composable big knowledge applications. In conjunction with HDM, a runtime framework is provided to support the execution, integration and management of HDM applications on distributed infrastructures. Based mostly on the purposeful knowledge dependency graph of HDM, multiple optimizations are applied to enhance the performance of executing HDM jobs.

The experimental results show that our optimizations can achieve improvements between ten to forty p.c of the work-Completion-Time for different varieties of applications in comparison with the current state of art, Apache Spark.

Click here to get the complete project:

For more Project topics click here

Leave a Reply

Your email address will not be published. Required fields are marked *