Projects

PySpark

Configuring PySpark Job Cluster
from Submitting job to refine data ,
Worked with window and statistic function
At last, It worked with World Economic Indices to visualize
for Time Series Data in Zeppelin

Real - Time Streaming
with Kafka

Updating World News Stream
and, Building Recommendation System

△ Logic for Implementation

State :
Made in dataframe to build recommendation modeling
And, Ready to build Java consumer to connect with jdbc
Further, Testing elastic search
Flink will be last task to reduce and map with Geology data (Through Jar file or Zeppelin flink)