Design and development around Apache SPARK, Python and Hadoop Framework.
Extensive usage and experience with RDD and Data Frames with in Spark.
Extensive experience with data analytics and working knowledge of big data infrastructure such as various Hadoop Ecosystems like HDFS, Hive, Spark etc.
Should be working with gigabytes/terabytes of data and must understand the challenges of transforming and enriching such large datasets.
Provide effective solutions to address the business problems strategic and tactical.
Collaborate with other technology teams and architects to define and develop cross- function technology stack interactions.
Read, extract, transform, stage and load data to multiple targets, including Hadoop and Oracle.
Develop automation scripts around Hadoop framework to automate processes and existing flows around.
Should be able to modify existing programming/codes for new requirements.
Unit testing and debugging. Perform root cause analysis (RCA) for any failed processes.
17 freelancers are bidding on average ₹1036/hour for this job
Hello Sir, I have extensive experience in scala spark, hive hdfs hadoop frameworks. I have developed many real-time and batch applications in spark. I can surely handle all the task. Thanks
My current project work includes tools -- Kafka Spark Machine Learning Hadoop Python. I hope that matches your project. So I would like to discuss more about your project.