Hello, I am looking for a personal trainer in Apache Spark with python along with some bigdata tools like Kafka, Hbase, etc
We are looking for someone to join us with out business venture, we can discuss more in person about where we are and at what stage, suffice to say we have relatively good technical backgrounds but need someone who has experience in Hbase, Cassandra, Django, redis, thrift, PostgreSQL.
setting up tables in Hbase create topics in kafka set up the corn job
DOMAIN : BIG DATA AND HADOOP TITLE : REAL TIME PROJECT - INSURANCE LANGUAGE : JAVA VM : CLOUDERA QUICKSTART VM 5.5 IDE : ECLIPSE IDE ABSTRACT : Analyze health reports across years for the US market and find the average of privately and public insured people from years 2001-2011. The Project was processed by MapReduce method and output achieved.
I am working with input, output, python and Mapreduce framework in Linux cloud. My project is to write multiple mappers and reducers using python to solve and submit solutions to 3 different problem scenarios , all of the solutions tested in Hadoop set up in Linux cloud. (strictly restricted to writing programs that would work on the cloud- I am using AWS Cloud). For each problem, along with the mapper and reducer, you will be supposed to create a final program (or script) to combine the partial output files into one file. The combining program is supposed to only be combining the part files and sorting them, not affecting the stateless-ness of the mapper and reducer in any way. This project is due on the 12th of October. Will appreciate if you could quote a reasonable charge for completing this task on time. I can send the exact project requirements, relevant attachments and any other info you might need asap, once you confirm if this is something you could do. Thanks!