Filter

My recent searches
Filter by:
Budget
to
to
to
Skills
Languages
    Job State
    134 jobs found, pricing in USD

    Hi, i have a big data project for my startup that I would love help with. Need help with optimizing my Cloudera 5.11 infrastructure as well as optimizing the performance of some scrapers for my wine analytics application. Also have sentiment analysis training that I need help with (my system does it natively using spark but needs to be trained still). Please be sure to list your experience with Cloudera in your application as well as any other strong skills you have that aren't listed in your profile.

    $35 / hr (Avg Bid)
    $35 / hr Avg Bid
    34 bids

    Hello, I am SDET in an Agile team which has build micro-services and big data applications. I had been doing manual+automation+performance testing for these services. My goal is to gain in-depth understanding of the already built services/apps and become a developer. I am looking for an expert who has hands-on experience on below skills: 1. Amazon Web Services (everything is on AWS) - Ex: ECS, Lambda, Elasticsearch Service, ElastiCache, CloudFormation, CloudWatch, EMR,....) 2. Scala (with Akka Actor Model) and sbt 3. Akka HTTP 4. Docker and Cloud-config service 5. Jenkins 6. Datastore, Protobuf, Catalogs 7. Apache Spark and Flink - we are running big data apps and filters on EMR 8. Bash/Shell Scripting 9. Java Thanks

    $15 / hr (Avg Bid)
    $15 / hr Avg Bid
    1 bids
    Hadoop Developer 3 days left

    I am looking for a candidate to work for me., i have a new project in my job and its totally on spark, kafka, scala and ETL., anyone interested please drop me a mail.

    $218 (Avg Bid)
    $218 Avg Bid
    21 bids

    Video Training on Big Data Hadoop. It would be screen recording and voice over. The recording will be approx 8 hrs Must cover Hadoop, MapReduce, HDFS, Spark, Pig, Hive, HBase, MongoDB, Cassandra, Flume

    $205 (Avg Bid)
    $205 Avg Bid
    4 bids

    Hello, I am looking for a personal trainer in Apache Spark with python along with some bigdata tools like Kafka, Hbase, etc

    $14 / hr (Avg Bid)
    $14 / hr Avg Bid
    10 bids

    I am developing an IoT solution for which i am expecting large data from edge devices to the cloud. looking for a freelancer to help me in: - writing requirements - designing solution - implementaton - integration with Google Cloud and other pieces of the solution for: - managing data streaming from edge (Kafka and probably spark/akka) that include connector/producer/consumer functionalities - storing the data in NoSQL (Cassandra) - analyzing and visualizing the data (Hadoop) Apart from pricing, if you can share an approach and thoughtful timeline, that will help me in decision making.

    $447 (Avg Bid)
    $447 Avg Bid
    9 bids

    Experts who are from Data Science domain

    $8 / hr (Avg Bid)
    $8 / hr Avg Bid
    11 bids

    - Assign weights to edges based on triangles exists

    $150 (Avg Bid)
    $150 Avg Bid
    2 bids
    Project for I D. Ended
    VERIFIED

    Hi I D., I noticed your profile and would like to offer you my project. We can discuss any details over chat.

    $21 / hr (Avg Bid)
    $21 / hr Avg Bid
    5 bids

    Search online to download reasonably large dataset. Define your own problem based on the dataset and provide a solution to it with your knowledge of Apache PySpark platform. You may obtain some idea for defining your own problem by referring to research papers. Include the reference in this case. Prepare a final report including 1) motivation, 2) design, and 3) relevant source code and screen shots. Also explain difficulties experienced and how to resolve them. Clearly indicate which item you attempt. Convert your report into PDF file.

    $158 (Avg Bid)
    $158 Avg Bid
    6 bids

    download a random large dataset. Define your own problem based on the dataset (such as solve an issue in dataset you take)and provide a solution to it with your knowledge of Apache PySpark platform.

    $138 (Avg Bid)
    $138 Avg Bid
    9 bids

    Search online and find a PySpark tutorial or project and follow it on your own. You should indicate the [login to view URL] a final report including 1) motivation, 2) design, and 3) relevant source code and screen shots. Also explain difficulties experienced and how to resolve them. or Search online to download reasonably large dataset. Define your own problem based on the dataset and provide a solution to it with your knowledge of Apache PySpark platform. You may obtain some idea for defining your own problem by referring to research papers. Include the reference in this case..Prepare a final report including 1) motivation, 2) design, and 3) relevant source code and screen shots. Also explain difficulties experienced and how to resolve them.

    $52 (Avg Bid)
    $52 Avg Bid
    5 bids

    I have an application in which user selects a folder from hdfs, and the application writes the results in hdfs/output/directory. so we need to write code in java for checking permissions of output directory before writing the results in hdfs/output/directory.

    $32 (Avg Bid)
    $32 Avg Bid
    15 bids

    I have a sql ETL code and i would like someone to convert the code to Python that can run in Apache Beam

    $1130 (Avg Bid)
    $1130 Avg Bid
    20 bids

    Help needed with Pyspark installation on Windows and running Pyspark MLlib module

    $14 / hr (Avg Bid)
    $14 / hr Avg Bid
    3 bids

    below upload file implement the code on spark with output

    $20 (Avg Bid)
    $20 Avg Bid
    4 bids

    Data Frame1: EmpNo EmpName Salary E123 Tom 2000 E124 RAM 2000 E125 TAM 2000 E126 SAM 2000 E124 RAM 4000 E126 SAM 6000 E125 TAM 9000 E123 Tom1 4000 Transform this DataFrame to DataFrame2: EmpNo EmpName Salary rownum E123 Tom 2000 1 E124 RAM 2000 1 E125 TAM 2000 1 E126 SAM 2000 1 E126 SAM1 5000 2 E124 RAM 4000 2 E126 SAM 6000 3 E125 TAM 9000 2 E123 Tom1 4000 2 Here is the summary: -- Duplicate EMpnos should be indexed(as shown in the rownum column) -- The order of index should be based on salary. -- Need All of the below approaches -- Should be optimized and should be runnable on a cluster. 1. Using RDD(a. using SparkContext, [login to view URL]) 2. Using DataFrame(a. using SparkContext, [login to view URL]) 3. Using DataSet(a. using SparkContext, [login to view URL]) DataFrame Operations should contain both . notation and sql notation. Action Items: 1) Development 2) Testing 3) Demo 4) Any corrections/small enhancements(if required)

    $25 (Avg Bid)
    $25 Avg Bid
    15 bids

    This assignment is broken up into sections with bite-sized examples for demonstrating Spark functionality for log processing. For each problem, you should start by thinking about the algorithm that you will use to efficiently process the log in a parallel, distributed manner. This means using the various RDD operations along with lambda functions that are applied at each worker.

    $7 / hr (Avg Bid)
    $7 / hr Avg Bid
    13 bids

    I am looking for a freelancer with expertise in Hadoop, Nifi, Hive, Spark, Python, AWS Services. Expertise in the above is must specially ***Nifi*** You can directly contact me here [Removed by Freelancer.com Admin for offsiting - please see Section 13 of our Terms and Conditions]

    $654 (Avg Bid)
    $654 Avg Bid
    17 bids

    Top Spark Community Articles