Job Description

Skills:
PySpark, Apache Hive, DataFrame API, Spark SQL, AWS Glue, ETL Pipelines, Python,

Urgent Hiring for Pyspark

Experience:-5 to 9 years

Location:- Pan India

Role :- C2H

Jd

Ability to design and build Python-based code generation framework and runtime engine by reading Business Rules repository in order to.

Requirements

  • Minimum 3 years of experience in build & deployment of Bigdata applications using SparkSQL, SparkStreaming in Python;
  • Expertise on graph algorithms and advanced recursion techniques;
  • Minimum 5 years of extensive experience in design, build and deployment of Python-based applications;
  • Minimum 3 years of experience in the following: HIVE, YARN, Kafka, HBase, MongoDB;
  • Hands-on experience in generating/parsing XML, JSON documents, and REST API request/responses;
  • Bachelors degree in a quantitative field (such as Engineering, Computer Science, Statistics, Econometrics) and a minimum of 5 years of experience;
  • Expertise in handling complex large-scale Big Data environments preferably (20Tb+);
  • Hands-on experience writing complex SQL queries, exporting and importing large amounts of data using utilities.