Job Description
Skills:
PySpark, Apache Hive, DataFrame API, Spark SQL, AWS Glue, ETL Pipelines, Python,
PySpark, Apache Hive, DataFrame API, Spark SQL, AWS Glue, ETL Pipelines, Python,
Urgent Hiring for Pyspark
Experience:-5 to 9 years
Location:- Pan India
Role :- C2H
Jd
Ability to design and build Python-based code generation framework and runtime engine by reading Business Rules repository in order to.
Requirements
- Minimum 3 years of experience in build & deployment of Bigdata applications using SparkSQL, SparkStreaming in Python;
- Expertise on graph algorithms and advanced recursion techniques;
- Minimum 5 years of extensive experience in design, build and deployment of Python-based applications;
- Minimum 3 years of experience in the following: HIVE, YARN, Kafka, HBase, MongoDB;
- Hands-on experience in generating/parsing XML, JSON documents, and REST API request/responses;
- Bachelors degree in a quantitative field (such as Engineering, Computer Science, Statistics, Econometrics) and a minimum of 5 years of experience;
- Expertise in handling complex large-scale Big Data environments preferably (20Tb+);
- Hands-on experience writing complex SQL queries, exporting and importing large amounts of data using utilities.