Pyspark | 4 To 6 Years | Bengaluru Job in Capgemini
Pyspark | 4 To 6 Years | Bengaluru
Capgemini
4+ weeks ago
- Bengaluru, Bangalore Urban, Karnataka
- Not Disclosed
- Full-time
Job Summary
Job Description
- Must have hands on experience implementing AWS Big data lake using EMR and Spark
- Working experience with Spark Hive Message Queue or Pub Sub Streaming technologies for 3 years
- Have 6 years of experience developing data pipelines using mix of languages Python Scala SQL etc and open source frameworks to implement data ingest processing and analytics technologies
- Experience leveraging open source big data processing frameworks such as Apache Spark Hadoop and streaming technologies such as Kafka
- Hands on experience with newer technologies relevant to the data space such as Spark Airflow Apache Druid Snowflake or any other OLAP databases
- Experience developing and deploying data pipelines and real time data streams within a cloud native infrastructure preferably AWS
- PySpark
- AWS
- Experience in using CI CD pipeline Gitlab
- Experience in Code Quality implementation
- Used Pep8 Pylint tools or any other code quality tool
- Experience of Python Plugins operators like FTP Sensor Oracle Operator etc
Ref:
497707
Posted on:
August 21, 2020
Experience level:
Experienced
Contract type:
Permanent
Location:
Bangalore
Business units:
I and D Global Practice
Department:
Big Data & Analytics


Help us improve JobGrin
Need Help? Contact us