Senior Data Developer Job in Delhivery Pvt Ltd.

Senior Data Developer

Apply Now
Job Summary

Description

Looking for the big data engineer to help shape our technology and product roadmap. You will be part of the fast-paced, entrepreneurial team that enables Big Data and batch/ real-time analytical solutions leveraging transformational technologies (Python, Scala, Java, Hadoop, MapReduce, Kafka, Hive, HBase, Spark, Storm, etc.) to deliver innovative solutions.

As a data developer at Delhivery you are expected to work closely with Product team to do rapid development of new features / enhancements, work with the Data Services team to design and build data stores, write effective MapReduce / aggregations, write quality code and work with your lead and team to drive quality and efficiency.

Responsibility

  • You will be responsible for delivering high-value next-generation products on aggressive deadlines and will be required to write high-quality, highly optimized/high-performance and maintainable code
  • Manage ETL/ELT pipelines of various Microservices
  • Work on distributed/big-data system to build, release and maintain an Always On scalable data processing and reporting platform
  • Work on relational and NoSQL databases
  • Build scalable architectures for data storage, transformation and analysis
  • Ability to work in the fast-paced and dynamic environment

Qualifications

  • Overall 3+ years of IT experience in a variety of industries, which includes hands on experience in Big Data Analytics and development
  • 2+ yrs of experience in writing pyspark for data transformation.
  • 2+ years of experience with detailed knowledge of data warehouse technical architectures, ETL/ ELT, reporting/analytic tools, and data security
  • 2+ years of experience in designing data warehouse solutions and integrating technical components
  • 2+ yrs of experience leading data warehousing and analytics projects, including using AWS technologies - Redshift, S3, EC2, Data-pipeline and other big data technologies
  • 1+ yr of experience of BI implementation in the Cloud.
  • Exposure in at least one ETL tool.
  • Exposure to cloud Datawarehouse will be a plus.
  • Exposure in at least one reporting tool like Redash/Tableau/similar will be a plus.
  • Familiarity with Linux/Unix scripting
  • Expertise with the tools in Hadoop Ecosystem including Pig, Hive, HDFS MapReduce, Sqoop, Storm, Spark, Kafka, Yarn, Oozie, and Zookeeper.
  • Solid experience building APIs (REST), Java services, or Docker Microservices
  • Experience with data pipelines using Apache Kafka, Storm, Spark, AWS Lambda or similar technologies
  • Experience working with terabyte data sets using relational databases (RDBMS) and SQL
  • Experience using Agile/Scrum methodologies to iterate quickly on product changes, developing user stories and working through backlogs.
  • Experience with Hadoop, MPP DB platform, other NoSQL (Mongo, Cassandra) technologies will be a big plus
Experience Required :

3 to 7 Years

Vacancy :

2 - 4 Hires

Similar Jobs for you

See more recommended jobs