Lead - Big Data Developer Job in Sutherland Global Services

Lead - Big Data Developer

Apply Now
Job Summary Job Description:

Sutherland is seeking an analytical person to join us as Big Data Developer to be responsible for Predictive modelling and provide insights from the data with hands on experience in Big Data tools, Hadoop, Hive, Hbase, Impala and ETL techniques and frameworks. We are a group of dynamic and driven individuals. If you are looking to build a fulfilling career and are confident you have the skills and experience to help us succeed, we want to work with you!

Data Scientists in this role get to:

Keep management updated: Relay vital information in the form of timely and accurate reports
Impact the bottom line: Produce solid and effective strategies based on accurate and meaningful data reports and analysis and/or keen observations
Define Sutherlands reputation: Oversee and manage performance and service quality to guarantee customer satisfaction;
Strengthen relationships: Establish and maintain communication with clients and/or team members; understand needs, resolve issues, and meet expectations
Take the lead: Think critically and identify opportunities for improvement, then work with global SMEs to integrate efficiencies into daily activities



Qualifications:

To succeed in this position, you must:

Experience: 6+ years
This technical position will be responsible for development and on-going support of an enterprise data warehouse, data marts and supporting systems/applications.
Implementing ETL process with multiple sources systems such as SQL, Oracle, Files, Mail etc.
Develop efficient pig and hive scripts with joins on datasets using various techniques.
Assess the quality of datasets for a hadoop data lake.
Apply different HDFS formats and structure like Parquet, Avro, etc. to speed up analytics.
Monitoring performance and advising any necessary infrastructure changes
Design and implement schemas of Hive and HBase within HDFS.
Assign schemas and create Hive tables.
Fine tune hadoop applications for high performance and throughput.
Troubleshoot and debug any hadoop ecosystem run time issues.
Support QA (resolve issues and release fixes) UAT and production support when required.
Deployment and integration testing of developed components in Development and Test environments.

Key competencies:
Knowledge of hadoop ecosystem and its components HBase, Pig, Hive, Sqoop, Flume, Oozie, etc.
Know-how on the java essentials for hadoop.
Know-how on basic Linux administration
Knowledge of scripting languages like Python or Perl.
Good knowledge of concurrency and multi-threading concepts
Experience with Spark
Experience with integration of data from multiple data sources
Experience with Cloudera 6.0 and above versions
Analytical and problem-solving skills.
Good Communication skills

Experience Required :

Fresher

Vacancy :

2 - 4 Hires

Similar Jobs for you

See more recommended jobs