Country/Region:  IN
Requisition ID:  26152
Work Model: 
Position Type: 
Salary Range: 
Location:  INDIA - CHENNAI - RNTBCI

Title:  Lead Data Engineer - AWS

Description: 

Area(s) of responsibility

Empowered By Innovation
Birlasoft, a global leader at the forefront of Cloud, AI, and Digital technologies, seamlessly blends domain expertise with enterprise solutions. The company’s consultative and design-thinking approach empowers societies worldwide, enhancing the efficiency and productivity of businesses. As part of the multibillion-dollar diversified CKA Birla Group, Birlasoft with its 12,000+ professionals, is committed to continuing the Group’s 170-year heritage of building sustainable communities.

Role: Lead Data Engineer -AWS
Location: Bangalore /Chennai 
Experience: 5 – 7 Years

Job Profile:
•    Provide estimates for requirements, analyses and develop as per the requirement.
•    Developing and maintaining data pipelines and ETL (Extract, Transform, Load) processes to extract data efficiently and reliably from various sources, transform it into a usable format, and load it into the appropriate data repositories. 
•    Creating and maintaining logical and physical data models that align with the organization's data architecture and business needs. This includes defining data schemas, tables, relationships, and indexing strategies for optimal data retrieval and analysis.
•    Collaborating with cross-functional teams and stakeholders to ensure data security, privacy, and compliance with regulations.
•    Collaborate with downstream application to understand their needs and build the data storage and optimize as per their need. 
•    Working closely with other stakeholders and Business to understand data requirements and translate them into technical solutions. 
•    Familiar with Agile methodologies and have prior experience working with Agile teams using Scrum/Kanban
•    Lead Technical discussions with customers to find the best possible solutions.
•    Proactively identify and implement opportunities to automate tasks and develop reusable frameworks.
•    Optimizing data pipelines to improve performance and cost, while ensuring a high quality of data within the data lake.
•    Monitoring services and jobs for cost and performance, ensuring continual operations of data pipelines, and fixing of defects.
•    Constantly looking for opportunities to optimize data pipelines to improve performance

Must Have: 
•    Hand on Expertise of 4- 5 years in AWS services like S3, Lambda, Glue, Athena, RDS, Step functions, SNS, SQS, API Gateway, Security, Access and Role permissions, Logging and monitoring Services.  
•    Good hand on knowledge on Python, Spark, Hive and Unix, AWS CLI
•    Prior experience in working with streaming solution like Kafka .
•    Prior experience in implementing different file storage types like Delta-lake / Ice-berg. 
•    Excellent knowledge in Data modeling and Designing ETL pipeline.   
•    Must have strong knowledge in using different databases such as MySQL, Oracle and Writing complex queries. 
•    Strong experience working in a continuous integration and Deployment process. 
•    Pyspark,  AWS ,SQL, Kafka

Nice to Have: 
•    Hand on experience in the Terraform, GIT, GIT Actions. CICD pipeline and Amazon Q. 
•    Terraform, GIT, GIT Actions. CICD pipeline , AI