Country/Region:  IN
Requisition ID:  34449
Work Model: 
Position Type: 
Salary Range: 
Location:  INDIA - BENGALURU - BIRLASOFT OFFICE

Title:  Data Engineer-PySpark

Description: 

Area(s) of responsibility

About Birlasoft:
Birlasoft, a global leader at the forefront of Cloud, AI, and Digital technologies, seamlessly blends domain expertise with enterprise solutions. The company’s consultative and design-thinking approach empowers societies worldwide, enhancing the efficiency and productivity of businesses. As part of the multibillion-dollar diversified CKA Birla Group, Birlasoft with its 12,000+ professionals, is committed to continuing the Group’s 170-year heritage of building sustainable communities.

•    About the Job – We are seeking a highly skilled and motivated PySpark professional to join our team. The successful candidate will be responsible for developing, maintaining, and testing big data solutions. They will be involved in the design of data processing using PySpark and other big data technologies. The role requires a deep understanding of data processing, data structure optimization, and machine learning algorithms. The PySpark professional will work closely with our data science team to enable quick and reliable data analysis and ensure the highest level of data quality..
1.    Job Title – Data Engineer-Pyspark
2.    Location: Bangalore/Hyderabad
3.    Educational Background – Bachelor's degree in Computer Science, Information Technology, or related field.
Key Responsibilities –Develop, test, and maintain data pipelines using PySpark.
•    Analyze large, complex data sets to meet functional and non-functional business requirements.
•    Collaborate with data scientists and architects on several projects.
•    Optimize data delivery and develop monitoring for performance and troubleshooting.
Required Skills
•    Proficiency in PySpark and Python programming language.
•    Strong understanding of Big Data Analytics and Hadoop.
•    Experience with SQL databases and knowledge of query languages.
•    A Bachelor's degree in Computer Science, Information Systems, or a related field, with a strong background in data processing and machine learning algorithms.
Preferred Skills:Familiarity with machine learning algorithms. Experience with data visualization tools like Tableau or PowerBI.
•    Knowledge of cloud platforms like AWS, Google Cloud, or Azure.
•    Understanding of data warehousing and ETL techniques.
•    Experience with other programming languages like Java or Scala.
•    Familiarity with NoSQL databases like MongoDB or Cassandra.
•    Knowledge of data structures and algorithms.
•    Experience with other big data technologies like Hive, Pig, or HBase.
•    Understanding of distributed systems.
•    Familiarity with Agile/Scrum methodologies.

Required Experience: 3-5yrs
Skills Required – Pyspark