Qualification - BE, B.Tech, MCA
Job Responsibilities:
Design, build, and maintain data pipelines to ingest, process, and analyze large datasets using PySpark.
Work on Data Warehouse and Data Lake solutions to manage structured and unstructured data.
Develop and optimize complex SQL queries for data extraction and reporting.
Leverage AWS cloud services such as S3, EC2, EMR, Athena, and Redshift for data storage, processing, and analytics.
Collaborate with cross-functional teams to ensure the successful delivery of data solutions that meet business needs.
Monitor data pipelines and troubleshoot any issues related to data integrity or system performance.
Required Skills:
8 years of experience in data engineering or related fields.
In-depth knowledge of Data Warehouses and Data Lakes.
Proven experience in building data pipelines using PySpark.
Strong expertise in SQL for data manipulation and extraction.
Familiarity with AWS cloud services, including S3, EC2, EMR, Athena, Redshift, and other cloud computing platforms.
Preferred Skills:
Python programming experience is a plus.
Experience working in Agile environments with tools like JIRA and GitHub.