Job Description
- Design, develop, and maintain scalable data pipelines and ETL processes using PySpark, AWS Glue, and Redshift.
- Implement data warehousing solutions on AWS Redshift to support analytics and reporting.
- Optimize data workflows to ensure high performance, reliability, and scalability.
- Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver solutions.
- Develop and maintain scripts and automation using Python, PySpark, and AWS Glue.
- Monitor and troubleshoot data pipelines, ensuring data quality and integrity.
- Implement best practices for data management, including data security and governance.
- Stay updated with the latest trends and best practices in data engineering and AWS technologies.
- Create and maintain comprehensive documentation for all data engineering activities.
Good to have:
- Athena, Redshift, AWS DynamoDB, Pyspark.
- Languages: Python & SQL.
- AWS Services: S3,Lambda, CloudFormation, EventBridge, Cloud Watch, IAM, Glue or EMR Orchestration tools- Step function Or Airflow Or Glue workflow.
Role: Data Engineer
Industry Type: IT Services & Consulting
Department: Software & QA
Employment Type: Full Time, Permanent
Role Category: Advertising & Creative
Location: Hyderabad, Chennai, Bengaluru