Data Engineers (aws / Cloud-Based
HITEC City, Hyderabad
4 years
Azure Cloud ServicesPythonPySparkSQLGitHubAWS GlueAmazon S3AWS LambdaAWS Elastic MapReduce (EMR)AWS Step FunctionsAmazon Relational Database Service (RDS)Amazon AthenaAmazon CloudWatchETL TestingBig Data ProcessingSnowflakeApache AirflowDialectical Behavior Therapy (DBT)Data ModelingTerraform
Job Description:
We are looking for skilled Data Engineers with strong experience in AWS and modern data platforms to build scalable, cloud-native data solutions. As a Data Engineer, you will design, build, and optimize data pipelines and data platforms on AWS Cloud, enabling analytics, reporting, and advanced data use cases. You’ll work with large-scale datasets and modern ETL/ELT frameworks in a cloud environment.
Location- Hyderabad(4 Days WFO)
Experience Required- 3B(Min-4-6 Years)/4A(Min 6-8 Years)
🛠️ Core Technical Skills (Must Have)
🔹 AWS Data Engineering Stack
• Strong experience with AWS Cloud
• Python for data engineering
• PySpark
• SQL (advanced querying & optimization)
• GitHub / version control
🔹 Key AWS Services
Hands-on experience with most of the below:
• AWS Glue
• S3
• Lambda
• EMR
• Step Functions
• RDS
• Athena
• CloudWatch
______________
🔄 Data Engineering & Platform Skills
• Strong understanding of ETL / ELT processes
• Experience with Big Data processing
• Working knowledge of:
o Snowflake
o Apache Airflow
o DBT
o AWS Glue-based pipelines
• Data modeling and performance tuning
______________
☁️ Cloud & DevOps Exposure
Good understanding of:
• Cloud-based data architectures
• CI/CD for data pipelines
• Data migration & data modernization projects
______________
⭐ Good to Have
• Terraform (Infrastructure as Code)
• Exposure to AI tools / Copilot
• DevOps tools (CI/CD, Octopus, etc.)
• Database management knowledge
______________
🎯 Key Responsibilities
• Design and develop scalable data pipelines on AWS
• Build and maintain ETL/ELT workflows
• Optimize data processing using PySpark & SQL
• Work with cloud data lakes and warehouses (S3, Snowflake, Athena)
• Monitor and troubleshoot pipelines using CloudWatch
• Collaborate with analytics, BI, and application teams
• Support data platform modernization initiatives

