Data Engineers (aws / Cloud-Based

Data Engineers (aws / Cloud-Based

HITEC City, Hyderabad

4 years

Azure Cloud ServicesPythonPySparkSQLGitHubAWS GlueAmazon S3AWS LambdaAWS Elastic MapReduce (EMR)AWS Step FunctionsAmazon Relational Database Service (RDS)Amazon AthenaAmazon CloudWatchETL TestingBig Data ProcessingSnowflakeApache AirflowDialectical Behavior Therapy (DBT)Data ModelingTerraform

Job Description:

We are looking for skilled Data Engineers with strong experience in AWS and modern data platforms to build scalable, cloud-native data solutions. As a Data Engineer, you will design, build, and optimize data pipelines and data platforms on AWS Cloud, enabling analytics, reporting, and advanced data use cases. You’ll work with large-scale datasets and modern ETL/ELT frameworks in a cloud environment.

Location- Hyderabad(4 Days WFO)

Experience Required- 3B(Min-4-6 Years)/4A(Min 6-8 Years)

🛠️ Core Technical Skills (Must Have)

🔹 AWS Data Engineering Stack

• Strong experience with AWS Cloud

• Python for data engineering

• PySpark

• SQL (advanced querying & optimization)

• GitHub / version control

🔹 Key AWS Services

Hands-on experience with most of the below:

• AWS Glue

• S3

• Lambda

• EMR

• Step Functions

• RDS

• Athena

• CloudWatch

______________

🔄 Data Engineering & Platform Skills

• Strong understanding of ETL / ELT processes

• Experience with Big Data processing

• Working knowledge of:

o Snowflake

o Apache Airflow

o DBT

o AWS Glue-based pipelines

• Data modeling and performance tuning

______________

☁️ Cloud & DevOps Exposure

Good understanding of:

• Cloud-based data architectures

• CI/CD for data pipelines

• Data migration & data modernization projects

______________

⭐ Good to Have

• Terraform (Infrastructure as Code)

• Exposure to AI tools / Copilot

• DevOps tools (CI/CD, Octopus, etc.)

• Database management knowledge

______________

🎯 Key Responsibilities

• Design and develop scalable data pipelines on AWS

• Build and maintain ETL/ELT workflows

• Optimize data processing using PySpark & SQL

• Work with cloud data lakes and warehouses (S3, Snowflake, Athena)

• Monitor and troubleshoot pipelines using CloudWatch

• Collaborate with analytics, BI, and application teams

• Support data platform modernization initiatives