Senior/Lead Data Engineer - AWS & Airflow

India | Full-time | Partially remote

Apply by: No close date
Apply

 

About US:-

We turn customer challenges into growth opportunities.

Material is a global strategy partner to the world’s most recognizable brands and innovative companies. Our people around the globe thrive by helping organizations design and deliver rewarding customer experiences.

We use deep human insights, design innovation and data to create experiences powered by modern technology. Our approaches speed engagement and growth for the companies we work with and transform relationships between businesses and the people they serve.

Srijan, a Material company, is a renowned global digital engineering firm with a reputation for solving complex technology problems using their deep technology expertise and leveraging strategic partnerships with top-tier technology partners. Be a part of an Awesome Tribe

 

Job Title: Senior/Lead Data Engineer

Experience: 4+ years


About the Role

We are seeking a Senior/Lead Data Engineer to join our fast-growing data engineering team supporting a large-scale healthcare and nonprofit analytics program for a US-based client. This role will focus on building and optimizing cloud-native data pipelines using AWS services to support analytics, reporting, and stakeholder decision-making.

You will work closely with architects, analysts, and product teams to integrate disparate data sources, ensure data quality, and deliver business-critical data solutions at scale.


Key Responsibilities

  • Design and implement robust, scalable ETL/ELT pipelines using AWS-native tools
  • Ingest and transform data from multiple sources into S3, applying schema discovery via AWS Glue Crawlers
  • Develop and orchestrate workflows using Apache AirflowAWS Step Functions, and Lambda functions
  • Build and optimize data models in Amazon Redshift for analytics consumption
  • Manage and enforce IAM-based access control, ensuring secure data practices
  • Write clean, modular, and reusable code in PySpark and SQL for large-scale data processing
  • Implement monitoring, alerting, and CI/CD pipelines to improve deployment efficiency and reliability
  • Work closely with business stakeholders and analysts to understand data requirements and deliver meaningful insights
  • Participate in code reviews and knowledge-sharing activities across teams.
  • Understands scrum and comfortable working in an Agile environment.

Required Skills

  • 4+ years of experience as a Data Engineer, with at least 3+ years working in cloud-native environments (preferably AWS)
  • Hands-on experience with S3RedshiftGlue (ETL & Crawlers)LambdaStep Functions, and Airflow
  • Strong programming skills in PySpark and SQL
  • Experience designing and implementing data lakesdata warehouses, and real-time/near-real-time pipelines
  • Familiarity with DevOpsCI/CD pipelines, and infrastructure as code tools (e.g., Git, CloudFormation, Terraform)
  • Understanding of data governancedata security, and role-based access control in cloud environments
  • Strong problem-solving skills and ability to work independently as well as collaboratively
  • Excellent written and verbal communication skills

Nice to Have

  • Experience working in domains such as nonprofit, healthcare, or campaign marketing
  • Familiarity with AWS Notebooks, Athena, and CloudWatch
  • Exposure to data observability tools, testing frameworks, or event-driven architectures
  • Experience mentoring junior engineers or leading small teams

Why Work With Us?

  • Opportunity to work on meaningful data problems impacting real-world decision-making
  • Collaborative and supportive work culture with flexible remote working options
  • Emphasis on innovation, ownership, and growth
  • Competitive compensation and career advancement opportunities