Data Engineer AWS – LATAM

LATAM

About Distillery

Distillery Tech Inc accelerates innovation through an unyielding approach to nearshore software development. The world’s most innovative technology teams choose Distillery to help accelerate strategic innovation, fill a pressing technology gap, and hit mission-critical deadlines. We support essential applications, mobile apps, websites, and eCommerce platforms by placing senior, strategic technical leaders and deploying fully managed technology teams that work intimately alongside our client’s in-house development teams. At Distillery Tech Inc, we’re not here to reinvent nearshore software development, we’re on a mission to perfect it. Distillery Tech Inc is committed to diversity and inclusion. We actively seek to cultivate a workforce that reflects the rich tapestry of perspectives, backgrounds, and experiences present in our society. Our recruitment efforts are dedicated to promoting equal opportunities for all candidates, regardless of race, ethnicity, gender, sexual orientation, disability, age, or any other dimension of diversity.


About the Position

We are looking for a Senior Data Engineer to join our client, a leading company in the automobile e-commerce industry. In this role, you will drive operational excellence across a large-scale AWS-based data platform. You will be responsible for maintaining and optimizing approximately 1,000 Airflow DAGs and AWS Glue jobs that support critical business operations. Your work will directly contribute to enhancing the performance, cost efficiency, and reliability of the company's data infrastructure while partnering with stakeholders to deliver actionable data insights.


Responsibilities

  • Monitor, maintain, and optimize ~1,000 production Airflow DAGs and AWS Glue jobs.
  • Identify and resolve performance bottlenecks; optimize resource usage and reduce execution times.
  • Implement cost optimization strategies across AWS services (Redshift, Glue, S3, compute resources).
  • Enhance pipeline reliability with improved error handling, retry logic, and data validation.
  • Establish and improve SLAs, monitoring, alerting, and observability.
  • Standardize patterns and reduce technical debt across DAGs.
  • Maintain and optimize data architectures using AWS (S3, Redshift, Glue, EMR, Lambda).
  • Improve Redshift performance, data modeling, and cluster efficiency.
  • Manage infrastructure as code and deploy automation processes.
  • Ensure compliance with data security and governance best practices.
  • Build internal tooling and self-service capabilities.
  • Collaborate with analysts, data scientists, and business stakeholders.
  • Translate business needs into technical solutions and models.
  • Document pipelines, schemas, and support ad-hoc analysis.
  • Proactively monitor pipelines and troubleshoot production issues.
  • Implement robust logging, alerting, and CI/CD deployment practices.
  • Participate in on-call rotation if applicable and conduct root cause analysis.



Requirements

  • 3+ years of experience in Python for data engineering (pandas, boto3, SQL libraries).
  • Strong hands-on experience with AWS data stack:
    • Amazon Redshift (performance tuning, data modeling).
    • AWS Glue (ETL, crawlers, data catalog).
    • Apache Airflow / MWAA (DAGs, operators, sensors).
    • S3, Lambda, Step Functions, EMR (preferred).
  • Advanced SQL skills and experience with query optimization.
  • Understanding of cloud infrastructure, networking, IAM, and security.
  • Proficiency with Git and collaborative development workflows.
  • Excellent communication skills and ability to work cross-functionally.
  • Detail-oriented with strong problem-solving skills in a fast-paced environment.


Nice To Have

  • Experience with dbt (data build tool).
  • Familiarity with other orchestration tools (Dagster, Prefect, Step Functions).
  • Exposure to streaming technologies (Kafka, Kinesis, Flink).
  • Knowledge of DataOps/MLOps and CI/CD for data pipelines.
  • AWS certifications (Solutions Architect, Data Analytics, etc.).
  • Experience with data warehousing concepts (Kimball, star schema, SCD).
  • Familiarity with Terraform or CloudFormation.
  • Exposure to observability tools (Monte Carlo, DataDog, Great Expectations).


Why You'll Like Working Here

Join a global team committed to Distillery's core values: Unyielding Commitment, Relentless Pursuit, Courageous Ambition, and Authentic Connection.

100% Remote Work: Enjoy the freedom to work from anywhere while collaborating with a diverse, multinational team. Competitive Compensation: Generous and competitive package in USD, along with a comprehensive benefits plan.

Flexible Hours: Create a schedule that aligns with your life and priorities.

Home Office Setup: Receive all the hardware and software needed to succeed from home. Innovative Workplace: Collaborate with the global Top 1% of talent in a multicultural and dynamic environment.

Focus on Growth: Pursue professional and personal development while contributing your unique talents to a team where you can truly shine!