Austin, TX, USA

Description

We are seeking a talented Azure Data Engineer to design, build, and maintain scalable data infrastructure in Microsoft Azure. The ideal candidate will have deep expertise in Azure Data Services and hands-on experience building data pipelines, transforming datasets, and enabling business insights. Join us to work on cutting-edge cloud data projects across various industries.

Requirements

  • Bachelor’s Degree in Computer Science, Information Systems, or a related field.
  • 3+ years of experience in data engineering, with a focus on Azure technologies.
  • Proficient in Azure Data Factory, Synapse Analytics, Databricks, Azure SQL DB, and Azure Data Lake.
  • Strong programming skills in SQL, Python, or Scala.
  • Experience with big data tools like Spark or Hive.
  • Familiar with Azure DevOps, CI/CD pipelines, and infrastructure-as-code (e.g., ARM/Bicep).
  • Understanding of data modeling, data governance, and data security.
  • Excellent communication, analytical, and problem-solving skills.

Bonuses

  • Performance-based annual/quarterly bonuses.
  • Employee referral bonus program.
  • Signing bonus (for eligible candidates).

Benefits

  • Medical, Dental, and Vision Insurance.
  • 401(k) with employer match.
  • Generous PTO, Holidays, and Sick Leave.
  • Remote work flexibility.
  • Ongoing learning and certification support.
  • Relocation assistance (if applicable).

Responsibilities

  • Design and implement robust data pipelines using Azure Data Factory, Synapse Analytics, and Databricks.
  • Develop, manage, and optimize ETL/ELT processes for structured and unstructured data.
  • Work closely with data analysts, architects, and business stakeholders to understand data requirements.
  • Ensure high data quality, integrity, and governance.
  • Implement data lake and data warehouse solutions in Azure.
  • Optimize performance and cost of data storage and compute resources.
  • Develop CI/CD pipelines for data workflows using tools like Azure DevOps.
  • Monitor, troubleshoot, and resolve issues related to data pipeline performance and availability.