Data & Analytics Engineer, AiDP
Location
Austin, TX
Posted
Today
The Developer Experience Platform team is building the next generation of AI-powered tools that accelerate how applications are developed across Apple. We are looking for a Data & Analytics Engineer to help design, build, and scale the data foundation that powers this platform. In this role, you will develop robust data pipelines and analytics systems that enable AI agents, autonomous workflows, and data-driven insights—directly impacting how software is built at scale.
As a hands-on engineer, you will:
- Design, build, and maintain scalable data pipelines and ELT workflows to support AI and analytics use cases
- Develop clean, reliable, and well-modeled datasets for both batch and real-time consumption
- Partner closely with AI/ML engineers and platform teams to deliver high-quality data for model training, inference, and agent workflows
- Implement data quality, observability, and monitoring systems to ensure trust and reliability across pipelines
- Build and optimize data models in modern cloud data warehouses (e.g., Snowflake, BigQuery, Databricks)
- Use tools like DBT to create modular, testable, and well-documented transformation layers
- Orchestrate and manage workflows using tools such as Airflow, Prefect, or Dagster
- Optimize pipelines and queries for performance, scalability, and cost efficiency
- Contribute to the design of the data architecture supporting AI agents and autonomous workflows
- Enable self-service analytics and reporting for engineering and product teams
-
Collaborate across teams to define and implement best practices for data engineering in an AI-first platform
-
3+ years of hands-on experience in data engineering, analytics engineering, or a related role in a production environment
- Proficiency in Python and SQL, including pipeline development, automation, and performance optimization
- Hands-on experience with cloud data warehouses (e.g., Snowflake, BigQuery, or Databricks)
- Experience implementing monitoring, logging, and observability for data pipelines
- Experience with data modeling
-
B.S. in Computer Science or similar or equivalent industry experience
-
Experience building AI/LLM-powered data pipelines, including RAG systems and integrations with APIs such as OpenAI or Anthropic
- Experience with real-time/streaming data systems such as Apache Kafka, Flink, or Spark Structured Streaming
- Experience with workflow orchestration tools such as Airflow, Prefect, or Dagster
- Knowledge of MLOps workflows, including feature engineering, model deployment, and monitoring (e.g., MLflow, Vertex AI)
- Experience with data quality, governance, and lineage tools (e.g., Great Expectations, Monte Carlo)
- Experience building and maintaining ELT pipelines using DBT
- Experience building dashboards and analytics using tools like Tableau, Looker, or Power BI
- Working knowledge of cloud platforms (AWS, GCP, or Azure) and associated data services (e.g., S3, Glue, Dataflow)
Stay informed about the latest analytics engineering opportunities. Subscribe to our weekly newsletter.