✨ About The Role
- The role involves designing and building large-scale distributed data systems that power analytics, AI/ML, and business intelligence.
- The candidate will develop batch and streaming solutions to ensure data reliability, efficiency, and scalability across the company.
- Managing data ingestion, movement, and processing through core platforms like Snowflake and real-time streaming systems is a key responsibility.
- The position requires improving data reliability, consistency, and performance to ensure high-quality data for various stakeholders.
- Collaboration with AI researchers, data scientists, product engineers, and business teams to understand data needs and build scalable solutions is essential.
âš¡ Requirements
- The ideal candidate will have over 6 years of experience in designing and building distributed data infrastructure at scale.
- Strong expertise in batch and streaming data processing technologies such as Spark, Flink, Kafka, or Airflow/Dagster is essential.
- A proven track record of impact-driven problem-solving in fast-paced environments is crucial for success in this role.
- The candidate should possess excellent technical communication skills, enabling effective collaboration with both technical and non-technical counterparts.
- A strong sense of engineering excellence, focusing on high-quality, reliable, and performant systems, is necessary.
- Experience mentoring and supporting engineers, fostering a culture of learning and technical excellence, is highly valued.