✨ About The Role
- The role involves building real-time data pipelines to process millions of unstructured financial documents.
- You will be responsible for developing ML-based parsers to intelligently chunk and tag data for indexing.
- The position requires working with cutting-edge data engineering technologies at the AI frontier.
- The job emphasizes a fast-paced environment with a focus on releasing updates early and often.
- You will be part of a team that values integrity, honesty, and continuous learning, inspired by the principles of Warren Buffett.
⚡ Requirements
- The ideal candidate will have at least 3 years of experience deploying production code in a company with a large infrastructure.
- A strong proficiency in Spark, Python, and Postgres is essential for success in this role.
- Familiarity with LLMs and experience in building real-time data pipelines will be highly beneficial.
- Candidates with knowledge of Elastic, Next.js, and TypeScript will have an advantage.
- A collaborative mindset and the ability to work effectively in a small team environment are crucial.