We are Belvo, an open finance API platform with the bold vision of democratizing access to financial services in Latin America. We enable any financial innovator to access and interpret financial data, as well as initiate payments from their end-users accounts. We're turning the messy complexities of the Latin American financial ecosystem into a modern set of tools to access and interpret data and move money in a seamless and secure way.
We're a highly-technical, passionate, and driven team. We are more than 90 people and our team currently represents 20 nationalities. We have offices in São Paulo and Mexico City – while a large portion of us work remotely.
We are tackling a very stimulating problem: connecting fintech innovators with legacy financial infrastructure. We strive to go beyond the limits of what is possible today and to do so in an elegant and developer-first way.
Since starting our adventure in May 2019, we have raised $71m from the leading VC investors globally.
We're looking for a seasoned Senior Data Engineer to join our Data Platform team. This is a team that has the goal to support data understanding at scale, by architecting and developing the infrastructure to develop data pipelines, moving and transforming complex datasets coming from different sources, and improving data discoverability and data literacy. The ideal candidate is a player who is sought out for technical guidance and can act as an owner for projects across the company. Ideally has been building data infrastructure and is familiar with Data Mesh concepts.
As part of the team, you will be in contact with our stakeholders, ranging from data insights teams of analysts to deeply technical backend product teams to better define and develop the platform, and you will be a central part of the roadmap definition for the team. You will have full ownership of some projects and will have the opportunity to define new data platform products. The current platform uses the latest technologies, like EMR Studio and Apache Iceberg, and as part of the team, you will also be responsible for maintaining and evolving it.
Our platform infrastructure is fully defined with Terraform and we are processing over a thousand events per second. We run daily processes that read over 40 terabytes of data using dbt over Athena and Spark on EMR clusters, everything orchestrated with Dagster. We are moving some of our processes to Streaming processing using Kinesis and Flink.
This position may be for you if:
Amazing if:
Our tech stack:
Our process steps:
Our perks: