Design and build scalable, reliable data infrastructure and pipelines. Mentor junior and mid-level engineers through code reviews, coaching, and knowledge sharing. Identify and implement process improvements (automation, optimization, monitoring). Collaborate with valuation professionals to understand the business needs and requirements. Lead complex technical implementation work, ensuring performance and quality standards. Collaborate with global teams to deliver fault-tolerant, high-quality data solutions. Drive engineering excellence by enforcing coding standards, automated unit testing, and CI/CD best practices. Lead and participate in global trainings to enhance the understanding of the infrastructure by valuation professions. Contribute to architecture discussions and proof-of-concepts, providing input without direct ownership of roadmap.
Bachelor’s degree in computer science, Engineering, or related field (master’s preferred). Equivalent experience considered. 5+ years of experience in data engineering or software engineering, with at least 2 years in a senior or lead role. Proven experience building and managing ETL/ELT pipelines Advanced proficiency with Azure, AWS, and Databricks (with a focus on data services) Deep knowledge of Python, Spark ecosystem (PySpark, Spark SQL), and relational databases Experience building REST APIs, Python SDKs, libraries, Spark-based data services and tools like FastAPI, Pydantic, Polars, Pandas, Delta Lake, Docker, Kubernetes Understanding of Lakehouse architecture, Medallion architecture, and data governance Experience with pipeline orchestration tools (e.g., Airflow, Azure Data Factory) Strong communication skills and ability to work cross-functionally with international teams Skilled in data profiling, cataloging, and mapping for technical data flows Understanding of API product management principles, including lifecycle strategy, documentation standards, and versioning Experience with CI/CD, Git, and DevOps best practices
Familiarity with financial statements and valuation methodologies (income approach, market approach, etc.). Broader cloud architecture awareness (networking, security, cost optimization). Experience tuning complex SQL/Spark queries and pipelines for performance. Hands-on experience building Lakehouse solutions using Azure Databricks, ADLS, and/or Snowflake as part of a modern data platform architecture.