Data Integration Engineer

Engineering /
Full-time /
We seek a dynamic and results-driven professional to join our team as a Data Integration Engineer. As a rapidly growing player in the industry, we are looking for an individual with a strong background on data warehousing and pipeline development. You will play a pivotal role in designing, developing, and maintaining our data infrastructure, ensuring the accuracy and efficiency of our data processes. This role reports directly to our Director of Business Intelligence.


    • Design, develop, maintain, and document Python-based data processing and ETL jobs ensuring data is complete, accurate, and that system failures are minimized and resolved efficiently
    • Develop and maintain a state-of-the-art Data Warehouse, ensuring it meets the evolving needs of the business
    • Utilize Apache Airflow to orchestrate and manage ETL workflows, ensuring optimal performance and reliability
    • Adhere to coding standards, participate in code reviews, and assist with controlled releases as part of best practice workflows
    • Proficient in API integrations with third-party software to retrieve and integrate data seamlessly into the system
    • Ensure data pipelines are optimized to minimize costs, while also being efficient enough to meet the reporting needs of the business
    • Ensure that requirements are captured effectively and efficiently, translating business needs into technical solutions
    • Communicate results of technical projects to both technical and non-technical users, ensuring clear understanding and alignment with business goals.


    • Minimum of 5 years of experience in a Data Engineering role, with a strong focus on data warehousing and pipeline development
    • Expertise in writing complex SQL and programming in Python
    • Experience with dbt, Apache Airflow, and Google BigQuery, demonstrating a deep understanding of these technologies
    • Proficient in developing and integrating APIs with third-party software to facilitate data exchange and workflow automation
    • Experience in building and maintaining robust data processing pipelines and ETL/ELT processes
    • In-depth knowledge of data warehousing methodologies and hands-on experience with major Relational Database systems
    • Experience in Event Streaming a plus
    • Demonstrated ability to work effectively in a distributed team environment, with a strong sense of ownership, urgency, and drive
    • Comfortable with Git and experienced in conducting code reviews and adhering to best coding practices
    • Strong communication skills and the ability to document all code and features for maintainability and future reference
    • A Master’s Degree in Engineering, Computer Science, or related field, or equivalent work experience. is an Equal Opportunity / Affirmative Action employer committed to diversity in the workplace. All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, age, national origin, disability, protected veteran status, gender identity or any other factor protected by applicable federal, state, or local laws.