About the position
Flexport is seeking a creative and technically minded Data Engineer to tackle complex data problems in global trade. The role involves developing scalable and innovative data solutions to support internal and customer-facing analytics, as well as providing actionable insights. The ideal candidate should have experience in building scalable data models and setting data standards for large enterprises. They should also possess advanced skills in SQL, experience with data warehouses and pipelines, and proficiency in programming languages. The Data Engineer will play a crucial role in ensuring data quality, driving metadata enrichment, and advocating for development standards and best practices.
Responsibilities
- Architect, build, publish, and maintain performant and reliable data models and pipelines
- Enable self-service data consumption throughout the enterprise
- Support flexible querying and data visualization
- Support advanced analytical and scientific use cases
- Serve as a data steward and subject-matter expert for specific business and technology domains
- Collaborate with upstream data producers and consumers to define successful analytical outcomes
- Drive data quality and enrich metadata context
- Provide feedback to the analytics platform team for tooling improvement
- Develop and evangelize development standards and best practices for data modeling and working with data and tools
- Have 8+ years of work experience in data curation lifecycle or a bachelor's degree in a quantitative field with 5+ years of industry experience
- Possess advanced skills in writing clean, performant, scalable SQL
- Have experience with Snowflake, BigQuery, Redshift, etc.
- Be proficient in schema design and data modeling strategies
- Have experience with dbt or similar tools, Spark-based or similar data pipelines
- Have experience with general-purpose programming languages (e.g., Python, Java, Go)
- Maintain a "compliance first" attitude
- Be familiar with canonical modeling techniques, domain-driven design, and data mesh principles
- Be able to solve ambiguous problems independently
- Be detail-oriented and eager to learn new skills and tools
- Have excellent documentation and communication skills
- Have a passion for high data quality and building scalable and user-friendly data solutions
- Be a team player with a bias for action and willingness to mentor junior developers
- Have a strong sense of ownership
- Collaborate with various organizational stakeholders
- Be familiar with modern BI tools such as Looker, Superset, Metabase
- Work with technologies like Kafka, Snowflake, Astronomer/Dagster, dbt, Amplitude, Fullstory, etc.
- Be part of shaping the future of the global trade industry.
Requirements
- 8+ years of work experience in data curation lifecycle or a bachelor's degree in a quantitative field with 5+ years of industry experience
- Advanced skills in writing clean, performant, scalable SQL
- Advanced experience with Snowflake, BigQuery, Redshift, etc.
- Extensive experience in schema design and data modeling strategies
- Significant experience with dbt or similar tools, Spark-based or similar data pipelines
- Significant experience with general-purpose programming (Python, Java, Go)
- "Compliance first" attitude to ensure regulatory compliance
- Familiarity with canonical modeling techniques, domain-driven design, data mesh principles (preferred)
- Ability to solve ambiguous problems independently
- Detail-oriented and eager to learn new skills and tools
- Ability to write clear, concise documentation and communicate with precision
- Passion for high data quality and building scalable, user-friendly data solutions
- Team player with a bias for action and willingness to mentor junior developers
- Strong sense of ownership
- Collaboration with organizational stakeholders
- Familiarity with modern BI tools (Looker, Superset, Metabase) (preferred)
Benefits
- Competitive salary range: $165,000-$183,000 USD
- Opportunity to work on challenging data problems in global trade
- Access to the most up-to-date information and insights about global supply chains, transportation, and logistics
- Collaboration with Software Engineering teams to develop scalable and innovative data solutions
- Support for internal analytics and customer-facing products and experiences
- Effortlessly accessible intelligence and actionable insights for end users
- Data stewardship and subject-matter expertise in dedicated business and technology domains
- Collaboration with upstream data producers and consumers to define successful analytical outcomes
- Focus on data quality and enriching metadata context
- Feedback provision to the analytics platform team for tooling improvement
- Development standards and best practices for data modeling and working with data and tools
- Advanced skills in writing clean, performant, scalable SQL
- Experience with Snowflake, BigQuery, Redshift, etc.
- Expertise in schema design and data modeling strategies
- Familiarity with dbt and Spark-based data pipelines
- Proficiency in general-purpose programming languages (e.g., Python, Java, Go)
- Compliance-first attitude to ensure regulatory compliance
- Familiarity with canonical modeling techniques, domain-driven design, and data mesh principles
- Ability to solve ambiguous problems independently
- Detail-oriented and eager to learn new skills and tools
- Strong communication and documentation skills
- Passion for high data quality and building scalable and user-friendly data solutions
- Team player with a bias for action and willingness to mentor junior developers
- Strong sense of ownership
- Collaboration with a wide variety of organizational stakeholders
- Familiarity with modern BI tools (e.g., Looker, Superset, Metabase)