Data Engineer

BounteousDallas, TX
$110,000 - $125,000

About The Position

Bounteous is a premier end-to-end digital transformation consultancy dedicated to partnering with ambitious brands to create digital solutions for today’s complex challenges and tomorrow’s opportunities. With uncompromising standards for technical and domain expertise, we deliver innovative and strategic solutions in Strategy, Analytics, Digital Engineering, Cloud, Data & AI, Experience Design, and Marketing. Our Co-Innovation methodology is a unique engagement model designed to align interests and accelerate value creation. Our clients worldwide benefit from the skills and expertise of over 4,000+ expert team members across the Americas, APAC, and EMEA. By partnering with leading technology providers, we craft transformative digital experiences that enhance customer engagement and drive business success. Information Security Responsibilities Promote and enforce awareness of key information security practices, including acceptable use of information assets, malware protection, and password security protocols Identify, assess, and report security risks, focusing on how these risks impact the confidentiality, integrity, and availability of information assets Understand and evaluate how data is stored, processed, or transmitted, ensuring compliance with data privacy and protection standards (GDPR, CCPA, etc.) Ensure data protection measures are integrated throughout the information lifecycle to safeguard sensitive information Responsibilities: Pipeline Migration Logic & Scheduling: Refactoring and migrating extraction logic and job scheduling from legacy frameworks to the new Lakehouse environment. Data Transfer: Executing the physical migration of underlying datasets while ensuring data integrity. Stakeholder Engagement: Acting as a technical liaison to internal clients, facilitating "handoff and sign-off" conversations with data owners to ensure migrated assets meet business requirements Consumption Pattern Migration Code Conversion: Translating and optimizing legacy SQL and Spark-based consumption patterns (raw and modeled) for compatibility with Snowflake and Iceberg. Usage analysis: Understand usage patterns to deliver the required data products. Stakeholder Engagement: Acting as a technical liaison to internal clients, facilitating "handoff and sign-off" conversations with data owners to ensure migrated assets meet business requirements. Data Reconciliation & Quality A rigorous approach to data validation is required. Candidates must work with reconciliation frameworks to build confidence that migrated data is functionally equivalent to that already used within production flows.

Requirements

  • Education: Bachelor’s or Master’s degree in Computer Science, Applied Mathematics, Engineering, or a related quantitative field.
  • Experience: Minimum of 5 years of professional "hands-on-keyboard" coding experience in a collaborative, team-based environment. Ability to trouble shoot (SQL) and basic scripting experience.
  • Languages: Professional proficiency in Python or Java.
  • Methodology: Deep familiarity with the full Software Development Life Cycle (SDLC) and CI/CD best practices & K8s deployment experience.
  • Core Data Engineering Competencies: Candidates must demonstrate a sophisticated understanding of the following modeling concepts to ensure data correctness during reconciliation:
  • Temporal Data Modeling: Managing state changes over time (e.g., SCD Type 2).
  • Schema Management: Expertise in Schema Evolution (Ref: Iceberg Apache) and enforcement strategies.
  • Performance Optimization: Advanced knowledge of data partitioning and clustering.
  • Architectural Theory: Balancing Normalization vs. Denormalization and the strategic use of Natural vs. Surrogate Keys.
  • Technical Stack Requirements:
  • Extraction & Logic: Kafka, ANSI SQL, FTP, Apache Spark
  • Data Formats: JSON, Avro, Parquet
  • Platforms: Hadoop (HDFS/Hive), Snowflake, Apache Iceberg, Sybase IQ
  • Candidate will also need to work with our internal data management platform, and must have an aptitude for learning new workflows and language constructs is essential.

Responsibilities

  • Refactoring and migrating extraction logic and job scheduling from legacy frameworks to the new Lakehouse environment.
  • Executing the physical migration of underlying datasets while ensuring data integrity.
  • Acting as a technical liaison to internal clients, facilitating "handoff and sign-off" conversations with data owners to ensure migrated assets meet business requirements
  • Translating and optimizing legacy SQL and Spark-based consumption patterns (raw and modeled) for compatibility with Snowflake and Iceberg.
  • Understand usage patterns to deliver the required data products.
  • Acting as a technical liaison to internal clients, facilitating "handoff and sign-off" conversations with data owners to ensure migrated assets meet business requirements.
  • Candidates must work with reconciliation frameworks to build confidence that migrated data is functionally equivalent to that already used within production flows.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service