About The Position

This position will be responsible for serving as the primary data solutions architect for a data services/visualization platform for a large federal agency. The architect will lead the design of data lakes, data warehouses, and visualization services and work with data/cloud engineers to design data pipelines to ingest, clean and transform both batch and streaming data.

Requirements

  • Data Architect Experience (10+ years).
  • Designing Bronze (raw), Silver (cleansed), and Gold (aggregated) layers using Delta Lake.
  • Strong experience with Databricks, Spark, Python, SQL, and Cloud platforms (AWS).
  • Building automated ingestion and transformation pipelines (Databricks Workflows, Delta Live Tables).
  • Design normalized relational schemas (3NF) for transactional systems, translating business requirements into logical and physical models that balance data integrity, referential constraints, and query performance across heterogeneous source systems.
  • Architect star and snowflake schemas with appropriate fact table grain, conformed dimensions, and slowly changing dimension strategies (Type 1, 2, 3) to support enterprise analytical workloads and cross-functional reporting requirements.
  • Implementing data lineage, metadata management, and validation to ensure accuracy.
  • Tuning query performance, managing data warehouse costs, and optimizing storage.
  • Data Architecting a data lake and/or data warehouse model including defining the use of standard conventions using cloud components (AWS S3, AWS Redshift, AWS Aurora Postgres).
  • Creating logical and physical data models for data repositories containing a number of heterogenous sources.
  • Creating work streams for the development team using an Agile framework; developing a roadmap Initiatives and Epics and work with a scrum master to define stories with clear acceptance criteria.
  • Optimizing data warehouses for high performance for varying types of use cases (high frequency loads, large loads, high frequency queries).
  • Implementing various archiving and partitioning approaches to optimize data ingestion, data storage costs and meet data retention requirements.
  • Software Development, Leadership and Customer Experience (5+ years).
  • The Data Architect should be able to elicit requirements from the client including complex process flows and design data models that will support the development of the application. Ability to communicate effectively with senior leadership including translating complex technical concepts into specific and understandable technology recommendations.
  • Demonstrate an understanding of cloud components in either AWS or Azure to design a cloud-native technical solution for the application e.g. the use of AWS Lambdas, SNS/queues, API Gateway, Cloudfront, CloudFormation. Will lead the data model and process flow controller design to support an efficient and intuitive user interface view for the application.
  • The Data Architect will be responsible for creating work streams for the development team to build out the functional infrastructure in Data bricks and create the application using an Agile framework; developing a roadmap Initiatives and Epics to accomplish the development plan within an Agile framework and work with a scrum master to define stories with clear acceptance criteria.
  • Experience designing data models for transactional applications.
  • Experience working with Data bricks or other data engineering, machine learning, data analytics platform to build a data Lakehouse and data warehouse solution.
  • Experience with data replication, high availability, and data migrations tools and techniques.

Nice To Haves

  • Use of integrating visualization tools, such as Tableau, with data lakes and data warehouses).
  • AWS Solution Architect Certification, Databricks certification.
  • Bachelor's Degree in a related area or equivalent experience.

Responsibilities

  • Leading the design of data warehouses, data lakes and datamarts including both the logical and physical models.
  • Assist in soliciting and assessing requirements for the data warehouses/data lakes among data consumers/analysts, visualization specialists, programmers, and engineers.
  • Facilitating change control processes for the data model within the repositories including integrating with CI/CD pipelines.
  • Designing and implementing a metadata repository to be used for data discovery/catalog services across the enterprise.
  • Provide technical guidance and mentorship to other members of the technical team.

Benefits

  • Company-supported medical, dental, vision, life, STD, and LTD insurance
  • Benefits include 11 federal holidays and PTO
  • Eligible employees may receive performance-based incentives in recognition of individual and/or team achievements.
  • 401(k) with company matching
  • Flexible Spending Accounts for commuter, medical, and dependent care expenses
  • Tuition Assistance
  • Charitable Contribution matching
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service