Data Engineer

ClaspMcDonald, OH
$100,000 - $125,000Hybrid

About The Position

We are seeking a Data Engineer to build & operate reliable data pipelines that ingest and make accessible to our team and partners the data on hundreds of millions of dollars worth of student loans, educational enrollment statuses and employment data. The ideal candidate has both strong programming experience in python and is passionate about both high performance data pipelines as well as the analytics that they fuel. At Clasp we run a DevOps culture where the engineers have full ownership of the code they write & the infrastructure on which it runs. Candidates should be enthused about making substantial contributions to the architecture driving the product roadmap and the Stride business, and achieving tremendous personal growth with us along the way! Our modern Data Technology Stack consists primarily of Airflow, dbt, Postgresql, Superset, BigQuery, Python and SQL

Requirements

  • 3+ years of experience as a data engineer, or equivalent experience building data pipelines
  • 2+ years of experience with elements of the following technologies: Data Analytics: Airflow, DBT
  • Business Intelligence Systems: Tableau, Looker, Sisense, Apache Superset, etc.
  • Languages: Python, SQL
  • Databases: SQL; NoSQL a bonus
  • Strong communication and collaboration skills, with the ability to effectively communicate the complexities of technical programs to both technical and nontechnical stakeholders
  • Desire to mentor and collaborate with other members of the team
  • Willingness to roll your sleeves up to rapidly acquire competencies in a wide range of technical disciplines
  • Bachelor’s degree in Computer Science, Software Engineering, Information Systems, or equivalent experience

Nice To Haves

  • Bonus – Data Warehousing: Snowflake, BigQuery, etc.
  • Bonus -- Cloud Infrastructure: AWS or Google Cloud Experience
  • Bonus -- DevOps: Experience with modern cloud and container tooling such as Docker, Kubernetes, Terraform, etc.

Responsibilities

  • Build and maintain scalable data pipelines that ingest and transform financial, education, and employment data
  • Ensure data is reliable, timely, and accessible across internal teams and external partners
  • Proactively identify and resolve data quality issues, including upstream dependencies
  • Partner with stakeholders in finance, operations and business development departments to ensure that their data sets are reliably ingested into our data warehouse and their questions or reports can be answered programmatically
  • Ensure via automated testing and edge case handling that we can detect any errors with upstream data or data processing and that all reports contain the data as expected
  • Develop processes to anonymize and protect sensitive data across environments
  • Support the configuration and optimization of data warehouse, storage, and compute resources
  • Partner in managing infrastructure as code (e.g., Terraform) to ensure reproducibility and scalability
  • Monitor performance and cost efficiency of data workloads, identifying opportunities for optimization
  • Contribute to improving the reliability, scalability, and observability of our data platform

Benefits

  • Modern tech stack with the ability to have impact to many different user personas - recruiters, students and more
  • High autonomy working in a highly collaborative team
  • Can grow into a more formal people management role – expectation though is to be very hands on
  • Step into being a 10x engineer and ride the AI wave with a team doubling down on how this technology allows us to focus on harder problems and solve for our customer’s needs without compromising on quality
  • attractive equity component as part of our compensation package, providing an opportunity for eligible employees to share in the success and growth of our company
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service