Red Hat-posted 10 days ago
$178,131 - $220,680/Yr
Full-time • Mid Level
Remote • Boston, MA
5,001-10,000 employees

Design, build, and manage scalable data pipelines and enterprise integration workflows using tools like Kafka and Python-based frameworks. Build highly efficient and scalable ETL/ELT processes that support both incremental and full data loads, using frameworks like DBT and orchestrated through Apache Airflow. Telecommuting permitted: work may be performed within normal commuting distance from the Red Hat, Inc. office in Boston, MA. What You Will Do: Develop and maintain Python-based applications for data processing and transformation using libraries such as Pandas and NumPy. Architect and deploy cloud-native data solutions on public cloud platforms, ensuring scalability, security, and performance. Build and manage containerized applications using Red Hat OpenShift, enabling reliable deployment and orchestration across environments. Write and optimize complex SQL queries (CRUD and analytics) for structured and unstructured data across various databases. Integrate machine learning models and advanced analytics into production-grade data pipelines. Implement and maintain CI/CD pipelines based on GitOps methodology, with a focus on automation, testing, and streamlined deployment cycles. Troubleshoot complex system and data issues, minimizing downtime and ensuring data quality and reliability. Follow Agile methodologies and DevOps practices, actively contribute to sprint planning, retrospectives, and delivery cycles. Mentor junior engineers, conduct code reviews, and promote knowledge sharing across the team.

  • Design, build, and manage scalable data pipelines and enterprise integration workflows using tools like Kafka and Python-based frameworks.
  • Build highly efficient and scalable ETL/ELT processes that support both incremental and full data loads, using frameworks like DBT and orchestrated through Apache Airflow.
  • Develop and maintain Python-based applications for data processing and transformation using libraries such as Pandas and NumPy.
  • Architect and deploy cloud-native data solutions on public cloud platforms, ensuring scalability, security, and performance.
  • Build and manage containerized applications using Red Hat OpenShift, enabling reliable deployment and orchestration across environments.
  • Write and optimize complex SQL queries (CRUD and analytics) for structured and unstructured data across various databases.
  • Integrate machine learning models and advanced analytics into production-grade data pipelines.
  • Implement and maintain CI/CD pipelines based on GitOps methodology, with a focus on automation, testing, and streamlined deployment cycles.
  • Troubleshoot complex system and data issues, minimizing downtime and ensuring data quality and reliability.
  • Follow Agile methodologies and DevOps practices, actively contribute to sprint planning, retrospectives, and delivery cycles.
  • Mentor junior engineers, conduct code reviews, and promote knowledge sharing across the team.
  • Bachelor's degree (U.S. or foreign equivalent) in Computer Science, Information Systems or related field and five (5) years of experience in the job offered or related role OR Master's degree (U.S. or foreign equivalent) in Computer Science, Information Systems or related field and three (3) years of experience in the job offered or related role.
  • Must have three (3) years of experience with: designing, building, and maintaining enterprise-scale data processing pipelines and integration workflows using distributed orchestration frameworks and advanced data processing tools; leveraging advanced SQL functions, such as window functions, recursive queries, and in-database analytics to perform complex analytical operations on large-scale datasets; and developing extensible Python frameworks focused on microservices architecture, data APIs, and modular design patterns to deliver scalable, reusable, and maintainable enterprise solutions.
  • Must have one (1) year of experience with: implementing advanced CI/CD workflows using GitOps methodologies, and incorporating automated testing, artifact management, and multi-environment deployment strategies on Red Hat OpenShift; and deploying and administering containerized applications on Red Hat OpenShift, ensuring production-grade configurations for security, autoscaling, secrets and configuration management, and integration with external systems.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service