Data Engineer Sr. - TMD

Sequoia Connect
16hRemote

About The Position

Our client represents the connected world, offering innovative and customer-centric information technology experiences, enabling Enterprises, Associates, and Society to Rise™. They are a USD 6 billion company with 163,000+ professionals across 90 countries, helping 1279 global customers, including Fortune 500 companies. They focus on leveraging next-generation technologies, including 5G, Blockchain, Metaverse, Quantum Computing, Cybersecurity, Artificial Intelligence, and more, on enabling end-to-end digital transformation for global customers. Our client is one of the fastest-growing brands and among the top 7 IT service providers globally. Our client has consistently emerged as a leader in sustainability and is recognized amongst the ‘2021 Global 100 Most sustainable corporations in the World by Corporate Knights. We are currently searching for a Data Engineer Sr.: Responsibilities The Data Engineer is responsible for designing, building, and maintaining scalable data pipelines and infrastructure to support data analytics, reporting, and business intelligence. This role ensures data is accessible, reliable, and optimized for performance across various systems.

Requirements

  • Proficiency in SQL and experience with relational databases (e.g., Oracle, MySQL, SQL Server).
  • Strong programming skills in Python, PL/SQL, Java, or Scala.
  • Experience with big data technologies (e.g., Hadoop, Spark, Databricks) and cloud platforms (AWS, Azure, GCP).
  • Hands-on experience with OpenShift or other container orchestration platforms (e.g., Kubernetes).
  • Knowledge of data warehousing concepts and tools (e.g., Snowflake, Redshift, BigQuery).
  • Familiarity with workflow orchestration tools (e.g., Airflow, Luigi).
  • Understanding of data governance, security, and compliance.
  • Advanced Oral English.
  • Native Spanish.

Nice To Haves

  • Experience with streaming data (Kafka, Kinesis).
  • Background in DevOps practices for data pipelines.
  • Knowledge of machine learning workflows and integration with data pipelines.

Responsibilities

  • Design, develop, and maintain ETL/ELT pipelines for ingesting and transforming data from multiple sources.
  • Build and optimize data models for analytics and reporting.
  • Implement and manage data storage solutions (e.g., relational databases, data lakes, cloud storage).
  • Ensure data quality, integrity, and security across all systems.
  • Collaborate with data scientists, analysts, and business teams to understand requirements and deliver solutions.
  • Monitor and improve data pipeline performance and troubleshoot issues.
  • Stay updated with emerging technologies and best practices in data engineering and cloud platforms.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service