Peraton-posted 3 months ago
$112,000 - $179,000/Yr
Full-time • Mid Level
Fort Gordon, GA
5,001-10,000 employees
Professional, Scientific, and Technical Services

Peraton seeks a Data Engineer to provide technical contributions to Elasticsearch, Apache NiFi, Apache Kafka, Apache Spark, and Python-based Extract, Transform, Load (ETL) data processing pipelines and data quality monitoring using Cyber and Electromagnetic Spectrum (EMS) data. Location: Fort Gordon, GA.

  • Translate (potentially vague) proposed business processes into corresponding automated data pipelines to ensure the right data is available to supported teams at the right location and format.
  • Create, modify, and automate production-grade ETL pipelines and data monitoring tools like Grafana, Prometheus, or Logstash.
  • Communicate with other data and machine learning engineers about combining compressed, JSON, parquet, database, and data lake-based datasets into a cohesive dataset capable of facilitating business processes.
  • Participate in the software development and deployment using Elasticsearch, Apache NiFi, Apache Kafka, Dask, Prefect, or Pachyderm, Apache Spark, Grafana, Markdown.
  • Minimum of 8 years with BS/BA; Minimum of 6 years with MS/MA. Will consider HS + 12 years.
  • 5+ years of experience, with a focus on experiences that directly relate to deploying and managing Data Science tools in AWS and Azure, and demonstrate practical, hands-on experience with multi-cloud environments.
  • 1+ years of direct technical experience with Apache ecosystem data engineering projects (NiFi, Kafka, Airflow, Spark, Iceberg, etc.).
  • 1+ years of direct technical implementation experience with Python based ETL of JSON, compressed, and columnar datasets.
  • 1+ years of bash or PowerShell scripting.
  • Direct technical experience; translating (potentially vague) proposed business processes into corresponding automated data pipelines; direct technical implementation experienced with creating, modifying, and automating production-grade ETL pipelines and data monitoring tools like Grafana, Prometheus, or Logstash.
  • Comfortable with detailed technical communication with other data and machine learning engineers about combining compressed, JSON, parquet, database, and data lake-based datasets into a cohesive dataset capable of facilitating business processes.
  • Expertise in AWS and Azure Platforms.
  • Must have in-depth knowledge of both AWS and Azure cloud services, with specific skills in implementing and managing services that support DevSecOps workflows, such as container orchestration, serverless computing, and automated deployments.
  • A passion for innovative technology and protecting national defense through responsible software development and deployment.
  • Must have experience with the following tools: Elasticsearch, Apache NiFi, Apache Kafka, Apache Spark, Dask, Prefect or Pachyderm, Grafana, Markdown and GitLab Pages based documentation.
  • Familiarity with at least one programming language preferred.
  • Familiarity with Git preferred.
  • Requires a Top Secret//SCI clearance with the ability to obtain a CI Polygraph and MEAD.
  • U.S Citizenship Required.
  • Experience with Dask, Prefect, Pachyderm, Airflow, or similar Python ecosystem data engineering projects preferred.
  • Heavily subsidized employee benefits coverage for you and your dependents.
  • 25 days of PTO accrued annually up to a generous PTO cap.
  • Eligible to participate in an attractive bonus plan.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service