Dataflow Software Engineer, Distributed Analytics - Fully Cleared

Intelliforce-IT Solutions GroupColumbia, MD
1d$179,000 - $237,000Onsite

About The Position

At Intelliforce, we build mission-aligned technology that enables real-time insight, resilient analytics, and operational confidence across the Intelligence Community. This role supports the mission by designing and optimizing dataflow and distributed processing systems that power large-scale analytics. You’ll work at the intersection of data engineering and software development, helping transform raw data into actionable intelligence while operating in high-trust, high-impact environments. This position is ideal for an engineer who thrives in Linux environments, enjoys working with large datasets, and wants their work directly tied to mission outcomes.

Requirements

  • Clearance: Active Top Secret Clearance with Full Scope Polygraph (required)
  • Citizenship: Must be a U.S. Citizen
  • Master’s degree in Computer Science or related discipline and 5+ years of software engineering experience
  • OR
  • Bachelor’s degree in Computer Science or related discipline and 7+ years of software engineering experience
  • OR
  • 9+ years of software engineering experience in programs of similar scope, type, and complexity
  • Strong experience working in Linux environments using the CLI and native tools
  • Proficiency developing automation scripts in Bash
  • Recent hands-on software development experience using Python and Java
  • Experience with distributed big data processing engines, including Apache Spark
  • Experience using Jupyter Notebook for analytics and development workflows
  • Data wrangling and preprocessing experience using tools such as pandas and NumPy
  • Experience working with structured and unstructured data formats including Parquet, JSON, CSV, and XML
  • Familiarity with data quality, validation, and anomaly detection concepts
  • Experience using Git for source control and team collaboration

Nice To Haves

  • Experience orchestrating workflows using Apache Airflow, including DAG design and scheduling
  • Familiarity with SQL technologies such as MySQL, MariaDB, or PostgreSQL
  • Exposure to HPC job scheduling tools such as Slurm
  • Experience using Atlassian tools including Jira and Confluence

Responsibilities

  • Design, implement, and optimize data ingress and egress pathways supporting analytics workloads
  • Build and enhance distributed data processing solutions using Apache Spark
  • Develop and maintain Python and Java applications supporting large-scale data workflows
  • Automate operational tasks using Bash scripting in Linux environments
  • Work with structured, semi-structured, and unstructured data formats to support analytics pipelines
  • Collaborate with system engineers and stakeholders to refine requirements and improve performance
  • Write and maintain documentation supporting sustainment, testing, and operational use
  • Participate in code reviews, testing, and continuous integration workflows

Benefits

  • Ample PTO to rest and recharge—plus all federal holidays and your birthday off, just because.
  • Multiple medical plan options, including ones with zero deductible or premium for employees.
  • Generous 401(k) with immediate vesting—because your future matters now.
  • Exciting bonus opportunities, from profit sharing to quarterly awards and President’s Club recognition.
  • A culture of collaboration, connection, and fun, with regular team activities that go beyond the work.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service