Redhorse Corporation-posted 22 days ago
Mid Level
Arlington, VA
251-500 employees

Redhorse Corporation is seeking a motivated Data Engineer to join our National Security team focused on building robust data infrastructure for high-stakes analytical applications. In this role, you will play a key part in designing and implementing data pipelines that integrate information from diverse sources to create and maintain knowledge graph-based data products. Your work will directly support mission-critical initiatives in cyber forensics, financial forensics, and criminal investigations, ensuring data integrity, accessibility, and scalability. This position offers an excellent opportunity for growth in a collaborative environment where you’ll work with cutting-edge technologies to solve complex data challenges.

  • Design, develop, and maintain data pipelines to extract, transform, and load (ETL) data from disparate sources, ensuring seamless integration into knowledge graph structures.
  • Build and sustain knowledge graph data products using Neo4j, including modeling relationships, optimizing queries, and ensuring data consistency.
  • Utilize GraphAware Hume Orchestra for advanced graph processing and orchestration tasks.
  • Leverage Databricks and Spark for large-scale data processing, distributed computing, and analytics workflows.
  • Develop scripts and automation in Python to handle data ingestion, cleaning, and transformation.
  • Implement workflow orchestration using Airflow and data flow management with NiFi to automate and monitor pipeline operations.
  • Collaborate with cross-functional teams, including data scientists, analysts, and investigators, to support cyber forensics, financial forensics, and criminal investigations by providing reliable data access and insights.
  • Troubleshoot and optimize data pipelines for performance, scalability, and reliability.
  • Ensure compliance with data security, privacy, and regulatory standards relevant to forensic and investigative contexts.
  • Participate in code reviews, documentation, and knowledge sharing to foster best practices within the team.
  • Active Secret security clearance.
  • Bachelor’s degree in Computer Science, Data Engineering, Information Systems, or a related field (or equivalent experience).
  • 2-5 years of experience in data engineering, with hands-on exposure to ETL processes and data pipeline development.
  • Proficiency in Python for data manipulation and scripting.
  • Experience with big data technologies such as Spark and Databricks.
  • Familiarity with graph databases, particularly Neo4j, and building knowledge graphs.
  • Knowledge of workflow tools like Airflow and data integration platforms like NiFi.
  • Demonstrated expertise in supporting cyber forensics, financial forensics, or criminal investigations, including handling sensitive data and ensuring evidentiary integrity.
  • Strong problem-solving skills and attention to detail, with the ability to work independently and in a team setting.
  • Excellent communication skills to articulate technical concepts to non-technical stakeholders.
  • Experience with GraphAware Hume Orchestra or similar graph orchestration tools.
  • Understanding of data modeling techniques for knowledge graphs and semantic technologies.
  • Familiarity with cloud-based data environments (e.g., AWS, Azure) and version control systems like Git.
  • Background in forensic data analysis or investigative support in a professional setting
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service