Maxar Technologies-posted 4 months ago
$108,000 - $180,000/Yr
Full-time • Mid Level
Reston, VA
1,001-5,000 employees

Maxar is seeking a Mid-Level Data Scientist to play a key support role within the data science team, responsible for automating data workflows, cleaning and preparing large-scale datasets, and ensuring that senior data scientists have reliable, well-structured data for advanced analytics and modeling. This position requires strong technical skills in data wrangling, scripting, and big data processing frameworks. The ideal candidate is comfortable working in modern cloud environments, has familiarity with containerization, and is capable of handling data pipelines independently, with some exposure to tools like Clairvoyant for orchestration or pipeline development. Life with Us Project: This project supports a hard and deeply buried target mission set, providing critical analytic insights to government stakeholders. Success depends on the development of accurate, scalable, and mission-tailored data solutions that enable timely and informed decision-making. The work directly impacts national security objectives by transforming complex data into actionable intelligence in a highly specialized and sensitive domain. Your Career: We are quickly growing our team and this opportunity will provide ample opportunity for career growth and skillset development. You will have the opportunity to work closely with leadership to help set your own goals and ensure you are on a path to achieving them.

  • Collaborate with senior data scientists to prepare, clean, and process structured and unstructured datasets for analysis.
  • Automate ETL workflows and streamline repetitive data preparation tasks using Python, SQL, and scripting tools.
  • Operate in big data ecosystems using tools such as Spark, Hadoop, or their cloud-native equivalents (e.g., AWS Glue, Azure Synapse, Databricks).
  • Assist in the development and deployment of data pipelines in collaboration with data engineering and DevOps teams.
  • Implement basic statistical analyses, visualizations, and reporting to support exploratory data analysis and hypothesis validation.
  • Maintain detailed documentation of data preparation methods, scripts, and pipeline configurations.
  • Support integration of data into downstream modeling and LLM/NLP workflows.
  • Contribute to operationalizing data products, APIs, and internal data access tools.
  • Work in containerized environments (e.g., Docker) and CI/CD pipelines when deploying scripts or tools.
  • Gain exposure to knowledge management systems supported by LLMs and assist in tagging or curating training datasets for LLM models.
  • Current/active TS/SCI security clearance and be willing and able to obtain CI polygraph.
  • 5 years of professional experience in data science, analytics, or data engineering roles.
  • Bachelor’s degree in data science, computer science, engineering, statistics, GIS, or related discipline. Degree may be substituted with an additional 2 yrs of experience.
  • Proficiency in Python and SQL; familiarity with R is a plus.
  • Experience with data wrangling libraries (e.g., pandas, PySpark) and working with APIs or batch processing tools.
  • Familiarity with ETL pipelines and orchestration tools (e.g., Apache Airflow, Clairvoyant, or similar).
  • Comfortable working with large datasets in distributed environments (e.g., Spark).
  • Foundational understanding of containers (e.g., Docker) and how they are used in deploying data workflows.
  • Exposure to cloud platforms such as AWS, Azure, or GCP, especially in data-related services.
  • Strong attention to detail and documentation practices.
  • Master’s degree in a technical field (e.g., Data Science, Computer Science, Engineering).
  • Experience using Clairvoyant or similar orchestration platforms to manage and monitor data pipelines.
  • Familiarity with container orchestration tools like Docker, Rancher, or Kubernetes.
  • Basic knowledge of LLM and NLP concepts, including using pre-trained models for simple tasks like summarization or keyword extraction.
  • Basic familiarity with GIS tools and data.
  • Exposure to building internal dashboards or search tools powered by backend data models.
  • Experience working in agile teams or cross-functional technical environments.
  • Understanding of responsible AI practices and basic data ethics.
  • Dedicated professional development time.
  • Peer groups.
  • Education reimbursement.
  • Student loan forgiveness.
  • Paid time off.
  • Health and welfare insurance.
  • 401(k) to eligible employees.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service