Data Scientist

OMG TechnologyCary, NC
Hybrid

About The Position

We are seeking an experienced Data Scientist with strong expertise in Data Science and machine learning engineering, possessing hands-on experience in designing and deploying ML solutions in production. This role focuses on building scalable ML solutions, productionizing models, and enabling robust ML platforms for enterprise-grade deployments.

Requirements

  • Bachelor’s degree in Computer Science, Information Technology, Data Science, or related field.
  • 5+ years experience with Python (pandas, PySpark, scikit-learn; familiarity with PyTorch/TensorFlow helpful), bash, experience with Docker.
  • Design and implement predictive and prescriptive models for regression, classification, and optimization problems. Apply advanced techniques such as structural time series modeling and boosting algorithms (e.g., XGBoost, LightGBM).
  • 5+ years experience with SageMaker (training, processing, pipelines, model registry, endpoints) or equivalents (Kubeflow, MLflow/Feast, Vertex, Databricks ML).
  • 5+ years’ experience with Databricks DABS or Airflow or Step Functions, e-driven designs with EventBridge/SQS/Kinesis.
  • 3+ years experience with AWS/Azure/GCP on various services like ECR/ECS, Lambda, API Gateway, S3, Glue/Athena/EMR, RDS/Aurora (PostgreSQL/MySQL), DynamoDB, CloudWatch, IAM, VPC, WAF.
  • Snowflake Foundations: Warehouses, databases, schemas, stages, Snowflake SQL, RBAC, UDF, Snowpark.
  • 3+ years hands-on experience with CodeBuild/Code Pipeline or GitHub Actions/GitLab; blue/green, canary, and shadow deployments for models and services.
  • Proven experience with batch/stream pipelines, schema management, partitioning, performance tuning; parquet/iceberg best practices.
  • Unit/integration tests for data and models, contract tests for features, reproducible training; data drift/performance monitoring.
  • Incident response for model services, SLOs, dashboards, runbooks; strong debugging across data, model, and infra layers.
  • Clear communication, collaborative mindset, and a bias to automate & document.

Nice To Haves

  • AWS, GCP or other machine learning certifications
  • XGBoos, Timeseries, Pytorch, or Tensorflow experience
  • Experience in retail/manufacturing is preferred.

Responsibilities

  • Design and implement predictive and prescriptive models for regression, classification, and optimization problems, applying advanced techniques such as structural time series modeling and boosting algorithms (e.g., XGBoost, LightGBM).
  • Develop and tune machine learning models using Python, PySpark, TensorFlow, and PyTorch.
  • Work closely with stakeholders to understand business challenges and translate them into data science solutions, and work in the end-to-end solutioning.
  • Collaborate with cross-functional teams to ensure successful integration of models into business processes.
  • Rapidly prototype and test hypotheses to validate model approaches.
  • Build automated workflows for model monitoring and performance evaluation.
  • Create dashboards using tools like Databricks and Palantir to visualize key model metrics like model drift, Shapley values etc.
  • Build repeatable paths from experimentation to deployment (batch, streaming, and low-latency endpoints), including feature engineering, training, evaluation.
  • Stand up and operate core platform components—model registry, feature store, experiment tracking, artifact stores, and standardized CI/CD for ML.
  • Author robust data/ML pipelines (orchestrated with Step Functions / Airflow / Argo) that train, validate, and release models on schedules or events.
  • Implement end-to-end monitoring, data validation, model/drift checks, and alerting SLA/SLOs.
  • Enforce model/version lineage, reproducibility, approvals, rollback plans, auditability, and cost controls aligned to enterprise policies.
  • Collaborate with on-shore/off-shore teams; coach data scientists on packaging, testing, and performance; contribute to standards and reviews.
  • Prototype new patterns; troubleshoot production issues across data, model, and infrastructure layers.
© 2026 Teal Labs, Inc
Privacy PolicyTerms of Service