We’re looking for a Data Engineering Intern to support our Geospatial and AI Foundations work over a summer internship that transitions into part-time work through September. This role begins as a full-time, 3-month summer internship and then continues part-time through September. This internship is designed for a graduate-level candidate who wants to bridge geospatial research, production data engineering, and applied AI systems. You’ll contribute meaningful geospatial analyses, help build and improve production data pipelines, and experiment with AI-driven data access tools. This role plugs directly into the data and platform teams, supporting both research initiatives and operational data infrastructure. Your work will help transform complex geospatial insights into reliable, production-ready data assets used across the organization. You’ll split your time across three areas: geospatial data science, data engineering, and applied AI systems. GeoData Science (Core Research Contribution) Lead research-oriented analyses such as tree canopy classification, slope and terrain analysis, and spatial feature extraction at scale Design and document reproducible analytical workflows Translate complex geospatial methods into clear, accessible outputs for non-technical stakeholders Connect research outputs to production data pipelines Share learnings on emerging GeoAI methods with the team Data Engineering (Skill Building with Mentorship) Build or improve Airflow ELT pipelines with mentorship and clear documentation Write clean, well-structured SQL and Python pipelines Develop modular dbt models with semantic layer definitions and documented business logic Contribute to data quality systems, including schema validation and freshness monitoring Support DataHub adoption through schema documentation and lineage tracking Communicate progress through documentation, code reviews, and updates AI Engineering (Applied Learning) Build data agents using tools like LangGraph, LangChain, or Bedrock Agent Core Develop and maintain RAG pipelines for natural-language data access Iterate on text-to-SQL approaches and document failure modes Contribute to MCP server development as needed Evaluate agent outputs and refine prompts based on feedback
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Intern