Data Engineering Intern(Spring/Summer 2026)

AccuWeather CareersState College, PA
9d

About The Position

The Data Engineering Intern supports AccuWeather’s data engineering team in developing and automating systems that enhance data accessibility, reliability, and documentation. This role helps drive innovation across AccuWeather’s data infrastructure by contributing to automation initiatives, improving data discoverability, and exploring AI-driven solutions. Working under the guidance of experienced data engineers, the intern will gain hands-on experience with Databricks, Spark, Azure, and modern data engineering practices that directly support AccuWeather’s business and operational goals.

Requirements

  • Actively pursuing a Bachelor’s or Master’s degree in Computer Science, Software Engineering, Information Systems, or a related technical field.
  • Foundational knowledge of Python and SQL for data manipulation and analysis.
  • Familiarity with ETL concepts and structured data formats such as CSV, JSON, and Parquet.
  • Interest in cloud-based data platforms (Azure preferred) and modern data engineering tools.
  • Strong analytical and problem-solving skills with an eagerness to learn.
  • Effective communication and teamwork skills.

Nice To Haves

  • Exposure to Databricks, Apache Spark, or other distributed data frameworks.
  • Familiarity with Git or version control practices.
  • Interest in AI/LLM-based automation, data documentation, or metadata management.
  • Prior project or internship experience in data engineering or cloud technologies.

Responsibilities

  • Support the development and maintenance of data pipelines using Databricks, Spark, and similar technologies.
  • Write and optimize SQL and Python scripts for data transformation, integration, and automation tasks.
  • Develop automation scripts that populate metadata and comments across Databricks tables using structured definitions (e.g., CSV files).
  • Assist in building a proof-of-concept (POC) for an automated data dictionary that can be maintained using existing Databricks metadata.
  • Contribute to prototyping an AI-powered knowledge agent that leverages internal data and documentation to answer common questions.
  • Collaborate with team members to enhance data quality, cataloging, and metadata management across the ecosystem.
  • Participate in code reviews, design discussions, and sprint ceremonies to learn engineering best practices.
  • Document findings, workflows, and automation processes for future reuse.
  • Perform other duties as assigned.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service