Comcast-posted about 2 months ago
$90,010 - $210,963/Yr
Full-time • Mid Level

Universal Ads, a part of Comcast, enables any brand, of any size, to seamlessly make and buy commercials across premium video reaching new qualified audiences at scale. Universal Ads combines premium and brand-safe video content directly from the most influential media companies with the ease and familiarity of social ad buying. It is your one-stop shop for high-quality video ads delivering performance and unmatched scale. The position is responsible for designing, building and overseeing the deployment and operation of technology architecture, solutions and software to capture, manage, store and utilize structured and unstructured data from internal and external sources. The role involves establishing and building processes and structures based on business and technical requirements to channel data from multiple inputs, route appropriately and store using any combination of distributed (cloud) structures, local databases, and other applicable storage forms as required. The candidate will develop technical tools and programming that leverage artificial intelligence, machine learning and big-data techniques to cleanse, organize and transform data and to maintain, defend and update data structures and integrity on an automated basis. The role also includes creating and establishing design standards and assurance processes for software, systems and applications development to ensure compatibility and operability of data connections, flows and storage requirements.

  • Develops data structures and pipelines aligned to established standards and guidelines to organize, collect, standardize and transform data that helps generate insights and address reporting needs.
  • Focuses on ensuring data quality during ingest, processing as well as final load to the target tables.
  • Creates standard ingestion frameworks for structured and unstructured data as well as checking and reporting on the quality of the data being processed.
  • Creates standard methods for end users / downstream applications to consume data including but not limited to database views, extracts and Application Programming Interfaces.
  • Develops and maintains information systems (e.g., data warehouses, data lakes) including data access Application Programming Interfaces.
  • Participates in the implementation of solutions via data architecture, data engineering, or data manipulation on both on-prem platforms like Kubernetes and Teradata as well as Cloud platforms like Databricks.
  • Determines the appropriate storage platform across different on-prem (minIO and Teradata) and Cloud (AWS S3, Redshift) depending on the privacy, access and sensitivity requirements.
  • Understands the data lineage from source to the final semantic layer along with the transformation rules applied to enable faster troubleshooting and impact analysis during changes.
  • Collaborates with technology and platform management partners to optimize data sourcing and processing rules to ensure appropriate data quality as well as process optimization.
  • Creates and establishes design standards and assurance processes for software, systems and applications development to ensure compatibility and operability of data connections, flows and storage requirements.
  • Reviews internal and external business and product requirements for data operations and activity and suggests changes and upgrades to systems and storage to accommodate ongoing needs.
  • Develops strategies for data acquisition, archive recovery, and database implementation.
  • Manages data migrations/conversions and troubleshooting data processing issues.
  • Understands the data sensitivity, customer data privacy rules and regulations and applies them consistently in all Information Lifecycle Management activities.
  • Identifies and reacts to system notification and log to ensure quality standards for databases and applications.
  • Solves abstract problems beyond single development language or situation by reusing data file and flags already set.
  • Solves critical issues and shares knowledge such as trends, aggregate, quantity volume regarding specific data sources.
  • 3+ Years of experience in Data Engineering
  • Experience with Python
  • Experience with Big Data
  • Experience with Data Pipelines (Airflow)
  • Amazon Athena
  • Amazon Elastic Compute Cloud (Amazon EC2)
  • Apache Airflow
  • Data Modeling
  • PostgreSQL
  • Comcast provides best-in-class Benefits to eligible employees.
  • An array of options, expert guidance and always-on tools, that are personalized to meet the needs of your reality.
© 2024 Teal Labs, Inc
Privacy PolicyTerms of Service