Design , build and maintain scalable pipelines using Python/Databricks Leverage Spark and SQL to process and transform large scale datasets Develop and optimize ELT/ETL processes for high volume of data workflows ETL: Hands on experience of building data pipelines. Proficiency in data integration platforms such as Apache Spark or Talend Big Data: Experience of 'big data' platforms such as Hadoop, Hive or Snowflake for data storage and processing Data Warehousing & Database Management: Understanding of Data Warehousing concepts, Relational Oracle database design Data Modeling & Design: Good exposure to data modeling techniques; design, optimization and maintenance of data models and data structures Languages: Proficient in any programming languages commonly used in data engineering such as Python or Scala DevOps: Exposure to concepts and enablers - CI/CD platforms, version control systems (e.g. GIT), automated quality control management Data Quality & Controls: Exposure to data validation, cleansing, enrichment and data controls
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Industry
Professional, Scientific, and Technical Services
Education Level
No Education Listed
Number of Employees
5,001-10,000 employees