Design, build, and maintain highly scalable, reliable, and efficient data pipelines for extracting, transforming, and loading (ETL/ELT) data from various sources into data warehouses, data lakes, and other storage systems. Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and ensure the efficient flow of data across various platforms and systems. Develop, deploy, and maintain real-time data streaming solutions to process and deliver data with low latency (using tools like Apache Kafka, Flink, or Spark Streaming). Optimize and automate data workflows, ensuring that data pipelines are efficient, reliable, and capable of handling increasing data volumes. Oversee the development of data models and schema designs to ensure accurate, accessible, and high-performance data storage for analytics and reporting. Ensure data quality, integrity, and consistency by implementing data validation, monitoring, and error-handling mechanisms. Monitor and troubleshoot performance bottlenecks in data systems, resolving issues with data processing and data access. Stay current with emerging trends and best practices in data engineering, recommending new tools, technologies, and methods to enhance existing infrastructure. Mentor and guide junior data engineers, promoting best practices in data engineering and fostering a collaborative, high-performance team environment.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level