Responsible for designing, constructing, installing, and maintaining large-scale processing systems and other infrastructure. Ensure that data, whether structured or unstructured, is easily accessible and usable to analysts. Build ETL tools, migrating legacy systems to modern data ecosystems, and handling FHIR resources in healthcare data environments. Design data pipelines, optimizing data processing, and deliver actionable insights. Manage GCP services like BigQuery, Dataflow, Pub/Sub, and Cloud storage to deliver business-critical insights. Possess strong cloud-based data engineering, hands-on experience with GCP services. Optimize existing workflows for performance, scalability and cost-efficiency. Design and implement data pipelines using tools like Apache Beam, Dataflow or Cloud Composer (Airflow). Develop, optimize and manage large-scale ETL/ELT workflow and processes on GCP. Utilize BigQuery for data warehousing and analytics, writing complex SQL queries for reporting and analysis. Build and maintain real-time data streaming solution using Pub/Sub and Dataflow. Implement best practices for data security, governance and compliance (IAM roles, encryption). Manage and maintain GCP storage systems like Cloud Storage, ensuring high availability and scalability. Monitor and troubleshoot data pipelines and workflows, ensuring reliability and performance.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Number of Employees
5,001-10,000 employees