Minimum 7 Years of experience in design, development and deployment of big data applications and ETL jobs using PySpark APIs/sparkSQL. Experience in design, build and deployment of Python based applications. Experience in writing complex SQL queries/procedures using Relational databases like SQL Server or Oracle Experience in version control system like Git and CI/CD pipeline is a must. Experience in Delta lake APIs is a plus Experience in Docker and Kubernetes is a plus Knowledge of AWS services like S3, Athena, Glue, Lambda, Redshift or Cloud platform is a plus PySpark, Delta Lake, Shell scripting, AWS Services (like Glue, Lambda, Athena, Step functions, Event Bridge, EC2, Batch, MSK etc.), PostgreSQL, RedShift, Dynamo DB, Jenkins, Udeploy, GitHub
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Industry
Professional, Scientific, and Technical Services
Education Level
No Education Listed
Number of Employees
5,001-10,000 employees