Technical skills requirements The candidate must demonstrate proficiency in, · Strong expertise in Hadoop ecosystem – HDFS, Hive, Spark(PySpark/Scala),MapReduce · Hands-on experience with Sqoop for data ingestion · Strong programming skills in Python / Scala / Java · Advanced knowledge of SQL and query optimization · Experience in data lake architecture and maintenance · Strong understanding of distributed computing and parallel processing · Experience in ETL/ELT pipeline design and optimization · Exposure to cloud migration strategies and readiness planning· Experience with data governance tools and frameworks· Knowledge of CI/CD pipelines and DevOps practices Nice-to-have skills· Experience with cloud platforms (GCP preferred – Dataproc, BigQuery, Cloud Storage) · Familiarity with workflow orchestration tools (Airflow / Cloud Composer) Qualifications Overall 8 + years with 7-10 years of relevant work experience in Big Data/Hadoop B.Tech., M.Tech. or MCA degree from a reputed university
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Number of Employees
1-10 employees