- Minimum 6+ years of experience in software industry with at least 5 years' experience in Big Data (Spark, Hive, Hadoop eco system).
- Minimum 3 years of hands-on experience in core Spark, Spark SQL, Spark Streaming, and Spark performance tuning
- Minimum 3 year of hands-on coding experience in Python .
- Experience setting up Jupyter notebooks, installing JupyterLab extensions, Kernels Management
- Experience with building pip, Conda Packages, dependency resolution and setting up CICD pipelines.
- 5+ years of Red Hat Linux or UNIX experience
- Hadoop experience in deploying and administering Hadoop Clusters
- Experience in Data Integration projects involving data warehousing, data marts, ETL/ELT, and performance tuning
- Strong problem solving and analytic skills.
- Excellent oral and written communication skills.
- provided by Dice