Description:
• Strong project experience in Big Data, Cloudera Distribution 7.x, Cloud migration, RDBMS is a must.
• Strong project experience with Amazon EMR/Databricks/Cloudera CDP is must
• 4-5 experience building data pipelines using Hadoop components Sqoop, Hive, Solr, MR, Impala, Spark, Spark SQL., HBase.
• 4-5 years of programming experience in Python, Java and Scala is a must.
• Strong experience with REST API development using Python frameworks (Django, Flask etc.)
• Micro Services/Web service development experience using Spring framework is highly desirable.
• Strong experience in Amazon EMR/Data Bricks/Cloudera CDP
• Experience with Dask, NumPy, Pandas, Scikit-Learn
• Hands-on experience in Big Data, Cloudera Distribution 7.x, RDBMS
• Hands-on experience building data pipelines using Hadoop components Sqoop, Hive, Solr, MR, Impala, Spark, Spark SQL., HBase.
• Strong experience working in Real-Time analytics like Spark/Kafka/Storm
• Experience with Jenkins, JIRA
• Expertise in Unix/Linux environment in writing scripts and schedule/execute job.
• Provide vision, gather requirements and translate client user requirements into technical architecture.
• Design and implement an integrated Big Data platform and analytics solution.
• Design and implement data collectors to collect and transport data to the Big Data Platform.
• Implement monitoring solution(s) for the Big Data platform to monitor health on the infrastructure