Detailed job description - Skill Set:
• Expertise in Hadoop data modeling, ETL using Hive and Pyspark and Knowledge in Talend and Redwood
• Good Knowledge in Transmission & Distribution domain, especially in the field of Grid Modernization, Grid Analytics, etc
• Design, Develop and implement effective Analytics solutions and models with Hadoop
• Design, Architecture and implement effective Analytics solutions and models using technologies including, but not limited to - Spark, Hadoop and UNIX
• Examine and identify Datawarehouse structural necessities by evaluating business requirements.
• Assess Datawarehouse implementation procedures to ensure they comply with internal and external regulations.
• Prepare accurate Datawarehouse design and architecture reports for management and executive teams.
• Monitor the system performance by performing regular tests, troubleshooting and integrating new features.
• Recommend solutions to improve new and existing Datawarehouse solutions.
• Educate staff members through training and individual support.
• Offer support by responding to system problems in a timely manner.
• Understand and document data flows in and between different systems/applications
• Guidance to developers in preparing functional/technical specs to define reporting requirement and ETL process.
Technical Skills
Hadoop DW Architecture, Hive and PySpark
Snowflake knowledge is a plus
Utility T&D Domain knowledge, Grid modernization
Data Modeling, Framework build on Ingestion, Audit, Logging and Reconciliation