Azure Data Engineer

Toronto 12 days agoFull-time External
343.4k - 732.4k / yr
Responsibilities • Design, develop, and optimize large-scale data pipelines and ETL processes to efficiently extract, transform, and load data from various sources including cloud platforms like AWS and Azure Data Lake. • Build and maintain scalable data warehouses utilizing technologies such as SQL Server, Oracle, Hadoop, Apache Hive, Spark, and Big Data frameworks to support analytics and reporting needs. • Collaborate with cross-functional teams to understand data requirements and translate them into technical solutions using Python, Java, Shell Scripting, Bash (Unix shell), Talend, Informatica, and RESTful APIs. • Implement data models and database designs that ensure high performance, reliability, and security for enterprise data systems. • Develop automated workflows for model training, analysis tasks, and data validation to streamline operations. • Integrate linked data sources to enhance data richness and enable comprehensive analysis across disparate datasets. • Utilize analytics tools such as Looker for visualization and reporting to deliver actionable insights to stakeholders. • Participate in Agile development cycles to continuously improve data infrastructure while ensuring compliance with best practices in database management. • Conduct troubleshooting of complex data issues and perform performance tuning on large datasets. • Document processes thoroughly to ensure maintainability and knowledge sharing within the team. Qualifications • Proven experience designing and implementing ETL workflows using tools like Talend, Informatica or custom scripting languages such as Python or Shell Scripting. • Strong proficiency in SQL (Microsoft SQL Server, Oracle) along with expertise in database design principles for building efficient Data Warehouses. • Hands-on experience working with Big Data technologies including Hadoop ecosystem (HDFS), Apache Hive, Spark, and Azure Data Lake. • Familiarity with cloud platforms such as AWS (Amazon Web Services) and Azure for deploying scalable data solutions. • Knowledge of programming languages including Java and Python for developing robust data processing applications. • Experience working with RESTful APIs for integrating external systems or services into the data architecture. • Ability to analyze complex datasets using advanced analysis skills to identify trends or anomalies that inform business strategies. • Understanding of linked data concepts to connect diverse datasets effectively. • Skills in database modeling, schema design, and optimization techniques to support high-performance environments. • Experience with analysis tools like Looker or similar BI platforms for creating dashboards and reports. • Familiarity with Shell Scripting (Bash), VBA is a plus for automating tasks or integrating legacy systems. • Knowledge of Agile methodologies to collaborate effectively within fast-moving project teams. Join us if you're eager to leverage your technical expertise in a vibrant environment where innovation meets impact! Job Type: Full-time Pay: $67,850.26-$144,698.27 per year