Senior Big Data Architect

Los Angeles 23 months agoContractor External
90 - 100 / hr
Hi , Hope you are doing great Please find the below requirement and let me know your interest Job Title : Senior Big Data Architect Location : San Jose, CA Onsite Duration : 12+ Months Job Description : • Translate business requirements into technical specifications, including data streams, integrations, transformations, databases, data lakes, data warehouses, and data products • Define the data architecture framework, standards, and principles, including modeling, metadata, reference data, master data, and security for the PA LDS environment • Define reference architecture, which is a pattern others can follow to create and improve data systems • Define full life cycle of data process flows (data origins, organizational data flows and functions, how data flows are managed, and how data changes in transition) • Create procedures to ensure data accuracy, quality, timeliness, availability, and accessibility • Create and implement data management processes and procedures • Collaborate with other teams within the organization to devise and implement data strategies, build models, and assess shareholder needs and goals • Develop application programming interfaces (APIs) to retrieve data • Define and deploy large, complex data environments that meet functional / non-functional business area requirements • Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc. • Design the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and ‘big data’ technologies Qualifications: • Bachelor’s Degree in Computer Science or related field of study and minimum 10+ years of data/database background including 5+ years acting as a Data Architect. • Candidate must have 2-3 years of cloud based data services in AWS such as ECT, Glue, EMR, RDS, Redshift. • Must have real-time data streaming experience with Storm, Spark-Streaming, Kafka or similar.. • Strong data management skills to collect, store, and use data in an efficient and cost-effective manner • Experience in system development life cycle, project management approaches, and requirements, design, and test techniques • Experience in established and emerging data management and reporting technologies with knowledge of columnar and NoSQL databases, predictive analytics, data visualization, and unstructured data • Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases • Experience building and optimizing ‘big data’ data pipelines, architectures, and data sets. • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement • Strong project management and organizational skills • Experience supporting and working with cross-functional teams in a dynamic environment. • We are looking for a candidate with 5+ years of experience in a Data Architect role, who has attained a degree in Computer Science, Statistics, Informatics, Information Systems, or another quantitative field. They should also have experience using the following software/tools: • Experience with big data tools: Hadoop, Spark, Kafka, etc. • Experience in artificial intelligence and machine learning (AI/ML) to build scalable systems for handling big data • Experience using data modeling tools like ERWin or Visio to visualize metadata and database schemas and/or architectures • Experience with relational SQL and NoSQL databases, including Oracle, MS SQL Server, Postgres, Cassandra, etc. • Experience with data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc. • Experience with data integration services solutions from vendors such as Informatica, MuleSoft, Talend, TIBCO, etc. • Experience with cloud-based data services such as AWS (EC2, Glue, EMR, RDS, Redshift, etc.) • Experience with stream-processing systems: Storm, Spark-Streaming, Kafka etc. Experience with object-oriented/object function scripting languages: Python, R, Java, C++, Scala, etc