Responsibilities
• Contribute to the process of establishing a Data Lake / Data Warehouse (EDW), Operational Data Store (ODS)
• Play a key role in the process of data transformation (Data Collection, Data validation/quality, Data Cleaning, Data Exploration and Analysis activities) required for effective reporting, analytics, and visualization
• Develop and evolve the enterprise-wide data architecture strategy and roadmap to support delivery
• Build and maintain artefacts including Entity Relationship Models, Data dictionary, taxonomy, governance framework to aid data traceability
• Partner with other key team to define, document, and communicate principles, patterns, and reference architectures that provide a clear vision for delivering value from data
• Ability to identify fit for purpose data stores (relational, NoSQL, document, graph, etc.) to meet business requirements
• Oversee the migration of data from legacy systems to new solutions
• Ability to communicate, influence and persuade peers and leadershipRequired/Minimum Qualifications
• 3+ years of success in consultative/complex technical sales and deployment projects, architecture, design, implementation, and/or support of highly distributed applications
• Proven experience in architecting and implementing big data platforms like Azure synapse analytics, Snowflake, GCP Big Query, AWS Redshift etc.
• A comprehensive understanding of the principles and best practices behind data engineering, and the supporting technologies such as RDBMS, NoSQL, Cache & In-memory stores
• Experience with distributed data and analytics architectures in cloud and hybrid environments and handled data volumes exceeding 1 TB
• Experience using Azure, GCP, AWS cloud data platforms to design schema, build views and optimize data transformation processes or similar technologies
• Experience in implementing standards, best practices, and latest technical advancements in areas of data engineering and big data processing
• Experience in optimizing the infrastructure cost and application runtimes by monitoring the utilization stats using tools available and right sizing the underlying compute
• Awareness of visualization / reporting environments including Tableau, Power BI or other similar tools
• Deep knowledge of architectural patterns – such as data warehouse, data lake, and data hub – and the ability to leverage them to enable operational and analytical use cases
• Familiarity with data modeling skills,