About the job
Skills:
Statistical Analysis, Machine Learning, Python Programming, Data Visualization, Predictive Modeling, Big Data, SQL, Data Mining,
Responsibilities
Work with team members and the IA Leads to understand the requirements including business problem
or opportunity, data requirements, and technology requirements
Contribute to documentation of requirements and user stories in Jira/Confluence, as required
Build and deploy back-end software services, data applications/services and/or APIs
Build and deploy data pipelines on the data platform
Build and deploy ML models
Build and deploy front-end software for data products
3rd level support for incidents relating to appropriate productionised data products/services
Document relevant information required for knowledge sharing across the team and for operational
support handover
Collaborate with team and stakeholders to ensure successful delivery
Adhere to technology standards, data governance & privacy policies and security standards
Adhere to corporate governance policies
Attend/contribute to squad agile ceremonies (e.g. standups, sprint planning etc)
Attend/contribute to project meetings and deliverables
Build and deploy using devops and infrastructure as code
Comply with the change process and provide inputs, as required, and attend CAB, where required
Provide input into handover to the support team for new builds/enhancements
Knowledge sharing amongst the data engineering chapter and project/squad team members
Experience
Minimum of 8 years of experience in data engineering
At least 3 years experience in Machine Learning engineering
At least 4 years experience in full-stack development
Experience working in a large, complex Enterprise in a technology data-related role
Technical Skills (mandatory)
Software development skills and experience
Back-end development skills including Node.js, ASP.NET, MVC framework, Restful Services / APIs
Front-end development skills including HTML, Javascript, JQuery, Angular/React
Experience with distributed programming (e.g. Apache Spark)
Cloud data pipeline build, deploy and maintain/support skills
Python and Pyspark skills
Advanced SQL skills including the application of DBT
Data pipeline orchestration skills using Apache Airflow, Google Cloud Pub/Sub, Cloud Run and Cloud
Functions
An in-depth understanding of common databases and data structures.
Experience working with big data cloud platforms (Azure, Google Cloud Platform, AWS)
Data science knowledge and/or experience supporting ML models built on VertexAI
Experience working in agile environment
Devops & CI/CD experience
Infrastructure as code experience
Experience with Object-Oriented programming (OOP) in Java and/or C#
Experience with Unit Testing, TDD
Desired Skills and Experience
Statistical Analysis, Machine Learning, Python Programming, Data Visualization, Predictive Modeling, Big Data, SQL, Data Mining