- Professional
- Oficina en Bangalore
Hungry, Humble, Honest, with Heart.
The Opportunity
We are looking for a highly skilled Data Integration Engineer to design, build, and manage scalable data pipelines and integration solutions across cloud and on-premises platforms. The role requires strong expertise in ETL/iPaaS tools, APIs, and data platforms, with exposure to AI/ML-driven automation for smarter monitoring, anomaly detection, and data quality improvement.
About the Team
At Nutanix, the Data Science team is a dynamic and diverse group of 50 talented professionals spread across our offices in India (Bangalore and Pune) and San Jose. We pride ourselves on fostering a collaborative and supportive environment where innovation thrives. Our team is deeply committed to leveraging data in a results-oriented manner, ensuring our solutions remain customer-centric. We believe in transparency and trust, which allows for open communication and a fluid exchange of ideas. Being agile and adaptable, we embrace diverse perspectives to drive creativity and efficient problem-solving.
Your Role
- Design, develop, and optimize data integration workflows, ETL/ELT pipelines, and APIs.
- Work with iPaaS and ETL tools (Informatica) to integrate enterprise systems.
- Build pipelines across cloud platforms (AWS, Azure, GCP) and modern warehouses (Snowflake, Databricks, BigQuery, Redshift).
- Implement data quality, lineage, and governance frameworks to ensure reliable data flow.
- Leverage AI/ML models for data anomaly detection, pipeline monitoring, and predictive quality checks.
- Contribute to self-healing pipeline design by incorporating AI-driven automation.
- Collaborate with architects, analysts, and business teams to integrate structured, semi-structured, and unstructured data sources.
- Document integration patterns, best practices, and reusable frameworks.
What You Will Bring
- 6–8 years of experience in data integration, ETL/ELT design, and data pipelines.
- Strong expertise in Informatica, or similar ETL/iPaaS tools.
- Proficiency in SQL, Python, and automation scripting.
- Experience with cloud data platforms (Snowflake, Databricks, BigQuery, etc.).
- Familiarity with data governance practices (cataloging, lineage, DQ frameworks).
- Exposure to AI/ML concepts applied to data quality and pipeline optimization.
- Understanding of DevOps/CI-CD pipelines for data integration deployments.
Nice-to-Have
- Hands-on experience with Kafka, Spark, Airflow, or event-driven architectures.
- Knowledge of REST APIs, microservices, and real-time data integration.
- Conceptual understanding or hands-on exposure to ML frameworks (Scikit-learn, TensorFlow, PyTorch).
- Experience contributing to AI-augmented/self-healing pipelines.
Education - Bachelor’s or Master’s in Computer Science, Data Engineering, Information Systems, or related field
Work Arrangement
Hybrid: This role operates in a hybrid capacity, blending the benefits of remote work with the advantages of in-person collaboration. For most roles, that will mean coming into an office a minimum of 3 days per week, however certain roles and/or teams may require more frequent in-office presence. Additional team-specific guidance and norms will be provided by your manager.
--
Solicitar ahora