Netomi is the leading agentic AI platform for enterprise customer experience. We work with the largest global brands like Delta Airlines, MetLife, MGM, United, and others to enable agentic automation at scale across the entire customer journey. Our no-code platform delivers the fastest time to market, lowest total cost of ownership, and simple, scalable management of AI agents for any CX use case. Backed by WndrCo, Y Combinator, and Index Ventures, we help enterprises drive efficiency, lower costs, and deliver higher quality customer experiences.
Want to be part of the AI revolution and transform how the world’s largest global brands do business? Join us!
Do you believe in the missions of intelligence agencies? Are you interested in building state-of-the-art NLP models and solving complex technical challenges? Do you want to be a part of our journey in shaping the future of Automated Customer Service?
If you are interested in working on some of the most challenging technical and programmatic issues, we would love to discuss with you about the exciting work and career opportunities at Netomi.
Job Responsibilities
Design, develop, and deploy scalable, high-performance software systems and infrastructure to solve complex business problems.
Collaborate closely with Data Scientists to support the integration of machine learning models and data pipelines into production systems, focusing primarily on software engineering aspects (e.g., code optimization, deployment, and system integration).
Work with Product & Engineering teams to integrate solutions into products and services, emphasizing system design, coding, and best practices.
Develop and manage data pipelines to handle large datasets, ensuring efficient data ingestion, transformation, and storage to support both data scientists and engineering needs.
Architect and implement scalable software systems, optimize existing systems, and build new features with a focus on system efficiency and scalability.
Manage databases and caching systems (e.g., MySQL, Redis, Elasticsearch), ensuring efficient data storage and retrieval, while supporting the needs of data scientists working with large datasets.
Deploy robust infrastructure solutions on AWS or GCP, ensuring high availability, fault tolerance, and scalability for data-heavy applications.
Conduct experiments and test hypotheses to improve system performance and reliability, leveraging collaboration with data scientists to ensure models and algorithms are properly integrated into production.
Communicate technical details and insights to both engineering and non-engineering stakeholders, fostering cross-functional understanding.
Stay updated on the latest developments in software engineering, system design, and infrastructure best practices, sharing knowledge with the broader team.
Provide technical mentorship and guidance to junior team members, encouraging continuous learning and development.
Ensure system compliance with data security and privacy regulations, incorporating best practices into development processes.
Requirements
3+ years of experience as a software engineer, with a focus on system design, coding, and best practices, preferably in a product development environment.
Strong programming skills in Python or other relevant programming languages, with experience in building scalable systems.
Experience collaborating with Data Scientists or working on projects involving data pipelines, with an understanding of how to support machine learning model deployment and data-driven systems from a software perspective.
Proficiency in using key components of the tech stack including Elasticsearch, Redis, MySQL, and AWS services (e.g., EC2, RDS, S3, Lambda), with a deep understanding of system design principles.
Excellent communication skills, with the ability to explain complex concepts to technical and non-technical stakeholders.
Strong problem-solving and analytical skills, with a keen interest in continuous learning and skill development.
Experience optimizing infrastructure and software deployment to reduce latency and improve cost efficiency.
Optional: Experience with modern machine learning and data processing technologies such as Apache Airflow, Kubernetes, MLflow, and Hugging Face for managing machine learning workflows and distributed systems.
Additional Information
Netomi is an equal opportunity employer committed to diversity in the workplace. We evaluate qualified applicants without regard to race, color, religion, sex, sexual orientation, disability, veteran status, and other protected characteristics.
These cookies are necessary for the website to function and cannot be turned off in our systems. You can set your browser to block these cookies, but then some parts of the website might not work.
Security
User experience
Target group oriented cookies
These cookies are set through our website by our advertising partners. They may be used by these companies to profile your interests and show you relevant advertising elsewhere.
Google Analytics
Google Ads
We use cookies
🍪
Our website uses cookies and similar technologies to personalize content, optimize the user experience and to indvidualize and evaluate advertising. By clicking Okay or activating an option in the cookie settings, you agree to this.
The best remote jobs via email
Join 5'000+ people getting weekly alerts with remote jobs!