We are back with another episode of True ML Talks. In this, we again dive deep into MLOps pipelines and LLMs Applications in enterprises as we are speaking with Labhesh Patel.
Labhesh was a CTO and Chief Scientist at Jumio Corporation, where he worked in leveraging ML / AI in identity verification space. He has held multiple leadership positions, both in engineering and science roles in the past, with leading organizations.
📌
Our conversations with Labhesh will cover below aspects:- Interesting Research Papers and Patents- Utilizing AI to solve business problems- Building the MLOps Pipeline- Breaking Down Silos: Building Cohesive MLOps Teams for Success- Navigating Cloud Provider Roadblocks- Future of Generative AI
There are a lot of challenges and opportunities in transforming manual processes with AI. Here are some key takeaways:
By prioritizing business needs, focusing on data quality, and building a strong team, you can navigate the complexities and unlock the true potential of AI to transform your operations.
For anyone building complex ML systems, there are some things you can keep in mind.
In the fast-paced world of MLOps, collaboration is king. But too often, teams become fragmented, with data scientists building models in isolation and engineers struggling to deploy and maintain them. The result? Slow progress, missed opportunities, and frustrated stakeholders.
So how do we break down these silos and build MLOps teams that thrive?
Imagine a cross-functional team of 8-10 individuals, each with unique expertise: product managers, data engineers, DevOps, security, ML engineers, QA, and even customer support. This diverse group, united by a common goal (e.g., reducing fraud), becomes a powerful force for innovation and efficiency.
Here's why this approach works:
It is of the utmost importance to do targeted hiring. You need data engineers with a strong understanding of ML pipelines and ML engineers who appreciate software engineering principles. This combination of diverse skills is the secret sauce to a high-performing MLOps team.
Breaking down silos isn't just about structure, it's about culture. Encourage open communication, celebrate diverse perspectives, and create an environment where everyone feels empowered to contribute. By doing so, you'll build a cohesive MLOps team that can turn your ML dreams into reality.
There are a lot of potential roadblocks you can encounter when heavily relying on a Cloud Provider. In such scenarios, it is very important to be able to pivot when such a roadblock arises.
Here are some common challenges that can arise
When dealing with sensitive data (PII, healthcare records), strict regulations like GDPR and CCPA come into play. Cloud providers, while compliant with general standards, might not offer specific tools for secure access and audit trails.
The potential solutions to these are:
Sometimes, cloud providers hold back specific features or release them on their schedule, leaving clients waiting for crucial functionalities.
The potential solution to this is to be proactive in communicating with your Point of contact for that cloud provider.
Giving direct feedback to the POC and communicating the blockers you face can sometimes land you and your team early access to private beta programs, ensuring that you don't miss out on future solutions.
Remember, even with roadblocks, a proactive and adaptable mindset can turn challenges into opportunities in the ever-evolving world of cloud-based MLOps.
Generative AI, particularly LLMs (Large Language Models), is all the rage. However, currently, LLMs are in a "hype phase", praised for their magical abilities to handle diverse tasks. Developers resort to throwing API calls at LLMs, leading to issues like rate limiting and high costs.
There might be a shift towards SLMs, trained for specific tasks and domains within enterprises.
This "routered architecture" would direct queries to the appropriate SLM for faster and more efficient responses.
Smaller models also address cost and scalability concerns, making them more accessible to businesses.
The transition will likely happen gradually, driven by the practical limitations of LLMs and the increasing availability of effective SLMs.
Cost reduction and improved latency will play key roles in accelerating the adoption of SLMs.
Keep watching the TrueML YouTube series and reading the TrueML blog series.
TrueFoundry is a ML Deployment PaaS over Kubernetes to speed up developer workflows while allowing them full flexibility in testing and deploying models while ensuring full security and control for the Infra team. Through our platform, we enable Machine learning Teams to deploy and monitor models in 15 minutes with 100% reliability, scalability, and the ability to roll back in seconds - allowing them to save cost and release Models to production faster, enabling real business value realisation.
Join AI/ML leaders for the latest on product, community, and GenAI developments