We are back with another episode of True ML Talks. In this, we dive deep into GenAI and LLMOps strategy in Level AI as we are speaking with Abhimanyu Talwar
Abhimanyu is Staff AI Research Engineer at Level AI. Level AI a conversational intelligence company. They use machine learning to get insights out of conversational data.
📌
Our conversations with Abhimanyu will cover below aspects:- Embracing Past Conversations for Customer Service AI- Tackling the Hidden Challenges of Generative AI- Open Source vs. Fine-Tuning- Understanding Agents- Demystifying Contact Center QA with AI- Customer Service AI with GPT-4- MLOPs Revolutionizing Customer Success
Customer service AI is all the rage, but relying solely on static knowledge bases has its limits. Agent Assist and AgentGPT, two innovative tools fromLevel AI, that unlock the power of past conversations to boost agent efficiency and customer satisfaction.
Forget keyword matching! Level employs a powerful retrieval pipeline:
To give better results, they further do the following:
By harnessing both the power of static knowledge bases and the dynamic insights hidden in past conversations, Agent Assist and AgentGPT offer a glimpse into the future of customer service AI. This future is one where agents are empowered with the right information, leading to faster resolutions, happier customers, and a more efficient contact center.
Gen AI is a powerful tool, but its success hinges on careful data curation, robust evaluation methods, and a data-driven approach
Picking the right data from your vast corpus is like finding the perfect ingredients for a delicious dish. Poor data selection leads to a model that's, well, inedible.
First thing is focus on your data. Basically, have a very good handle on what your data mix looks like. What is the quality of your annotations? All of that really matters a lot. Otherwise it will be garbage in, garbage out – Abhimanyu
After picking the right data, you will need to evaluate your Gen AI models. This isn't as straightforward as traditional AI tasks. Forget simple metrics like N-gram overlap – they miss the nuances of correctness. A single swapped word ("yes" to "no") can make all the difference.
For this you can use:
Don't rush into large models. Conduct experiments with smaller checkpoints to find the optimal data mix and task weightage.
It's tempting to stick with ChatGPT's apparent versatility. While GPT-4 shines in unconstrained scenarios, businesses operate on real-world limitations. High traffic volume demands efficient, cost-effective solutions without sacrificing performance or responsiveness.
This is where fine-tuning your own models can be advantageous:
Don't write off options like ChatGPT entirely. They can be valuable allies! Consider using them as a starting point to:
Think of them as specialized teams of LLMs, each playing a specific role in a larger workflow. Instead of single API calls, tasks involve multiple "agents" collaborating through sequential API calls.
Imagine writing a poem: you need creativity, rhyme analysis, and even grammar checking. One LLM might excel at generating initial verses, another at ensuring rhyme schemes, and a third at polishing the final draft. Agents let you leverage the unique strengths of different models to achieve superior results.
One of the offering of Level AI is Agent Assist, a powerful AI tool powered by GPT technology. It helps automate QA by analyzing conversations and providing insights into agent performance.
Here's how it works:
Benefits of Agent Assist:
GPT-4's arrival has sparked excitement in the AI world, but is it a one-stop shop for exceptional customer service experiences? Not quite. While its power is undeniable, there are a lot of hidden layers behind truly impactful AI solutions.
GPT-4's raw potential is remarkable, surpassing open-source models and APIs in generating answers. However, relying solely on its output overlooks the crucial parts of the AI pipeline: data selection, feature extraction, aggregation, and business knowledge.
You can read more about how companies leverage human expertise and AI for Customer Service in the blog below.
The customer success landscape is being redrawn by MLOPs, the magic formula behind efficient AI deployments. Here's a sneak peek at its 5-year impact:
MLOPs is the key to creating a customer-centric future, but responsible and ethical use is crucial. By embracing its potential while addressing challenges, we can build success stories worth shouting about.
You can read more about how Generative AI will shape the future of Customer Experience in the blog below.
Keep watching the TrueML youtube series and reading the TrueML blog series.
TrueFoundry is a ML Deployment PaaS over Kubernetes to speed up developer workflows while allowing them full flexibility in testing and deploying models while ensuring full security and control for the Infra team. Through our platform, we enable Machine learning Teams to deploy and monitor models in 15 minutes with 100% reliability, scalability, and the ability to roll back in seconds - allowing them to save cost and release Models to production faster, enabling real business value realisation.
Join AI/ML leaders for the latest on product, community, and GenAI developments