We are back with another episode of True ML Talks. In this, we again dive deep into MLOps pipelines and LLMs Applications in enterprises as we are speaking with Sandeep Singh
Sandeep is the head of machine learning applied AI at the company called Beans AI.
📌
Our conversations with Sandeep will cover below aspects:- Revolutionizing Location Intelligence with AI- The Machine Learning Engine Behind Beans.AI- Beyond the Cloud: A Hybrid Machine Learning Workflow- A Deep Dive into the Model Zoo- Managing Large Language Models- Experimentation vs. Cost in Model Hosting
Beans.AI offers a suite of software-as-a-service (SaaS) solutions that utilize AI to understand and navigate physical spaces. Their solutions go beyond traditional mapping, offering:
Beans.AI utilizes a combination of data sources, including:
From GIS Basics to Cutting-Edge Tech:
It is very important to understand the geographical information systems (GIS) before diving into AI. This foundation, combined with expertise from Esri, a leader in mapping solutions, forms the bedrock of their approach.
Beans.AI doesn't rely on a single set-up. They leverage a flexible mix of tools and platforms:
Beans.AI prioritizes speed and adaptability. They experiment with different tools and stay nimble, keeping an eye on evolving technologies. Their approach isn't about rigid processes, but about choosing the right tool for the job, allowing them to move fast and innovate.
Building these models requires close collaboration. Their Chief GIS Officer bridges the gap between geographical expertise and AI development, facilitating seamless communication and knowledge sharing.
While AI plays a crucial role, Beans.AI recognizes the value of human expertise. Their GIS knowledge and understanding of specific use cases guide the development process, ensuring models are well-aligned with real-world needs.
When experiments graduate to production-ready models, Beans.AI turns to GCP. From training complex algorithms to serving predictions at scale, GCP provides a robust and scalable infrastructure. They leverage Kubernetes clusters for seamless horizontal scaling, ensuring responsiveness during peak seasons when package deliveries soar.
Beans.AI recognizes that a single platform can't solve everything. They actively experiment with other solutions like Vertex AI for specific tasks. However, they advocate for flexibility and data ownership. Solutions like Landing.AI, which allow model portability beyond their platform, resonate with their philosophy of operational ease and cost optimization.
Beans.AI navigates the ever-evolving landscape of AI responsibly. They actively explore new solutions like Palm APIs and OpenAI's Falcon, prioritizing quality and agility. Balancing cost and functionality, they advocate for open model access after training, allowing for wider deployment and impact.
Beans.AI's approach is anything but monolithic. They constantly explore and experiment with various open-source models, tailoring them to specific needs:
Beans.AI champions open-source models, enabling transfer learning and customization:
They emphasize the importance of testing models on your own data and tasks, as benchmarks don't always translate to real-world performance.
Beans.AI is exploring exciting text-to-image applications:
They're exploring Stable Diffusion to create multiple variations of package photos, adding a touch of surprise and delight to the user experience.
There is a clear distinction between the need for training and inference when it comes to LLMs:
While Google remains their primary ecosystem, Beans.AI doesn't shy away from exploring other options:
Beans.AI emphasizes a flexible approach, adapting its strategy based on specific needs:
While latency and platform specifics aren't immediate concerns, Beans.AI, emphasizes upfront cost estimation:
Beans.AI navigates the trade-off between prompt engineering and fine-tuning:
They strategically combine both techniques for optimal results:
It is important to have cost awareness:
Keep watching the TrueML youtube series and reading the TrueML blog series.
TrueFoundry is a ML Deployment PaaS over Kubernetes to speed up developer workflows while allowing them full flexibility in testing and deploying models while ensuring full security and control for the Infra team. Through our platform, we enable Machine learning Teams to deploy and monitor models in 15 minutes with 100% reliability, scalability, and the ability to roll back in seconds - allowing them to save cost and release Models to production faster, enabling real business value realisation.
Join AI/ML leaders for the latest on product, community, and GenAI developments