20 November 2024
At ODSC 2024, VESSL showcased its MLOps platform, gaining industry attention for unified hybrid infrastructure and cost-effective LLM solutions.
San Mateo, CA – October 29-31, 2024 – ODSC 2024 was marked by strong participation from AI-related startups and enterprises, with vibrant discussions around the latest technological trends and innovative solutions. As a Gold Sponsor, VESSL AI showcased its end-to-end MLOps platform, which supports users in executing ML tasks seamlessly across diverse infrastructure environments. As a platform, VESSL is particularly powerful in its ability to unify both cloud and on-premises environments into a single managed cluster. Furthermore, VESSL has also introduced RAG and Agentic Workflow technologies to enable non-developers to build chatbots with ease.
At this event, VESSL AI connected with numerous enterprises utilizing AI and LLM. Many were working across multiple cloud environments and were actively seeking a solution to unify these management processes. VESSL’s cluster management feature garnered significant attention for its potential to resolve this issue. Additionally, VESSL highlighted its competitive on-demand rate of $1.8 per hour for A100 GPUs, which was met with enthusiasm from attendees.
Throughout the conference, we gathered essential insights from multiple sessions to capture the most impactful moments of ODSC 2024.
Jaeman An, CEO of VESSL AI, presented at ODSC West 2024 on strategies to reduce costs and address performance, security, and compliance challenges in large-scale Large Language Model (LLM) deployments using VESSL. He highlighted the platform's ability to deploy over 100 LLMs starting at just $10, significantly reducing initial setup costs. An also discussed how optimizing GPU resource allocation and implementing hybrid-cloud strategies can lead to substantial cloud cost savings, with organizations potentially saving over $1 million annually. VESSL AI's automated scaling and monitoring features ensure efficient resource utilization and maintain high-performance AI services. Additionally, VESSL AI's seamless integration with open-source tools like Hugging Face Inference API, LangChain, Kubernetes, and vLLM streamlines the development and deployment process.
A key theme at ODSC was the advancement of Retrieval-Augmented Generation (RAG) technologies. Bill DeWeese focused on essential RAG components, such as embedding and search, highlighting a significant industry shift toward hybrid search. This approach combines BM25 keyword search with Sparse and Dense models, providing improved accuracy and relevance, and is quickly becoming a new standard in RAG architectures.
Joon Park, a professor at Stanford University introduced "Generative Agents," exploring how NPCs with human-like interactions could simulate complex social behaviors. These agents offer a fresh approach to traditional A/B testing, enabling researchers to run intricate simulations that reflect real-world dynamics. By modeling social interactions and responses, Generative Agents provide a valuable tool for exploring problem-solving scenarios and understanding human-like decision-making in controlled environments.
Evaluating LLMs was another significant topic at the event. Unlike traditional AI models, LLMs require various roles, which calls for diverse evaluation methods. Anoop Sinha, Research Director of AI&Future Technologies at Google, outlined the challenges in LLM evaluation and shared insights on approaching it from perspectives such as human input, functional testing, and user feedback. His detailed explanation of evaluation methodologies, particularly in cycles where human feedback is incorporated in the training process of LLMs, was especially memorable.
As agents continue to advance, they are increasingly capable in the coding domain. However, creating a reliable coding agent, especially when handling code rather than text, presents unique challenges. Eno Reyes, CTO of Factory.co, presented on how they build reliable coding agents, explaining methodologies in planning, decision making, and environmental grounding. His in-depth discussion on the strengths and weaknesses of each approach provided considerable inspiration.
ODSC 2024 highlighted the transformative power of AI and machine learning across diverse fields, with speakers underscoring how these technologies are reshaping industries. Key discussions focused not only on enhancing efficiency but also on enabling groundbreaking applications, from advanced RAG architectures and generative agents to reliable coding systems. This year’s conference reflected the AI community’s commitment to innovation, collaboration, and real-world impact.
Our participation at ODSC 2024 has reinforced our dedication to advancing the fields of AI and MLOps. We are inspired to continue pioneering new solutions that address today’s challenges and create meaningful, scalable impacts. The insights and connections gained at this event will fuel our commitment to pushing the boundaries of what AI can achieve, fostering partnerships that drive us all toward a more intelligent and connected future.
Software Engineer
Software Engineer
Technical Communicator
Build, train, and deploy models faster at scale with fully managed infrastructure, tools, and workflows.