Skip to content

May 2024 AI, Machine Learning and Data Science Meetup

May 30, 2024 | 5:30 to 8:00 PM PT

Register for the event at GitHub's offices in San Francisco. RSVPs are limited!

By submitting you (1) agree to Voxel51’s Terms of Service and Privacy Statement and (2) agree to receive occasional emails.

Date, Time and Location

Date and Time

May 30, 5:30 PM to 8:00 PM Pacific


The Meetup will take place at GitHub’s offices in San Francisco. Note that pre-registration is mandatory.

88 Colin P Kelly Jr St, San Francisco, CA 94107

Talks and Speakers

Lessons Learned fine-tuning Llama2 for Autonomous Agents

Rahul Parundekar
A.I. Hero, Inc

In this talk, Rahul Parundekar, Founder of A.I. Hero, Inc. does a deep dive into the practicalities and nuances of making LLMs more effective and efficient. He’ll share hard-earned lessons from the trenches of LLMOps on Kubernetes, covering everything from the critical importance of data quality to the choice of fine-tuning techniques like LoRA and QLoRA. Rahul will share insights into the quirks of fine-tuning LLMs like Llama2, the need for looking beyond loss metrics and benchmarks for model performance, and the pivotal role of iterative improvement through user feedback – all learned through his work on fine-tuning an LLM for retrieval-augmented generation and autonomous agents. Whether you’re a seasoned AI professional or just starting, this talk will equip you with the knowledge of when and why you should fine-tune, to the long-term strategies to push the boundaries of what’s possible with LLMs, to building a performant framework on top of Kubernetes for fine-tuning at scale.

About the Speaker

Rahul Parundekar is the founder of A.I. Hero, Inc., a seasoned engineer, and architect with over 15 years of experience in AI development, focusing on Machine Learning and Large Language Model Operations (MLOps and LLMOps). AI Hero automates mundane enterprise tasks through agents, utilizing a framework for fine-tuning LLMs with both open and closed-source models to enhance agent autonomy.

Combining Hugging Face Transformer Models and Image Data with FiftyOne

Jacob Marks, PhD

Datasets and Models are the two pillars of modern machine learning, but connecting the two can be cumbersome and time-consuming. In this lightning talk, you will learn how the seamless integration between Hugging Face and FiftyOne simplifies this complexity, enabling more effective data-model co-development. By the end of the talk, you will be able to download and visualize datasets from the Hugging Face hub with FiftyOne, apply state-of-the-art transformer models directly to your data, and effortlessly share your datasets with others.

About the Speaker

Jacob Marks is a Machine Learning Engineer and Developer Evangelist at Voxel51, where he leads open source efforts in vector search, semantic search, and generative AI for the FiftyOne data-centric AI toolkit.

Prior to joining Voxel51, Jacob worked at Google X, Samsung Research, and Wolfram Research. In a past life, he was a theoretical physicist: in 2022, he completed his Ph.D. at Stanford, where he investigated quantum phases of matter.

Multi-Modal Visual Question Answering (VQA) using UForm tiny models with Milvus vector database

Christy Bergman

UForm is a multimodal AI library that will help you understand and search visual and textual content across various languages. UForm not only supports RAG chat use-cases, but is also capable of Visual Question Answering (VQA). Compact custom pre-trained transformer models can run anywhere from your server farm down to your laptop. I’ll be giving a demo of RAG and VQA using Milvus vector database.

About the Speaker

Christy Bergman is a passionate Developer Advocate at Zilliz. She previously worked in distributed computing at Anyscale and as a Specialist AI/ML Solutions Architect at AWS. Christy studied applied math, is a self-taught coder, and has published papers, including one with ACM Recsys. She enjoys hiking and bird watching.

Ash Vardanian
Unum Cloud

About the Speaker

Ash Vardanian is the Founder of Unum Cloud. With a background in Astrophysics, his work today primarily lies in the intersection of Theoretical Computer Science, High-Performance Computing, and AI Systems Design, including everything from GPU algorithms and SIMD Assembly to Linux kernel bypass technologies and multimodal perception.

Strategies for Enhancing the Adoption of Open Source Libraries: A Case Study on

Vladimir Iglovikov

In this presentation, we explore key strategies for boosting the adoption of open-source libraries, using as a case study. We will cover the importance of community engagement, continuous innovation, and comprehensive documentation in driving a project’s success. Through the lens of’s growth, attendees will gain insights into effective practices for promoting their open source projects within the machine learning and broader developer communities.

About the Speaker

Vladimir Iglovikov is a co-creator of, a Kaggle Grandmaster, and an advocate for open source AI technology. With a Ph.D. in physics and deep expertise in deep learning, he has significantly contributed to advancing the machine learning field.