LLM Powered Search Recommendations and Growth Strategy

In this deep exploration of employing Large Language Models (LLMs) for enhancing search recommendation systems, we will conduct a technical deep dive into the integral aspects of developing, fine-tuning, and deploying these advanced models. Our focus will be on the architecture of the models, strategies for effective data collection, and methodologies to pinpoint relevance and accuracy in recommendations. 

Following this, we will examine the necessary steps to ensure the model's efficiency and stability once integrated into production environments. A significant portion of the discussion will elaborate on a practical case study using Pinterest, demonstrating how LLMs can be strategically utilized across the full marketing funnel—enhancing user activation, conversion, and retention. By the end of this presentation, attendees will gain insights not only into the technical implementation but also into deriving substantial growth and engagement by leveraging LLM technology in practical, scalable ways.


Speaker

Faye Zhang

Staff Software Engineer @Pinterest, Lead on GenAI Search Traffic Projects, Speaker, Expert in AI/ML with a Strong Background in Full-Stack Development

The speaker is a seasoned Senior Software Engineer with deep expertise in AI and Machine Learning, and a strong background in full-stack development. Intent on building the future with scalable, innovative data solutions, and significantly impact user acquisition worldwide. Notably, at Pinterest, she led its first-ever GenAI production experiment that increased traffic by 20% and user engagement by 60%. She is currently leading Pinterest AI-driven search traffic projects. Additionally, she is a well-regarded speaker, frequently contributing her insights at AI events in San Francisco and Paris.

Read more
Find Faye Zhang at:

Date

Tuesday Nov 19 / 02:45PM PST ( 50 minutes )

Location

Ballroom BC

Topics

AI/ML Architecture Generative AI Growth

Slides

Slides are not available

Share

From the same track

Session LLMOps

Navigating LLM Deployment: Tips, Tricks, and Techniques

Tuesday Nov 19 / 01:35PM PST

Self-hosted Language Models are going to power the next generation of applications in critical industries like financial services, healthcare, and defense.

Speaker image - Meryem Arik

Meryem Arik

Co-Founder @TitanML, Recognized as a Technology Leader in Forbes 30 Under 30, Recovering Physicist

Session Generative AI

GenAI for Productivity

Tuesday Nov 19 / 11:45AM PST

At Wealthsimple, we leverage Generative AI internally to improve operational efficiency and streamline monotonous tasks. Our GenAI stack is a blend of tools we developed in house and third party solutions.

Speaker image - Mandy Gu

Mandy Gu

Senior Software Development Manager @Wealthsimple

Session AI/ML

10 Reasons Your Multi-Agent Workflows Fail and What You Can Do About It

Tuesday Nov 19 / 03:55PM PST

Multi-agent systems – a setup where multiple agents (generative AI models with access to tools) collaborate to solve complex tasks – are an emerging paradigm for building applications.

Speaker image - Victor Dibia

Victor Dibia

Principal Research Software Engineer @Microsoft Research, Core Contributor to AutoGen, Author of "Multi-Agent Systems with AutoGen" book. Previously @Cloudera, @IBMResearch

Session Machine Learning

A Framework for Building Micro Metrics for LLM System Evaluation

Tuesday Nov 19 / 05:05PM PST

LLM accuracy is a challenging topic to address and is much more multi dimensional than a simple accuracy score. In this talk we’ll dive deeper into how to measure LLM related metrics, going through examples, case studies and techniques beyond just a single accuracy and score.

Speaker image - Denys Linkov

Denys Linkov

Head of ML @Voiceflow, LinkedIn Learning Instructor, ML Advisor and Instructor, Previously @LinkedIn

Session

Scaling Large Language Model Serving Infrastructure at Meta

Tuesday Nov 19 / 10:35AM PST

Running LLMs requires significant computational power, which scales with model size and context length. We will discuss strategies for fitting models to various hardware configurations and share techniques for optimizing inference latency and throughput at Meta.

Speaker image - Charlotte Qi

Charlotte Qi

Senior Staff Engineer @Meta