Focus.AI Labs / Talk Page
Day 1 / 2:05 PM
← Back to Day 1 schedule
Talk detail

How to quantize models (without killing quality)

Day 1 / 2:05 PM β€” one-screen context page for this session.

Slot: 2:05 PM
Day: Day 1
Type: Talk
Speakers: Philip Kiely
Venue: Hyatt Regency Miami
Guide page: Focus.AI Labs style
01

Official Session Summary

Pulled from the live conference page.

Four-bit quantization has a bad reputation for destroying model quality. While it’s true that post-training quantization in 4-bit integer formats makes models noticeably worse, new microscaling data formats like MXFP4 and NVFP4 deliver on the promise of fast low-precision inference without meaningful quality loss. This talk introduces these data formats along with a shift from quantization as a binary decision to quantization as a granular process with model-level considerations (quantization across weights, activations, KV cache, attention) and layer-level considerations (quantization of input, output, and hidden layers) to help you preserve quality while accessing improved performance and cost characteristics from low-precision inference.

02

Speaker Background

Quick context on the person or people on stage.

Philip KielyBasetenHead of Developer Relations

Head of Developer Relations at Baseten, translating model serving and inference tradeoffs into practical guidance for builders shipping AI products.

03

Why This Slot Matters

A compact framing layer for navigating the conference.

This is one of the more substantive abstract-backed sessions on the schedule; worth opening when you need enough context to decide whether to stay in the room.