Claire/making models go fast, wondering what they learned
About
role
SWE, Neural Network Performance
optimizing inference latency on custom silicon
team
Autopilot @ Tesla
deer creek road, palo alto. since summer '24
focus
FSD inference on AI4/AI5 silicon
end-to-end driving models, real-time constraints
education
UIUC CompE '24
grainger college of engineering, '20-'24
research
IMPACT group (compiler optimization)
under Prof. Wen-mei Hwu, '22-'24
interests
interpretability, Chinese AV, what models learn
the gap between optimizing models and understanding them
i make the models that drive cars run fast enough to actually
drive cars. most of my time goes into fitting end-to-end neural
networks within hard latency budgets on custom silicon. milliseconds
matter when the car is moving.
mostly i think about what sits between the math and the metal.
been reading a lot of interpretability research lately, trying to
understand what these models actually learn. not just how to make
them faster.
readingQiu Miaojin, Last Words from MontmartrelisteningChinese Football, Ichiko Aoba, Alex GclimbingV5 project at Movement Sunnyvalecooking红烧肉 attempt #14. getting closer.bassLongview. the syncopation is a problem.