this is a space to document and distill my thoughts on publications and articles within the research community. literature reviews are personal and generally messy (marked in pink); blog post are intended for a wider audience (marked in gray ).

i write for my own edification; and posts are subject to my own interest for the topic at the time of writing.

llama/phi-3, scaling laws, and the benchmarking conundrum

April 23, 2024

For the first time since its release, people seem to be wavering in their confidence in Chatbot Arena. This thread caught my attention. It seems as if the release of Llama 3, and the models initial ranking has led people to question not only Llama 3 but the whole leaderboard itself. After letting the dust settle, I can confidently say that this vexation is unfounded.Llama 3 is a great model, and the Chatbot Arena remains a solid proxy for human judgement. Should they have waited a few days to...


making sense of floating points

March 21, 2024

NVIDIA GTC has just passed and with it came the announcement of their new architecture 'Blackwell'. It seems that Jensen's way of outpacing Moore's Law is just to reduce precision: "Blackwell is up to 30x faster than Hopper with over 20 petaflops of FP4 power". FP4. That's 16 numbers. It's funny to imagine a functioning representation scheme with just 16 numbers, i mean you can hold the entire set in a 2 byte lookup table! Floating point numbers was something I just took at face value for a long...