Episode 11 Thumbnail

Exploring Llama 4, OpenRouter, and Model Comparison Tools

In this episode, Jason and Ryan explore the freshly released Llama 4 model from Meta, which was just released over the weekend. They dive into its capabilities, testing it on Hugging Face, and discuss its groundbreaking 10+ million token context window. The conversation covers whether such a massive context window might eliminate the need for RAG (Retrieval Augmented Generation) and how it could simplify prompt engineering by allowing for more detailed system prompts and guardrails. They also explore two model comparison platformsβ€”OpenRouter and LM Arenaβ€”which allow users to test and compare different AI models side by side. During their exploration, they discover a lesser-known model called LunarCall that surprisingly outperforms others on a specific test. This episode provides valuable insights into the rapidly evolving landscape of AI models and practical tools for comparing their performance.

Jump To

Resources

Key Takeaways

Full Transcript

Scroll or use controls to navigate
πŸ’‘ Tip: Click on the πŸ•’ icons above or timestamps in the transcript below to jump to that point in the YouTube video.
πŸ•’ Jason Hand: Hey, Ryan. What's up? 

πŸ•’ Ryan MacLean: Not much. It feels like it's just been a little while this time. Only like a couple days. Yeah, three days. 

πŸ•’ Jason Hand: It's Monday and we just did this on Friday, but this is good. This is, we're staying on top of things. This is the important thing. 

πŸ•’ Ryan MacLean: Yeah, on that note, like I, I thought I was doing good watching the news during the week, but I didn't think there was going to be a model dropped over the weekend. Yeah, Saturday seems like a, an odd day for me at least to drop a big model, that kind of stuff. Here we are.

πŸ•’ Jason Hand: It's been at least 24 hours, so something big has definitely happened in the news of AI. So yeah, LLama 4 is indeed is the thing that just came out that we just spoke right before we hit record. And we thought we would just talk a little bit about Llama 4 since it just came out and some of the tools that are. I have been on our list of things to look at a little bit closer. Some of 'em allow us to compare some different models and so maybe we'll do something like that. And keep it a little bit more on the briefer side for today's recording.

πŸ•’ Ryan MacLean: Yeah, sure thing. So I'll share my screen first. I will admit that I know some about this model just because of, watching YouTube, that kind of stuff. But it's really new. So I, I assume that I could just, turn on a Lama and add it. And of course it's only been a couple days, so it's, it hasn't been added yet. It is however hosted on Hugging Face. And if you've not been to Hugging Face before, it's I don't wanna say a model repository 'cause it's a lot more than that. And we talked a little bit about Gradio before, which kind came out of Hugging Face. So this interface should look very similar to what I've been demoing in a couple earlier sessions.

[Full transcript continues...]
                

Episode Navigation

×

Sandy?