Is "Reasoning" Just Another API Call?

What we can learn from o1 models and "Thinking Claude" The AI landscape has shifted dramatically. We now have access to both "smart" and "dumb" models, where smart model families o1 take time to think and reason before answering. But here's where

Is RAG Dead? The Rise of Cache-Augmented Generation

As language models evolve, their context windows keep getting longer and longer. This evolution is challenging our assumptions about how we should feed information to these models. Enter Cache-Augmented Generation (CAG), a new approach that's making waves in the AI community. What is CAG? Cache-Augmented Generation loads all

Your AI Might Be Overthinking: A Guide to Better Prompting

Recent research suggests that modern AI language models, particularly reasoning-focused LLMs like o1, often engage in excessive computation. Here's what this means for prompt engineering and how you can optimize your AI interactions. The Overthinking Problem Consider this striking example: when asked to solve a simple "2+

How OpenAI's o1 model works behind-the-scenes & what we can learn from it

The o1 model family, developed by OpenAI, represents a significant advancement in AI reasoning capabilities. These models are specifically designed to excel at complex problem-solving tasks, from mathematical reasoning to coding challenges. What makes o1 particularly interesting is its ability to break down problems systematically and explore multiple solution paths—

All you need to know about prompt engineering

I recently recorded a podcast with Dan Shipper on Every. We covered a lot, but most interestingly spoke a lot about prompt engineering from first principles. Figured I would out all the highlights in blog form. The reports of prompt engineering's demise have been greatly exaggerated. The Three

The first platform built for prompt engineering