<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title><![CDATA[FoundationFrontier_]]></title><description><![CDATA[SWE exploring AI × DeFi]]></description><link>https://foundationfrontier.com</link><image><url>https://cdn.hashnode.com/uploads/logos/5fbf127d08c16c34b89c4a2a/ff308a66-180e-460c-973a-9e2e066122c2.jpg</url><title>FoundationFrontier_</title><link>https://foundationfrontier.com</link></image><generator>RSS for Node</generator><lastBuildDate>Fri, 17 Apr 2026 10:46:54 GMT</lastBuildDate><atom:link href="https://foundationfrontier.com/rss.xml" rel="self" type="application/rss+xml"/><language><![CDATA[en]]></language><ttl>60</ttl><item><title><![CDATA[From Language Models to AI Engineering]]></title><description><![CDATA[This is the second post while reading AI Engineering by Chip Huyen. Chapter 1 already touches on very interesting topics: where language models come from, what foundation models actually are, and how ]]></description><link>https://foundationfrontier.com/from-language-models-to-ai-engineering</link><guid isPermaLink="true">https://foundationfrontier.com/from-language-models-to-ai-engineering</guid><category><![CDATA[AI Engineering]]></category><dc:creator><![CDATA[Garo Sanchez]]></dc:creator><pubDate>Wed, 25 Mar 2026 15:56:57 GMT</pubDate><enclosure url="https://cdn.hashnode.com/uploads/covers/5fbf127d08c16c34b89c4a2a/e4670314-9122-47b6-8fab-3892b2f22087.jpg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>This is the second post while reading <em>AI Engineering</em> by Chip Huyen. Chapter 1 already touches on very interesting topics: where language models come from, what foundation models actually are, and how AI engineering emerged as a discipline.</p>
<h2>The Statistical Nature of Language</h2>
<p>The idea that language follows statistical patterns isn't new. We've known for centuries that certain letters appear more frequently than others, for example: the letter E is the most common in English.</p>
<p>In 1951, Claude Shannon published a landmark paper on the statistical nature of language, and many of his concepts are still used today (i.e. <em>entropy</em>).</p>
<p>This matters because language models are, at their core, statistical machines. They encode patterns about how language works and use those patterns to predict what comes next.</p>
<h2>Tokens: The Basic Unit</h2>
<p>The fundamental unit of a language model is the <em>token</em>. Tokenization is the process of splitting text into these units. Tokens strike a better balance than characters or full words for how models process text. If you want to see this in action, OpenAI has a <a href="https://platform.openai.com/tokenizer">tokenizer tool</a> where you can input any text and see exactly how it gets split:</p>
<img src="https://cdn.hashnode.com/uploads/covers/5fbf127d08c16c34b89c4a2a/00fb5631-552e-44f4-bd78-1901edb947a1.png" alt="" style="display:block;margin:0 auto" />

<p>I think its really cool that you can also see how a language model represents text internally:</p>
<img src="https://cdn.hashnode.com/uploads/covers/5fbf127d08c16c34b89c4a2a/91853dd0-d12f-4e4b-860d-cb10c0fb1814.png" alt="" style="display:block;margin:0 auto" />

<h2>Two Kinds of Language Models</h2>
<p>The book distinguishes two types of language models:</p>
<p><strong>Masked language models</strong> are trained to fill in the blank — given a sentence with a missing word, predict what goes there. BERT is the classic example. These are useful for tasks like text classification and <strong>code debugging</strong>.</p>
<p><strong>Autoregressive language models</strong> predict the next token in a sequence. This is the architecture behind GPT and the models we associate with generative AI.</p>
<p>Claude actually created a diagram to explain it to me:</p>
<img src="https://cdn.hashnode.com/uploads/covers/5fbf127d08c16c34b89c4a2a/b53fbaa2-f7ae-489c-846a-118fe0e5d4ce.png" alt="" style="display:block;margin:0 auto" />

<h2>The Breakthrough That Made LLMs Possible</h2>
<p>In traditional machine learning, there are 2 main approaches: <strong>supervised learning</strong> and <strong>unsupervised learning</strong>. For supervised learning, you need <em>labeled data</em>: inputs paired with known correct outputs. Creating those labels requires humans, and that can be very expensive.</p>
<p>The breakthrough for text was <strong>self-supervision</strong>: models learned to infer the labels directly from the input data. This meant models could train on essentially all the text available on the internet without anyone manually labeling anything. That's how we went from language models to <em>large</em> language models.</p>
<p>To get a sense of the scale: GPT-1 (2018) had 117 million parameters. GPT-2 (2019) jumped to 1.5 billion. GPT-3 (2020) reached 175 billion. GPT-4 is estimated at 1.7 trillion, though at this point the big labs have stopped disclosing these numbers.</p>
<h2>Foundation Models and Multi-Modality</h2>
<p>LLMs were trained on text, but the same principles extended to images, video, audio, and other formats. This is called multi-modality, and it gave us <strong>Large Multi-modal Models</strong> (LMMs). Both LLMs and LMMs are foundation models: general-purpose models trained on massive datasets by a handful of well-resourced labs, then adapted by everyone else.</p>
<p>The adaptation happens mostly through three main techniques:</p>
<ul>
<li><p>Prompt engineering (giving the model specific instructions)</p>
</li>
<li><p>RAG (connecting the model to external data sources)</p>
</li>
<li><p>Fine-tuning (further training on domain-specific data)</p>
</li>
</ul>
<h2>From Foundation Models to AI Engineering</h2>
<p>AI engineering is the discipline of building applications on top of these foundation models. Training foundation models is so prohibitively expensive that only a few organizations can do it, which created the <em>"model as a service"</em> paradigm, and with it, a need for engineers who can take these models and adapt them to solve real-world problems.</p>
<p>What kind of applications can be built?</p>
<p>Chip references AWS's categorization of generative AI into three buckets:</p>
<ul>
<li><p>Customer experience</p>
</li>
<li><p>Employee productivity</p>
</li>
<li><p>Process optimization</p>
</li>
</ul>
<p>AI has already proven valuable in marketing, code generation (Chip notes that experts say AI is significantly better at generating frontend code than backend code, uh-oh 😅), information aggregation, and workflow automation: from booking restaurants and planning trips for end users, to lead management and invoicing for enterprises.</p>
<h2>The Moat Question</h2>
<p>One of the most interesting discussions in this chapter is about product defensibility. The low barrier to entry in AI is both a blessing and a curse. If something is easy for you to build, it's easy for your competitors too.</p>
<p>Chip mentions a VC general partner crushing view: many startups' entire products could become a feature inside Google Docs or Microsoft Office. If their product takes off, what stops a tech giant from assigning three engineers to replicate it in two weeks?</p>
<p>Chip further argues that in AI, competitive advantages come from three places: <strong>technology, data, and distribution</strong>. Most startups would be using the same foundation models so technology is not a moat. Distribution advantages tend to belong to big companies. That leaves data as the real edge for startups and individuals: even if you couldn't use your data to train a model, the behavioral data you collect — what your users want, how they interact with your product — is invaluable.</p>
<h2>The AI Engineering Stack</h2>
<p>AI engineering differs from ML engineering in a fundamental way: it's <strong>not</strong> about developing models from scratch, but about adapting and evaluating them. The core responsibilities boil down to three things:</p>
<ul>
<li><p>Evaluation</p>
</li>
<li><p>Prompt engineering</p>
</li>
<li><p>Building the AI interface</p>
</li>
</ul>
<p>Evaluation is especially critical because foundation models are open-ended. Unlike traditional ML models with narrow, well-defined outputs, these models can produce almost anything, which makes measuring success much harder.</p>
<p>Prompt engineering is about extracting desirable behaviors from a model without modifying its weights. This includes providing context, connecting tools and managing memory systems.</p>
<h2>Why Full-Stack Engineers Have an Edge</h2>
<p>Finally, Chip describes how AI engineering has flipped the traditional ML workflow. Before, you started with data and models, and the product came last. Now, as an AI engineer you can start with the product, validate it with users, and only then invest in data and model optimization. This rewards fast iterators.</p>
<p>In terms of programming languages, Python still dominates, but JavaScript is growing (LangChain.js, Vercel AI SDK, etc.), making this space increasingly accessible to full-stack engineers. Chip puts it clearly: full-stack engineers have an advantage over traditional ML engineers in their ability to quickly turn ideas into demos, get feedback, and iterate.</p>
]]></content:encoded></item><item><title><![CDATA[AI Engineering book: First Notes]]></title><description><![CDATA[I started reading AI Engineering by Chip Huyen. These are my notes on the concepts I found interesting. This is the first entry in an ongoing series where I document what I learn, chapter by chapter.
]]></description><link>https://foundationfrontier.com/ai-engineering-book-first-notes</link><guid isPermaLink="true">https://foundationfrontier.com/ai-engineering-book-first-notes</guid><category><![CDATA[AI Engineering]]></category><dc:creator><![CDATA[Garo Sanchez]]></dc:creator><pubDate>Mon, 23 Mar 2026 19:15:18 GMT</pubDate><enclosure url="https://cdn.hashnode.com/uploads/covers/5fbf127d08c16c34b89c4a2a/797e97f2-37c5-4c76-b54b-a6b37f795340.jpg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I started reading <em>AI Engineering</em> by Chip Huyen. These are my notes on the concepts I found interesting. This is the first entry in an ongoing series where I document what I learn, chapter by chapter.</p>
<p>Chip opens with what surprised her about ChatGPT: a relatively small improvement in model quality led to an explosion of possibilities — new applications, new use cases, a whole ecosystem seemingly overnight. She reminds us that the seed of these technologies has been around for a while; the papers powering them were published as early as the 1950s.</p>
<p>The core goal of this book is to train us in <strong>how foundation models work</strong> so we can adapt them to solve real-world problems.</p>
<p>Regardless of how fast the tooling evolves, the best practices for working with AI models stay the same:</p>
<ul>
<li><p>Systematic experimentation</p>
</li>
<li><p>Rigorous evaluation</p>
</li>
<li><p>Optimization toward cheaper and faster models</p>
</li>
</ul>
<p>The book provides a framework for adapting foundation models — both language models (LLMs) and multi-modal models (LMMs) — and helps answer a set of practical questions that I found very interesting to think of:</p>
<ul>
<li><p>Should I build this AI application?</p>
</li>
<li><p>How can I evaluate and measure my app?</p>
</li>
<li><p>Why do models hallucinate, and how can I prevent it?</p>
</li>
<li><p>How can I get the most out of prompt engineering?</p>
</li>
<li><p>What is RAG and how do I use it?</p>
</li>
<li><p>What is an agent? How do I build and evaluate one?</p>
</li>
<li><p>When should I fine-tune a model?</p>
</li>
<li><p>How do I make my model faster, cheaper, and safer?</p>
</li>
<li><p>How do I create a feedback loop for continuous improvement?</p>
</li>
</ul>
<p>Beyond these questions, the book also covers model types, evaluation benchmarks, use cases, and AI application design patterns.</p>
<p>Chip also wrote <em>Designing Machine Learning Systems</em> (DMLS) and considers both books complementary. Some topics are more relevant to ML Engineering and get deeper coverage in DMLS.</p>
<p><em>AI Engineering</em> is not a tutorial. It's about understanding the fundamentals of this role — the practical concepts needed to build AI applications that solve real-world problems.</p>
<p>Chip says you don't need deep Machine Learning knowledge to build AI applications, but it helps to be familiar with a few core concepts:</p>
<ul>
<li><p><strong>Probability:</strong> sampling, determinism, and distribution</p>
</li>
<li><p><strong>Machine Learning:</strong> supervision, self-supervision, log-likelihood, gradient descent, backpropagation, loss functions, and hyperparameter tuning</p>
</li>
<li><p><strong>Neural network architectures:</strong> feedforward, recurrent, and transformer</p>
</li>
<li><p><strong>Metrics:</strong> accuracy, F1, precision, recall, cosine similarity, and cross entropy</p>
</li>
</ul>
<p>All of these are explained throughout the book as they become relevant.</p>
<p>The book lists several reasons to read it. The ones that really resonate with me:</p>
<ul>
<li><p>Identify underserved areas in AI engineering and better understand real use cases</p>
</li>
<li><p>Deeply understand what the role involves and what a career in AI engineering looks like</p>
</li>
<li><p>Understand how this technology works out of pure curiosity</p>
</li>
</ul>
<p>Finally, here's an overview of what the book will cover chapter by chapter:</p>
<ul>
<li><p>Use cases and the current state of the industry</p>
</li>
<li><p>How foundation models work under the hood</p>
</li>
<li><p>Evaluation techniques — measuring the behavior of AI applications</p>
</li>
<li><p>Prompt engineering and security</p>
</li>
<li><p>RAG (Retrieval-Augmented Generation)</p>
</li>
<li><p>Fine-tuning</p>
</li>
<li><p>Data — how to generate the best possible data for your application</p>
</li>
<li><p>Inference optimization</p>
</li>
</ul>
<p>I'm really excited to start this new adventure. Onwards.</p>
]]></content:encoded></item><item><title><![CDATA[Hello, world]]></title><description><![CDATA[A couple months ago I suddenly found myself with a lot of free time on my hands. After taking care of some pending stuff and touching some grass, I decided to stop ignoring what we all know and unders]]></description><link>https://foundationfrontier.com/hello-world</link><guid isPermaLink="true">https://foundationfrontier.com/hello-world</guid><dc:creator><![CDATA[Garo Sanchez]]></dc:creator><pubDate>Sat, 21 Mar 2026 20:59:45 GMT</pubDate><enclosure url="https://cdn.hashnode.com/uploads/covers/5fbf127d08c16c34b89c4a2a/51ef037f-48f7-417b-9ca8-77b5af794301.jpg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>A couple months ago I suddenly found myself with a lot of free time on my hands. After taking care of some pending stuff and touching some grass, I decided to stop ignoring what we all know and understand: <em>AI is eating software</em>. It's revolutionizing industries, changing the shape of the world — everything. The interesting stuff, the funding, the innovations, the business cases, the practical day-to-day use cases — it's all happening here. So my curiosity naturally led me to decide to learn more about AI.</p>
<p>Eight years ago, when I started my tech career at Accenture, there was an AI lab that anyone could participate in. From that time I remember the differences between supervised and unsupervised learning, deep Learning terms like CNNs and GANs that I always thought were sexy but never truly understood.</p>
<p>Eight years later, the name of the game is different. It's not about learning linear algebra or calculus (although I definitely want to do that too), or about training models from scratch. It's about understanding how generative AI works — the agents, the LLMs — about being able to work with these models, optimize them, deploy them, monitor them, and in one word, being the person capable of steering them so they give us the results we want.</p>
<p>This is definitely an adventure too interesting to ignore. But where do you even start? Simple. Last week I spent around 40 hours watching and summarizing "AI Engineer Roadmap" videos, reading blogs, podcasts, interviews, and book summaries. Fortunately, the path is straightforward — there's one book that's basically the bible of AI Engineering. Unsurprisingly, that book is called <em>AI Engineering</em> by Chip Huyen. So I subscribed to O'Reilly and started reading it online. That will be my first step, and I will use this blog to update anyone interested about my learning journey.</p>
<p>We will lay our foundation and then, the frontier awaits.</p>
]]></content:encoded></item></channel></rss>