<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0"><channel><title><![CDATA[Building Systems]]></title><description><![CDATA[Follow a backend engineer's journey from 0 to 1 in AI infrastructure. Simplifying deep tech papers and scaling AI systems.]]></description><link>https://blog.udbhavsomani.com</link><image><url>https://cdn.hashnode.com/uploads/logos/6273a40377679f3cf14d7206/b6687230-ac26-4dcc-8778-fd71aeebc159.png</url><title>Building Systems</title><link>https://blog.udbhavsomani.com</link></image><generator>RSS for Node</generator><lastBuildDate>Thu, 14 May 2026 23:55:49 GMT</lastBuildDate><atom:link href="https://blog.udbhavsomani.com/rss.xml" rel="self" type="application/rss+xml"/><language><![CDATA[en]]></language><ttl>60</ttl><item><title><![CDATA[A Backend Developer’s First Look at AI: The "0 to 1" Glossary]]></title><description><![CDATA[Intro: I am a backend dev learning AI at scale. I spent years building load balancers and microservices, but AI always felt like a "black box" of math. This is my attempt to crack that box open. Here ]]></description><link>https://blog.udbhavsomani.com/a-backend-developer-s-first-look-at-ai-the-0-to-1-glossary</link><guid isPermaLink="true">https://blog.udbhavsomani.com/a-backend-developer-s-first-look-at-ai-the-0-to-1-glossary</guid><category><![CDATA[AI]]></category><category><![CDATA[beginner]]></category><category><![CDATA[Beginner Developers]]></category><category><![CDATA[distributed system]]></category><category><![CDATA[backend]]></category><dc:creator><![CDATA[Udbhav Somani]]></dc:creator><pubDate>Thu, 14 May 2026 22:06:38 GMT</pubDate><enclosure url="https://cdn.hashnode.com/uploads/covers/6273a40377679f3cf14d7206/3841b09e-43d1-42e9-860e-53fa67abf9cd.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>Intro:</strong> <em>I am a backend dev learning AI at scale. I spent years building load balancers and microservices, but AI always felt like a "black box" of math. This is my attempt to crack that box open. Here is how I’m making sense of the jargon, using analogies a 5-year-old would get, and why it actually matters for our systems.</em></p>
<h3><strong>Question: What is Artificial Intelligence (AI) in simple terms?</strong></h3>
<p><strong>The ELI5:</strong> Imagine you have a robot friend. Usually, you have to tell him exactly what to do: <em>"Walk 5 steps, turn left, pick up the red ball."</em> That is traditional programming. <strong>AI</strong> is when you show the robot 100 videos of people picking up balls and let him figure out the steps himself. It’s teaching a computer to "guess" based on what it has seen before.</p>
<h3><strong>Question: What is a "Model" in AI?</strong></h3>
<p><strong>The ELI5:</strong> A model is like a <strong>recipe book</strong> that the robot wrote for itself after watching those videos. It’s the "brain" in a file. When we "deploy a model," we are just putting that recipe book into a kitchen (a server) so it can start cooking (making predictions).</p>
<h3><strong>Question: What is the difference between Training and Inference?</strong></h3>
<p><strong>The ELI5:</strong> <strong>Training</strong> is <strong>School.</strong> It’s the robot sitting at a desk looking at 1 million pictures of cats until it knows what a cat looks like. This is very slow and expensive.</p>
<ul>
<li><strong>Inference</strong> is <strong>The Test.</strong> It’s when you show the robot a <em>new</em> picture and ask, "Is this a cat?" The robot says "Yes!" in milliseconds. This is what our backend systems handle most of the time.</li>
</ul>
<h3><strong>Question: What is a "Weights" and "Parameters"?</strong></h3>
<p><strong>The ELI5:</strong> Think of a model as a giant control panel with millions of tiny knobs. During <strong>Training</strong>, we turn these knobs slightly left or right until the robot gets the answer right. The final position of those knobs is what we call <strong>Weights</strong>.</p>
<hr />
<h3><strong>The Backend Bridge: Why should we care?</strong></h3>
<p>As backend developers, we are used to <strong>Deterministic Systems</strong>: If I send <code>User_ID=123</code> to a database, I get the same record every single time.</p>
<p>AI is <strong>Probabilistic</strong>. If I ask a model to summarize a paragraph, it might give me a slightly different answer every time. This creates a massive challenge for us in the backend:</p>
<ol>
<li><p><strong>State Management:</strong> How do we cache a "vibe" instead of a specific ID?</p>
</li>
<li><p><strong>Latency:</strong> A SQL query takes 10ms; a Large Language Model (LLM) might take 2 seconds. How do we build "snappy" UIs around that?</p>
</li>
<li><p><strong>Cost:</strong> Running a standard API is cheap. Running a GPU-backed inference engine is like burning money if your load balancing isn't perfect.</p>
</li>
</ol>
<hr />
<h3><strong>Real-World Use Case: The "Smart" Support Agent</strong></h3>
<p>Imagine you are building a backend for a delivery app.</p>
<ul>
<li><p><strong>Old Way:</strong> If a user types "Where is my pizza?", your code looks for the keyword "Where" and "pizza" and triggers a status check. If the user types "My pie is late!", the code breaks because it doesn't know "pie" means "pizza."</p>
</li>
<li><p><strong>AI Way (Agentic):</strong> You send the text to a model. The model understands the <strong>intent</strong>.</p>
</li>
<li><p><strong>The Backend Challenge:</strong> Your backend now has to:</p>
<ol>
<li><p>Call the AI model (Inference).</p>
</li>
<li><p>The AI says: "The user is hungry and annoyed; check the GPS."</p>
</li>
<li><p>Your backend then calls the GPS service, gets the coordinates, and sends them back to the AI to "explain" it to the user.</p>
</li>
</ol>
</li>
</ul>
<p>This is <strong>Agentic AI</strong>—where the AI isn't just chatting; it's using your backend tools to solve a problem.</p>
<hr />
<h3><strong>Final Thought</strong></h3>
<p>I’m currently on Day 1 of this journey. My next post will dive into <strong>Distributed Inference</strong>: How do we handle millions of these "Smart Agent" requests without the servers exploding?</p>
<p><em>Follow along as I move from 0 to 1.</em></p>
]]></content:encoded></item></channel></rss>