I just came back from maternity leave. And honestly? I felt like I'd missed a decade in six months. I talked about starting small in my other blog Lost in the AI Hype, I Started Small
I've spent the last fifteen years designing cloud systems. And even I felt behind. AI went from a thing people were experimenting with to a thing everyone's apparently building with, and I had no idea where to start.
So I did what any architect would do. I went back to first principles.
I'm rebuilding my AI mental model from scratch in public. No math. No expert-level coding. Just real problems, the architecture underneath, and honest notes on where things might break.
If you prefer video, please watch Episode 1 of my video series . If you prefer reading, you're in the right place.
The demo: AI adapts a recipe in under a minute
Before any theory, let me show you what these models can actually do.
I opened Amazon Bedrock Playground, pasted a real recipe, and asked three questions with each one pushing the model a little further:
1. Extract and summarise
"What are the core techniques in this recipe, strip off the fluff?"
Clean, fast, useful. You might think: that's a fancy Ctrl+F (search).
2. Interpret and advise
"Looking at this recipe, what's the thing that's most likely to go wrong for someone cooking it for the first time?"
Now we're somewhere a search tool genuinely can't go. The model is reasoning about the recipe like spotting the bit where people actually mess up.
3. Personalise
"I'm cooking this for six people on Saturday. One is vegan, one is gluten-free. Adapt the recipe, give me a shopping list, and a timeline starting from 4pm."
This is the moment. I asked it something I'd normally spend twenty minutes thinking through. It gave me a starting point in ten seconds.
If you're curious but not technical, that's already useful.
If you're a builder, you're probably already thinking so what happened here.
So what actually happened?
Here's the architecture, as simply as I can put it.
I sent text called a prompt to a foundation model.
People throw around terms like AI, LLMs, and foundation models like they all mean the same thing but they don’t.
AI is the broad umbrella. It includes everything from recommendation engines and fraud detection systems to generative AI tools like ChatGPT.
Foundation models are a subset of AI, they are large models trained on massive datasets that can be adapted for different tasks. These aren’t just text models; they can generate images, video, speech, code, and more. Platforms like Amazon Bedrock give access to many of these models.
LLMs (Large Language Models) are a specific type of foundation model built for language tasks like answering questions, summarizing text, writing, or coding. So in my recipe demo, I was technically interacting with an LLM.
The simplest way to think about it:
AI → Foundation Models → LLMs
So, in our case it means its a big model trained on a huge mix of data for your day to day general purpose.
The model is a piece of software trained on an enormous amount of text: books, articles, code, conversations. It is not searching the internet. It learned patterns from all that text beforehand.
When I give it my prompt, it predicts the most useful response based on everything it learned.
Input (prompt) → Foundation Model → Output (response)
I've been building distributed systems for years, and a foundation model call is simpler than most of the APIs I'm used to. It's an HTTP request with text in, text out.
The complexity isn't in the call itself, it is in what the model learned before you or I ever showed up.
And this exact loop is what the entire current wave of AI is built on.
Every time you see a new Claude, or GPT, or Llama land, what's actually happening is someone trained a bigger or smarter version of this same idea.
Same loop. More data. Better prediction.
Where it breaks
The model doesn't know if it's right. It's predicting what a useful answer looks like. Sometimes that prediction is brilliant. Sometimes it invents something that sounds plausible and is completely wrong.
Every time you use one of these tools, ask yourself: what would I need to double-check before I trusted this?
That question is the single most useful habit you can build right now. We'll dig into why this happens in the next post.
Where the models live: Amazon Bedrock
You might've noticed I wasn't using ChatGPT or Claude's own website. I was using Amazon Bedrock.
Bedrock is where a bunch of foundation models live on AWS. Anthropic's Claude, Meta's Llama, Mistral, Amazon's own models, they are all callable through Bedrock, no need to run or train anything yourself.
The Playground is the easy door in, just type and go. Later in this series, when we start building, we'll call these same models from code. Same models, different door.
A note on my stack
I work at AWS. So the tools I use in this series are AWS tools like Bedrock for the models, and later, an AI-powered IDE called Kiro for building.
The concepts, though, aren't AWS specific. Foundation models, tokens, context windows, RAG, agents, these work the same way on any cloud. I'm showing you my stack. And honestly, I'm still figuring out which parts of it are great and which parts are a pain. You'll know which is which.
Try it yourself
If you're just getting started: open any AI chat tool (Bedrock Playground, Claude, ChatGPT, whatever you have access to), paste a recipe, a contract, a long email and ask it three questions:
- One to summarise.
- One to interpret.
- One that's personal to you.
See what happens. That's your homework.
If you're more on the builder side: the mental model is simple: text in, model, text out. Everything we build in this series is a variation on that loop.
What's next
Next up: when AI sounds confident and is completely wrong. Why it happens, how to spot it, how to stop it.
This is a series. I'm learning this in public, building as I go, and being honest when things don't work. If that sounds useful, please follow along.
This post is part of the "Learning AI Out Loud" series, a cloud architect learning AI from first principles. Watch the video version or follow the series on dev.to.



Top comments (20)
I feel this so hard.
Looking forward to your series!
Thank you for your support!!! Appreciate it!
coming back to first principles was the right call. the surface layer changes fast but the foundational thinking doesn’t - and that’s what makes a 15-year cloud background actually useful in AI right now.
Appreciate your kind words! I am true believer of diving deep into fundamentals and foundational learning instead of just surface layer.
yeah, surface tools change every 6 months now. the fundamentals are the only thing that compounds. makes more sense to invest there.
One thing I’m intentionally doing in this series: not pretending this technology is magic. We will we keep peeling back layers and understand what is our job as builders in this new world of building with AI.
Your dive into AI fundamentals is spot-on- the landscape shifts fast. The recipe summarization demo is a great baseline for what LLMs can do.\n\nBut scaling this to critical domains like drug interaction graphs introduces immense complexity.
Oh yes definitely! I want to cover some of these complexities in future posts!
Thank you for sharing your experience! Definitely taking a break and coming back is overwhelming given how fast things move, there's a sense of constantly feeling behind. I loved your focus on mental models from first principles. It's super easy to get lost in all of the hype (which frankly a lot of the latest things are) but sticking with the underlying logic and understanding capabilities is key to breaking through the noise.
This is a great resource for anyone feeling overwhelmed in the current landscape.
Thank you for your kind words!
The relearning loop is the part that compounds. Every break from the field means rebuilding the mental model from scratch, and the cost isn't time spent rereading docs, it's the lost intuition about which abstractions held up and which were trend chasing. The same dynamic hits AI agents that lose working context between sessions. That's the layer we're closing at Mneme.
Thats so true! And the field is super fast paced, its important to keep the focus on the right things. I would love to learn more about Mneme, can you point me where do I get started?
The past six months have been wild! I love the idea of going back to establish a foundational understanding first. You got this Rohini!
Thank you Erin! It is a wild ride!!! :D
This is amazing, simple and useful. thanks!
Thanks! Appreciate your kind words!
Some comments may only be visible to logged-in visitors. Sign in to view all comments.