What Part 9 Is About
Final StageYou have built everything. Parts 1–8 gave you 27 modules of AI engineering skills — RAG systems, agents, production APIs, prompt engineering, containerisation, observability, and a specialisation track. Now you need to make those skills visible to employers.
Most engineers underinvest in the presentation layer. They build great things privately and expect recruiters to read their minds. This module teaches you to build the signal that gets you conversations: a curated GitHub portfolio, a metrics-driven resume, targeted outreach, and the ability to talk through your work technically.
The Job Search Stack
- Portfolio — 3 public GitHub repos with excellent READMEs and live demos
- Resume — quantified impact, AI keywords, 1 page unless 10+ years experience
- LinkedIn — posts showcasing what you built, searchable by recruiters
- Outreach — warm DMs to engineers at target companies, not cold applications
- Technical interview prep — system design for AI, coding, and ML fundamentals
💡 Three deployed projects beat fifty tutorial certificates. Employers in AI engineering hire for demonstrated capability. A live RAG API, a deployed SaaS app, and a fine-tuned model on HuggingFace Hub signal more than any course completion badge.
Your Portfolio: 3 Pinned Projects
SelectionGitHub lets you pin 6 repos. Pin 3 AI engineering projects. Everything else is secondary. Each pinned project should demonstrate a different layer of the stack.
Project 1 — Production RAG API
Your M18 "Chat With Your Docs" built to production standard: ChromaDB + Cohere reranker + grounded answers + citations + streaming FastAPI.
Project 2 — Agent System
Your M21 hardened research agent: LangGraph, human-in-the-loop, cost circuit breaker, structured logging. Or your track specialisation capstone.
Project 3 — Your Track Capstone
Track A: deployed SaaS with Stripe. Track B: fine-tuned model + eval report. Track C: enterprise automation. Track D: analysis pipeline with report generation.
What Makes a Project Portfolio-Ready
- Deployed and live — a URL you can click during an interview. Not "see the screenshots."
- Excellent README — architecture diagram, problem statement, demo GIF or screenshot, tech stack badges (see Tab 2)
- Quantified results — "RAG faithfulness: 0.91 on 30-question eval set" beats "good accuracy"
- Production code quality — type hints, tests, .env.example, no hardcoded API keys, no 500-line main.py
- Commit history — evidence of iteration, not a single "initial commit + everything"
⚠️ Never put real API keys, credentials, or personal data in a public repo. Use a .env.example file showing required environment variables. Add .env to .gitignore before your first commit — it is very difficult to remove secrets from git history after the fact.
Writing READMEs That Get You Interviews
SignalThe README is your project's cover letter. Most engineers write one paragraph. A great README takes 2 hours and generates 10x more recruiter interest.
# Project Name — One-Line Value Proposition [](live-demo-url) [](pypi) [](license) > **[Live Demo](https://your-demo-url.com)** | Built with Claude + FastAPI + ChromaDB ## What It Does 2–3 sentences. What problem does it solve? Who is it for? What's the result? "A production-ready RAG API that lets you query your private documents in plain English, with grounded answers and citations. Faithfulness: 0.91 on 30-question eval set." ## Demo [GIF or screenshot of the app working — this is the most important element] ## Architecture [Simple ASCII or Mermaid diagram showing: User → API → LLM/Vector DB → Response] ## Tech Stack - **LLM**: Claude 3.5 Sonnet via Anthropic API - **Retrieval**: ChromaDB + Cohere reranker (two-stage, retrieve-50 → rerank-5) - **Backend**: FastAPI + Celery + Redis - **Deployment**: Docker + Railway ## Results | Metric | Score | |--------|-------| | RAG Faithfulness | 0.91 | | Context Recall @5 | 0.84 | | p95 Query Latency | 1.2s | ## Quick Start ```bash git clone https://github.com/you/project cp .env.example .env # add your API keys docker compose up # Open http://localhost:8000/docs ``` ## Key Technical Decisions - Used two-stage retrieval (vector → reranker) because baseline hit_rate@5 was 0.65; improved to 0.84 - Chose ChromaDB over Pinecone for self-hosting cost control at this scale - Added Anthropic prompt caching for system prompt (90% cost reduction on repeated queries)
# Generate a demo GIF with terminalizer or vhs pip install terminalizer terminalizer record demo terminalizer render demo -o demo.gif # Or use asciinema for terminal demos pip install asciinema asciinema rec demo.cast asciinema upload demo.cast # get a shareable URL # Add shields.io badges to README # https://img.shields.io/badge/Python-3.12-blue # https://img.shields.io/badge/FastAPI-0.109-green # https://img.shields.io/badge/Live_Demo-Click_Here-orange # Pin your best repos on GitHub profile: # Profile → Customize profile → Pin repositories → select 3
AI Engineering Resume — Principles
Job SearchStructure (1 page, no exceptions under 10 years experience)
- Name + contact — email, GitHub URL, LinkedIn URL, portfolio/demo URL
- Summary (2 sentences) — role target + top 3 skills + years experience
- Skills — grouped: LLM APIs (Anthropic, OpenAI) | Frameworks (FastAPI, LangGraph, LlamaIndex) | Infrastructure (Docker, Redis, Prometheus) | Languages (Python, TypeScript)
- Projects — 3 bullet points each. Lead with impact, not description.
- Experience — reverse chronological. Reframe existing work using AI engineering language.
The STAR-Metric Formula for Each Bullet
# BAD — describes what you did "Built a RAG system using ChromaDB and the Anthropic API" # GOOD — leads with impact, shows measurement "Built production RAG API serving 500+ queries/day: 0.91 faithfulness score, 1.2s p95 latency, 90% cost reduction via Anthropic prompt caching" # BAD — vague agent project "Developed an AI agent using LangGraph" # GOOD — quantified with production details "Shipped LangGraph research agent with circuit breaker (capped at $0.50/session), human-in-the-loop approval for sensitive tools, structured logging — reduced manual research time by 60% for 3 team members" # BAD — generic Python experience "Used Python for backend development" # GOOD — specific AI engineering context "Led migration of 4 microservices to async FastAPI with LLM integration; added Prometheus metrics and Grafana dashboards reducing MTTD from 2h to 8min"
Keywords That Get Past ATS Filters
Include these naturally in your bullets and skills section:
# AI/LLM keywords (pick the ones you actually know) RAG · Retrieval-Augmented Generation · LLM · Large Language Models Anthropic Claude · OpenAI GPT · prompt engineering · fine-tuning LangChain · LangGraph · LlamaIndex · vector database · embeddings ChromaDB · Pinecone · Qdrant · pgvector · semantic search · reranking # Infrastructure keywords FastAPI · async Python · Docker · Docker Compose · Celery · Redis Prometheus · Grafana · structlog · distributed tracing · OpenTelemetry GitHub Actions · CI/CD · blue-green deployment # ML/Eval keywords (if Track B) HuggingFace · PEFT · LoRA · QLoRA · Unsloth · vLLM · GGUF SHAP · model evaluation · evals · DeepEval · Ragas
LinkedIn Content That Generates Inbound
VisibilityPost 2–3 times per week during your job search. Share what you built, what you learned, what failed. AI engineering content gets 3–5x more engagement than generic tech content right now. One viral post generates more recruiter messages than 100 cold applications.
Content Formats That Work
- Build in public — "Built a production RAG system this week. Here's what surprised me about chunking..." [screenshot]
- Lessons learned — "3 things I got wrong in my first agent loop and how I fixed them"
- Quantified results — "Compared Cohere reranker vs no reranker on my test set. Faithfulness went from 0.65 to 0.91. Chart:"
- Technical takes — "Agents are overused. 70% of 'agent' use cases I've seen would be better as simple chains. Here's the decision framework I use..."
Warm Outreach Template (DM to AI engineer at target company)
Hi [Name], I've been following [Company]'s work on [specific thing — blog post, open source project, paper]. I noticed [specific observation showing you did your research]. I'm an AI engineer with [X years] background in [your domain — networking, distributed systems]. I've been building RAG systems and LangGraph agents this year and published a few things: [link to best project]. I'd love to hear how [Company] is approaching [specific technical problem they work on]. Happy to share what I've learned from my production RAG work if that's useful. [Name]
💡 Warm outreach (specific, researched DM) converts 10-20% of the time. Cold applications convert 1-3%. Find engineers at target companies on LinkedIn. Read what they post. Comment thoughtfully for a week before DM-ing. When you DM, reference something specific about their work — not a generic "I'm interested in your company."
Job Boards That Work for AI Engineering
- Anthropic Careers (careers.anthropic.com)
- AIJobs.net — AI-specific job board
- Levels.fyi — for compensation research before negotiating
- LinkedIn Jobs filtered by "AI Engineer" + "LLM" + location
- Twitter/X "hiring AI engineer" — many startups hire this way
AI Engineering Interview Preparation
Get HiredSystem Design for AI — Common Questions
- "Design a document Q&A system for 10 million documents" → talk about chunking, vector DB selection, retrieval quality, caching, cost
- "Design an AI customer support system" → routing, RAG, escalation, human handoff, evaluation
- "How would you reduce LLM API costs by 50%?" → model routing, prompt caching, response caching, batching, smaller max_tokens
- "How do you evaluate a RAG system?" → faithfulness, context recall, hit rate, LLM-as-judge, eval datasets
- "How do you handle an agent that gets stuck in a loop?" → fingerprinting, max_turns, circuit breaker, structured recovery
# Framework for AI system design answers: # 1. Clarify requirements # "How many queries/day? What latency SLA? What's the document corpus size? # Is accuracy or speed more important? What's the budget?" # 2. High-level architecture (draw boxes) # User → API → [Retrieval: embed query → vector search → rerank] → [LLM: generate] → Response # 3. Deep dive on the interesting parts # "For retrieval, I'd use two-stage: ChromaDB top-50 then Cohere reranker to top-5. # Baseline hit_rate@5 is ~0.65; with reranking it's ~0.84 in my experience." # 4. Scale and reliability # "For 1M+ documents I'd use Qdrant or Pinecone. For cost, I'd cache # embeddings and responses. For reliability, retries with exponential backoff." # 5. Evaluation # "I'd run Ragas on a 50-question ground truth set. Monitor faithfulness drift in prod."
Coding Questions You'll Actually See
- Implement a retry decorator with exponential backoff and jitter
- Write a sliding window rate limiter
- Implement a simple embedding cache with LRU eviction
- Write async code: asyncio.gather for parallel API calls, return_exceptions=True
- Implement cosine similarity and explain why it's used for embeddings
Behavioural Questions with AI Framing
- "Tell me about a time you had to make a tradeoff between accuracy and cost in an AI system" — your cost monitoring and model routing work
- "How do you keep up with the rapidly changing AI landscape?" — your learning journal, the roadmap you built
- "Describe a production issue you've diagnosed" — your observability work, structured logging, Grafana alerts
The Grand Capstone — Your Signature Project
ShowpieceYour grand capstone is a project that combines multiple Parts of this roadmap into one cohesive system specific to your professional domain. For you, that means networking / DPDK / telecom infrastructure — the domain where you have 4+ years of professional experience and can build something genuinely useful.
Suggested Capstone: Intelligent DPDK/VPP Documentation Assistant
A production AI system that lets network engineers query DPDK and VPP documentation, analyze configurations, and get code-level advice.
- RAG backend — index official DPDK/VPP docs, DPDK programmer's guide, API references, community mailing list archives
- Agent layer — research agent that can search docs, retrieve code examples, compare versions, generate configuration snippets
- Production API — FastAPI with auth, rate limiting, streaming, Prometheus metrics, Grafana dashboard
- CI/CD — GitHub Actions: lint → eval → Docker build → deploy on push to main
- Evaluation — 30-question eval set covering mempool, ring buffer, PMD, DPDK EAL, VPP node graph
- Cost monitoring — per-query cost tracking, prompt caching for large doc system prompt
💡 Domain expertise is your moat. A generic "chat with PDFs" app has 10,000 competitors. An AI assistant for DPDK/VPP network engineers, built by someone who works professionally with these systems, has almost none. Your 4+ years of distributed systems and DPDK experience makes your AI product qualitatively better than anything a generalist could build.
Alternative Capstone Ideas
- Network configuration assistant: parse and explain complex DPDK/VPP configs with error detection
- Performance analysis pipeline: ingest benchmark results, generate reports, suggest optimisations via AI
- Telecom infrastructure knowledge base: RAG over 3GPP specs, ITU documents, and vendor documentation
Capstone Deliverables
- Public GitHub repo with excellent README, architecture diagram, and eval results
- Live demo URL (Railway, Render, or DigitalOcean — all have free/cheap tiers)
- Blog post or LinkedIn series: "How I built a production AI assistant for network engineers"
- HuggingFace Space demo (if Track B — fine-tuned model)
LAUNCH CHECKLIST — DO ALL OF THESE
PORTFOLIO
- 3 pinned GitHub repos — each with live demo URL in the repo description
- Each repo has: GIF/screenshot, architecture diagram, quantified results table, quick start instructions
- All repos have .env.example — no real credentials anywhere in git history
- Production code quality: type hints, tests, structured imports, no 500-line single files
- At least one project shows observability (Prometheus metrics or structured logs)
- At least one project shows evaluation methodology (faithfulness score, task success rate, or similar)
RESUME
- 1 page (unless 10+ years experience)
- Every project bullet leads with impact metric, not description
- AI keywords included: RAG, LLM, embeddings, FastAPI, LangGraph, ChromaDB, Anthropic
- GitHub URL and portfolio/demo URL in header
- Had at least 2 other engineers review it for clarity and technical accuracy
- Headline mentions AI/LLM engineering and your background domain
- About section includes the specific AI systems you've built and deployed
- Featured section shows links to your best 2 projects
- At least 5 posts published showcasing your work (screenshots, results, lessons)
OUTREACH
- List of 20 target companies built — prioritised by domain fit and team quality
- Identified 2–3 engineers or hiring managers at each target company on LinkedIn
- Sent at least 10 warm, personalised DMs referencing specific work of the recipient
- Applied to at least 20 positions (warm outreach + direct applications)
INTERVIEW PREP
- Can whiteboard a RAG system design including retrieval, reranking, generation, and evaluation
- Can implement retry with exponential backoff from memory in Python
- Can explain LoRA in one paragraph for a non-ML interviewer
- Can articulate the cost/quality tradeoff in choosing a retrieval strategy
- Have done at least 3 mock system design interviews (with peers or on Pramp)
CAPSTONE
- Grand capstone project deployed and publicly accessible
- Evaluation report published (not just "it works" — actual metrics)
- At least one blog post or LinkedIn post explaining what you built and why