I took a career break
Career break in 2025 & Age of AI
After budgeting and tracking expenses for a year and a half, I did it
Advice
- The global job market is not good; too many good candidates
- Do not do this on a whim
- I spent a year tracking my budget
- Worked 10 years with ~4 months in between jobs
- Good at planning and Excel sheets
- Think about it
A couple of reasons why
- I was tired
- When I started working in 2015, it was normal for people to take career breaks. I knew people from Serbia and other countries who went traveling for 6–9 months around the world.
- JS Fatigue (that's a term). I'm mainly a frontend dev (JS/TS world), low on backend production projects, but with a lot of side projects.
- I wanted to explore AI — let's call that a Bet
I was tired
Self-explanatory. I really felt I was stagnating in my career and wanted to learn something else, but the options for branching weren't interesting to me.
What I did while I didn't code or explore
- I read — mostly Sanderson (Stormlight Archive, Mistborn Era 2 read)
- Played video games, RDR2 (rerun), Clair Obscur: Expedition 33 — even if you don't play games, check out the soundtrack, it's on Spotify
- Coded for myself, related to the Bet
JS Fatigue
- I have been in the industry for 10 years, mostly web, full-stack/frontend
- In that time period, I've seen the loop: everything on the server, then with JS, then SPA, then again resurgence of server-rendered content - Main difference: the cloud
- Only abstractions are added, and the ways of fetching data have changed — the end result is the same
- In my opinion, there are more abstractions now. What really changed is how we draw, fetch data, and interact with the server
- For frontend devs, it's now normal to know more than just frameworks: cloud knowledge, CI/CD, etc., a bit of infra.
The Bet
- There is a big push in AI, heavily subsidized by Microsoft, Tencent, and private VCs. If you don't know that, you're living under a rock
- ROI must be achieved, so companies are pushing AI-assisted tools — my thought
- People in C-level positions at prominent companies mention that this is a bubble (duh), but I take their opinions with a grain of salt
- The LLM world, as of the time of writing (just after ChatGPT 5 went live), has plateaued — my opinion
- The cycle of bubbles is always there: dot-com, crypto, maybe AI is next
- Nobody knows what will happen in 5–10 years
- Chip providers are the choke point for cheap inference
- If the bubble bursts, my bet is that it'll be cheaper to rent a GPU and deploy your own infra; also, privacy
What I made
-
Blog analyzer Enter a blog post address in a CLI; it uses two LLMs (Terrasect and Qwen3) for OCR and summarization. Python for web scraping Blog Analyzer
-
D&D Agent with TUI, with Fast MCP for rolling dice, attacks, spawning items, and storing local history. Single-player D&D adventure creator that works on Qwen Instruct with tool calling and local Fast MCP for navigating the world with narrative storytelling
-
Obsidian → RAG I use Obsidian as a brain dump, so I first implemented barebones RAG with cosine similarity, which worked well for structured data. For unstructured data, I hardened it via libraries, and the results are better but not perfect. Obsidian RAG GitHub repo
-
Jira (UI & Barebones MCP) - Barebones MCP that can be plugged-and-played into VS Code (for example) and works via stdio - Has a bridge to Nest, which is connected to the UI in a SPA Vite app - Uses MCP to get the current data via HTTP bridge - Uses Gemma-Instruct for fuzzy calls to map them to real tool calls - Tool call returns JSON (task, issues, sprint, board, epic) - Chat about the risks, list subtasks, etc. via LLM (Qwen2.5)
- That video on the top ThreeJS scene maker - Using ThreeJS ChatGPT 5 for initial prompt. - Opened Codex, added prompt as an MD - Burned 1.9M tokens in two conversations until I made it - Two phases: first with voxels - Models - Will write a blog post about it
What I have learned
- How to find a model that actually works
- Hugging Face is an awesome place
- Downloading a GGUF is sometimes easier than building it
- llama.cpp maintainers are awesome
- Need to explore more stuff like vLLM
- Quantizing models to run on my hardware
- Model orchestration
- Llama hell
- Different model, different template hell
- Everybody hallucinates
- Stuff can be delivered as MVPs very easily
How I did it
- Using only AI-assisted coding
- Mostly deleting code, writing prompts, and changing environment variables and infra setup
- Will go over every example and maybe add some about me in the future
- I did not code at all (well, 98%)
- Max amount spent monthly: 40 bucks
Thanks for reading
Best of vibes,
Mrdjan