Closing Thoughts for 2025
As the year winds down, I thought I'd wrap up with some closing thoughts around AI/ML, engineering, and how it's going professionally.
We're still trying to figure out how to responsibly use LLMs / GenAI
It's undeniable that LLMs are super useful.
But use LLMs for everything and risk becoming a meme.
You'll stunt your growth if you make ChatGPT to do your homework. Karpathy highlights "the goal is that the students are proficient in the use of AI, but can also exist without it, and imo the only way to get there is to flip classes around and move the majority of testing to in class settings."
Bryan Cantrill from Oxide Computing (and one of the most compelling speakers and writers in the field and throwback from when I first heard him speak at Dockercon '15) published how they use LLMs at Oxide and it comes down to responsibility.
AI generated videos - or AI slop - is becoming its own category. I really think the AI cat videos are melting peoples brains. One Punch Man Season 3 animation is pretty bad and fans are accusing animators of using AI poorly.
Done right, you can get some really good advice for learning new concepts, studying for interviews, and planning a vacation.
The battle in the US is over. The competition is now with China.
ChatGPT came out in November 2022. That gave the AI labs and startups the chance to disrupt before Big Tech got their stuff together. Three years was all Google needed to focus their army of big brain computer scientists to top LMArena in text, vision, video, image, and search. Mic-drop.
We have lots of great private LLM companies but not so many open-source alternative. Shoutout to Olmo 3 from Ai2 leading efforts in the States. OpenAI also dropped gpt-oss which was generous of them.
But you see the speed and pace of models coming out of China and you just wish for the same here. DeepSeek-OCR provided a new way to look at inputs. Images of text might be easier to encode meaning - very applicable to coding. Karpathy's take here. And remember, these models are still blackboxes. We KNOW they can be jailbroken. Backdoors can be introduced by a small number of samples.
The White House annouced the Genesis Mission late November. Here's what Google AI Overview says: The mission aims to build an integrated national AI platform—called the "American Science and Security Platform"—that unifies federal datasets, high-performance computing resources (supercomputers), AI models, and robotic laboratories into a "closed-loop" discovery system. This system is designed to dramatically cut down R&D cycles from years to potentially days or hours.
Have you seen the Xiaomi Electric cars?
Lots of opportunities to learn, grow, and contribute
Don't let the AGI doom and gloom take over. You're here for the love of the game. Build cool things, contribute to project you use, go deep into the codebase with the help of DeepWiki.
All eyes are on AI but that leave lots of opportunity on other projects like Rails and OpenTelemetry. PyTorch has a good first issues tag. There are still people behind these projects, not AIs. And getting up to speed and helping is not trivial but a lot easier with the quick feedback you can get from asking uncle Claude.
Still bullish on Small Models and Structured Output
My most recent side project was figuring out whether people would like a fine-tuning platform for Apple Intelligence Foundation Model Adapters. I reached out to 20+ app developers using AFM, a few replied graciously, and only one used adapters. Learned a lot from using Rails 8 and Sagemaker but ultimately the market isn't there yet.
News broke that Apple is going to use Google for Siri. Hopefully the partnership goes beyond the Trillion parameter models and we get improvements with AFM too.
Structured output as a developer is still super exciting. All major LLMs now support structured outputs. Instead of begging the model to spit out a format and retrying if it didn't work, there's a structured supported way to ask. It's still not 100% but smoothing the edges to make development more reliable and not require re-engineering is great.
We saw a breakthrough with TinyRecursiveModels achieve 45% on ARC-AGI-1 and 8% on ARC-AGI-2 using 7M parameters. Special purpose built for solving mazes but incredibly small and the concept can be applied elsewhere.
Z-Image just came out from Alibaba and is a 6B parameters text-to-image model that's really good and open-source. I haven't read the paper but it's also apparently very open about the model architecture and data it used.
Fulfilling and meaningful work
I'm feeling grateful for the opportunity at Nava to lead an infrastructure team and go from zero to one. In a way, we're only now approaching Day 1. It just feels great to learn, grow, teach, support, and just try my best with a supportive team.
Strata is our push to make new software programs repeatedly successful through open source tooling. Happy to contribute to Strata a small bit.
Closing Statement
It's been an incredible year of discovery and learning. In the year of AI full steam ahead, the steps are the same.
As karpathy-sensei put it,
How to become expert at thing:
1 iteratively take on concrete projects and accomplish them depth wise, learning “on demand” (ie don’t learn bottom up breadth wise)
2 teach/summarize everything you learn in your own words
3 only compare yourself to younger you, never to others
So keep learning and keep building.
- ← Previous
On Communication