- What's brewing in AI
- Posts
- š§š¼ OpenAI's o3 mini is here
š§š¼ OpenAI's o3 mini is here
Also: DeepSeek's massive splash
Howdy, wizards.
What a week. The Chinese company DeepSeek made huge waves in the AI industry and beyond. OpenAI dropped o3-mini, and the supercharged o3-mini-high. Sam Altman spilled some juicy tea about upcoming releases in a Reddit AMA.
Letās get into it, shall we? Hereās whatās brewing in AI.
DARIOāS PICKS
OpenAI first announced their new reasoning model o3 in December. The little brother of the model familyāo3 miniāis now officially here, and itās quickly becoming a favourite among ChatGPT users.
o3 mini is the smaller, faster and cheaper version of the upcoming (full) o3 model. Itās optimized for STEM domains, and also top of the line when it comes to coding.

That yellow bar to the right is o3-mini-high leading the game when it comes to math. Itās also on top for other STEM benchmarks and coding. Image source: OpenAI
o3 mini is released both in ChatGPT and in the API (so devs can start building apps with it) right away.
o3 mini in ChatGPT:
Available now to Plus/Team/Enterprise/Pro users, thereās two new options in the model picker: o3-mini and o3-mini-high. The difference is in how hard the models thinks before answering, and how many messages you can send:
o3-mini: reasoning level set to āmediumā, limit of 150 messages per day
o3-mini-high: reasoning level set to āhighā, limit of 50 messages per week
ChatGPT Pro users (the $200/month tier) have unlimited messages to the new model
Free users get free 10 messages to o3-mini per day by clicking the āReasonā button
o3 mini in the API:
o3 mini has the cost efficiency and latency of o1 mini, but is a more advanced reasoning model.
It supports some features for developers that makes it way easier to build production-ready apps with, including function calling (can use 3rd party APIs), structured outputs and developer messages.
When using the API, you can set the aforementioned reasoning level yourself, to either ālowā, āmediumā or āhighāāwhichever suits your use case (e.g. you can choose to optimise for speed when problems arenāt that complex).
ā Why it mattersā ā I tested o3-mini-high on some fairly complex situations where it had to calculate my taxes based on different scenarios, and it rocked! GPT-4o spewed out nothing but faulty calculations (which looks convincing on the surface but are not correct) on the same questions.
However, since o3 mini doesnāt yet support file uploads, it canāt reason over documents and images. Pretty sure Iāll find myself still using o1 a lot until it does. Itās coming soon, though.
Oh, and if youāre wondering when the full o3 model is expected to launch, I interpret OpenAIās Reddit answer to mean ā 1-3 months from now.
TOGETHER WITH GAMMA
An entirely new way to present ideas
Gammaās AI creates beautiful presentations, websites, and more. No design or coding skills required. Try it free today.
DARIOāS PICKS
DeepSeek, a Chinese company founded by a hedge fund manager with a team apparently building AI models as a side project, trained a really advanced, open weights AI model. They did so with dramatically less compute and in less time compared to the leading, US-based tech companies ā with an order-of-magnitude reduction in costs. DeepSeek even reached the #1 most downloaded app on Appleās app store this week.
This led to a huge selloff in US tech stocks with Nvidia shaving off $600b in a market cap in a single day and even Trump calling it a wake-up call for the industry.
The impact of DeepSeekās release is multifaceted and extends far beyond mere technological advancementāinto topics like economics, politics, privacy, censorship, export controls, open source, and more. Or as The Dude would say: lotta ins, lotta outs, lotta what-have-yous.
Hereās the main deets:
Performance-wise, DeepSeekās model, R1, is a very capable reasoning model comparable to OpenAIās o1 and o3. While it doesnāt outshine the top models on the market including o1/o3 and Claude 3.5 Sonnet, it stands as one of the best free AI models. I would have said the best free model if it werenāt for OpenAI making o3-mini available to free users this week (albeit with very limited usage).
Arguably the most notable thing about DeepSeekās R1 is that they managed to train it super cheap, based on reinforcement learning and efficient architectures like Mixture-of-Experts. The model was reportedly trained for about $5.5Māvastly lower than GPT-4ās estimated $100M. However, notable voices such as Anthropic CEO Dario Amodei suggest that while the actual cost is probably lower than for US-based top models, the actual figure isnāt even close to what DeepSeek claims. Amodei also says the cost reduction isnāt that surprising but rather a āexpected point on an ongoing cost reduction curveā.
Something ironic in all this is that the US has been restricting exports of its most advanced AI chips to slow Chinaās progressāa move that now appears ironic as DeepSeekās breakthrough rattles the market (a sentiment echoed on Reddit š).
Privacy and free speech remains concerns with DeepSeek. The model is completely tight-lipped when discussing the Chinese government, and a recent data breach apparently exposed 1M log files. For increased security, itās totally possible to run it locally on your laptop through a tool like LM Studio, but the performance will be very much reduced.
Related: DeepSeekās virality this week mightāve prompted OpenAI to release ChatGPT Gov ā a specialized ChatGPT version for US agencies that contains their data in a secure Azure environment. According to the company, 3,500 US Government agencies are already using AI (hereās a list Iāve with 21 of them).
OpenAI is saying DeepSeek copied them, using their model to train their own. True or not ā itās meme-worthy:
Good. How security is tight on the back end now.
ā Jolly Roger (@dontcallmeraylo)
9:40 PM ā¢ Jan 31, 2025
ā Why it mattersā ā DeepSeekās model shows that thereās no wall in sight yet when it comes to acceleration of AI. It shows itās possible to train intelligent models with a fraction of the resources we thought we needed. It increases the competition at the model layer of the industry, and consumers are left with free, powerful AI models. It also seems to have made OpenAI rethink itās direction of closed models.
Short-term this is bad news for NVIDIAās stock price, no doubt. The company sits at the heart of the industry with its advanced AI chips that everyone relies on to train their models. And everyone is now scared theyāll not have the same demand since you need less compute for the same result. Another way to look at this is that A) it expands the market as a lot more companies will now have the money to train models from scratch and achieve great outputs, such as highly advanced models for very specific domains and B) the big companies will simply use the improved economics to train even more powerful models ā not cutting costs. Not financial advice, btw.
TOGETHER WITH HUBSPOT
Unlock the potential of your growing startup with HubSpot's Starter Customer Platform. For a discounted price of just $20/month, unlock access to the Starter edition of HubSpotās six core products for marketing, sales, and customer service ā powered by HubSpotās Smart CRM ā all for the price of one. Built for startups like yours, HubSpot Starter has all the essential tools you need to scale.
DARIOāS PICKS
Six key people from OpenAI did a Reddit AMA (Ask Me Anything) session yesterday and, similar to last time they did it, teased some pretty cool things.
Check out the full AMA to see all the answers.
Hereās the answers I found most interesting (ranked order):
OpenAI is thinking of going more, well, open. A Redditor asked if theyāre going to release model weights and more research, to which Sam Altman said theyāre discussing it, have been āon the wrong side of historyā with their current approach, and need to figure out a different open source strategy. He also noted not everyone internally agrees on this and that itās not currently their highest priority.
Very soon, OpenAIās models will start showing much more detail of how they are thinking soon, inspired by DeepSeekās R1. According to CPO Kevin Weil, though, itās a balance because showing all of the thinking makes it easier for competitors to copy.
Operator on the regular Plus plan is currently months away. Theyāre currently held back by costs, and are currently training specialised Operator models that are faster and cheaper.
It might become easier for EU companies to build with OpenAI. Theyāre currently working on a EU data residency offering, and will be doing a broad rollout of this at some point.
OpenAI is working on an image generator that uses GPT-4o. To be released in a couple of months, and claimed to be āworth the waitā.
Sam Altman thinks that a recursive self-improvement for AI looking more like a hard take off rather than gradual improvement is more likely than the thought a couple of years ago (translation: AI might get much better ā very fast).
OpenAI has a lot of components for a complete AI automation suite already: voice, video, screen share, tasks and operator. According to Sam Altman, the most critical missing pieces for a complete AI experience are improved reasoning capabilities, automation in any environment (not just the browser), and the ability to understand and use any apps. Also says, ācontinuous video in and video out would be really coolā.
Altman thinks the aspect of AGI that will improve quality of life the most is the ability to accelerate the rate of scientific discovery.
There will be a new model at some point named GPT-5, not GPT-5o. Altman has highlighted that solving their naming problem is a key goal for 2025.
QUICKFIRE HEADLINES

Side notes
Here are the other essential updates on AI you shouldnāt miss this week, that I didnāt have time to cover in-depth.
OpenAI is partnering with US National Laboratories giving scientists access to its most advanced AI models for research, including nuclear weapon security.
Related: The advocacy group Bulletin of the Atomic Scientists just moved their āDoomsday clockā to 89 seconds to midnight. It captures threats including climate change, nuclear weapons, instability in the Middle East, pandemic threats, and use of AI in military operations.
Anthropic CEO Dario Amodei weighs in on DeepSeek in a new essay
New, clearer guidelines on AI-generated works from the US copyright office
THATāS ALL FOLKS!
Was this email forwarded to you? Sign up here. Want to get in front of 13,000 AI enthusiasts? Work with me. This newsletter is written & curated by Dario Chincha. |
What's your verdict on today's email? |
Affiliate disclosure: To cover the cost of my email software and the time I spend writing this newsletter, I sometimes link to products and other newsletters. Please assume these are affiliate links. If you choose to subscribe to a newsletter or buy a product through any of my links then THANK YOU ā it will make it possible for me to continue to do this.