The most important AI news and updates from last month: Dec 15, 2025 - Jan 15 2026
AI Socratic Events
Models Leaderboard

Link: aisocratic.org/leaderboard
NVIDIA Launches Vera Rubin
NVIDIA launches Vera Rubin, the successor of Blackwell. Here how it compares to Blackwell:
- 10x lower inference costs
- 4x fewer GPU required for MoE training
- 5x more energy efficient
- 5x longer uptime
- 10x higher reliability
- 18x faster assembly and maintainance.


sources: https://x.com/nvidia/status/2008357978148130866
We're In "Vibe" Code Age
Claude Code is an all rage right now. Last month Anthropic released Opus 4.5 and that changed everything. Some are calling it AGI in the OpenAI terms as a system that "outperform humans in most economically valuable work" — of course is not the case in most fields, but it's the case for coding.
Claude Code was released Apr/May 2025. It started as a side project and now is arguably the most productive coding tools ever created.
Claude Code Chrome Extension
Claude just added a Chrome extension that enable to run automations from the CLI or directly from the browser. So while Arc/Dia, OpenAI, Perplexity, and Google all launched their own browser, Anthropic went the extension way — in our opinion the correct way.
Random
My Twitter feed is all about “Claude Code". Yet 80% of the American public have no idea what is Anthropic.
The main reason why most engineers are preferring Claude Code to Cursor is simplicity. Claude Code is just a CLI and nothing else to it. The question then becomes for how long Cursor will maintain the lead as IDE if all you need is a CLI now.
You might remember @Karpathy last month, talking at the @Dwarkesh podcast talking about AI slowing down and expanding his projection for AGI. Well he recently shared his sentiment "I've never felt this much behind as a programmer" and admitting that Claude Code is a big paradigm shift in how we code and is only going to increase https://x.com/karpathy/status/2004607146781278521.
Another sign of the tectonic shift is the fact that engineers today build in hours what used to take weeks. This Google Engineer tweet is also all over my twitter feed:
Ralph Wiggum: Autonomous Loops for Claude Code
Claude Code’s official plugin marketplace includes a curious entry: ralph-wiggum. Named after The Simpsons character, it implements autonomous development loops where Claude works for hours without human intervention.
Pretty much what is does is looping through Claude Code until the problem asked is solved. Yes is bad. Yes, it might actually work.
while :; do cat PROMPT.md | claude ; done
https://paddo.dev/blog/ralph-wiggum-autonomous-loops/
Second Order Effects
Tailwind just laid off 75% of the people on their engineering team "because of the brutal impact AI has had on our business."

link: https://github.com/tailwindlabs/tailwindcss.com/pull/2388#issuecomment-3717222957
Research & Papers
DeepSeek 🐋 > mHC: Manifold-Constrained Hyper-Connections
This is a new banger paper from DeepSeek!
Traditional residual connections (e.g., in ResNets and Transformers) add the layer output to the input, preserving an "identity mapping" that enables stable training in very deep networks. Hyper-Connections (HC), a more recent idea, expand this by widening the residual stream (multiple parallel streams instead of one) and using learned mixing matrices for richer information flow and better expressivity. However, unconstrained HC breaks the identity property, leading to severe training instability (exploding/vanishing gradients) and high memory overhead, limiting scalability.Core Innovation: mHCmHC fixes HC by projecting the mixing matrices onto a specific mathematical manifold — the Birkhoff polytope (doubly stochastic matrices, where rows/columns sum to 1). This is achieved efficiently using the Sinkhorn-Knopp algorithm (an iterative normalization from 1967, ~20 iterations suffice).Key benefits:
- Restores bounded signal propagation (gain stays ~1-1.6 across layers, vs. exploding to 3000+ in plain HC).
- Enables stable widening of the residual stream (e.g., 4-8x wider) for better performance.
- Promotes controlled information mixing across depths, improving representation learning.
Efficiency OptimizationsDeepSeek added heavy infrastructure tweaks (kernel fusion, recomputation, communication overlapping) to keep overhead low (~6-7% extra training time).ResultsExperiments on models up to 27B parameters show:
- Better downstream performance (e.g., on reasoning benchmarks like GSM8K) than standard residuals or unstable HC.
- Superior scalability, with hints from "in-house large-scale experiments" suggesting it's production-ready (likely for DeepSeek's next models, e.g., V4).
In essence, mHC makes a theoretically superior but previously impractical idea (wider, diversified residuals) viable at scale, potentially unlocking new ways to improve LLMs beyond just more parameters or data. It's seen as a fundamental advance in topological architecture design, with community excitement around implementations and combinations (e.g., with value residuals). The original X thread you linked is a fan announcement hyping it as a "huge model smell" breakthrough.

Sources:
- Paper: mHC: Manifold-Constrained Hyper-Conncetions
- Blog: Understanding Manifold Constrained Hyper Connections
Neural Networks at scale all converge to a statistical model of reality and internal structure.
🌌 The Universal Weight Subspace Hypothesis
Johns Hopkins University reveals that neural networks, regardless of task or domain, converge to remarkably similar internal structures.
Their analysis of 1,100+ models (Mistral, ViT, LLaMA) shows they all use a few key "spectral directions" to store information.
This universal structure outperforms assumptions of randomness, offering a blueprint for more efficient multi-task learning, model merging, and drastically cutting AI's computational and environmental costs.
🏛️ The Platonic Representation Hypothesis
Neural networks, trained with different objectives on different data and modalities, are converging to a shared statistical model of reality in their representation spaces. Vision models, language models, different architectures are all slowly approximating the same underlying model of reality.
If this holds up, it's a huge unlock. We could translate between models instead of treating each one like a sealed black box, reuse interpretability wins across systems, and maybe align models at the representation level, not just by policing outputs.
The crazier implication is philosophical. Maybe MEANING isn't just a human convention. Maybe there are natural coordinates in reality and sufficiently strong learners keep rediscovering them.
So what's actually driving the convergence? The data, the objective, some deep simplicity bias? And where does it break?
Sources
- The Universal Weight Subspace Hypothesis > tweet
- The Platonic Representation Hypothesis > tweet
- Michael Levin - Symposium on The Platonic Space
Videos And Podcasts
The Ridiculous Engineering Of The World's Most Important Machine
The insane machines that make the most advanced computer chips from Veritaseum.
Other podcasts and videos from December 2025:
- MLST: PhD Bodybuilder Predicts The Future of AI, 97% Certain — Dr. Mike Israetel, this is mostly a fun cross-over debate between 2 podcasters. Please take it with a grain of salt and just as a fun watch.
- MLST: AutoGrad Changed Everything, Not Transformers — Dr. Jeff Beck, Dr. Jeff Beck, mathematician turned computational neuroscientist, this is a deep dive into why the future of AI might look less like ChatGPT and more like your own brain.
- Dwarkesh Podcast: Adam Marblestone – AI is missing something fundamental about the brain, Adam Marblestone is CEO of Convergent Research and a former DeepMind neuroscientist. We discuss how the brain learns so much from so little, what AI can learn from neuroscience, and Ilya’s question of how the genome encodes abstract reward functions—ultimately, the same problem.
Funding
It's interesting to see how the new purchases are taking place, they're not anymore the classical acquisition, but closer to M&A, why is that? Because businesses found a way to purposefully avoid triggering antitrust scrutiny. The script is the same as Meta and ScaleAI or Google and WIndsurf. Now NVIDIA and Groq, and Meta and Manus follow the same script.
NVIDIA "buys" Groq at $20B
Groq has entered into a non-exclusive licensing agreement with Nvidia for Groq’s inference technology. Groq Cloud will continue to operate without interruption.
This blog post breaks down the antitrust loophole that enabled NVIDIA to close this $20B deal: https://ossa-ma.github.io/blog/groq.
What Nvidia Actually Bought (And What It Didn't)
Nvidia acquired:
- All of Groq's intellectual property and patents
- Non-exclusive licensing rights to Groq's inference technology
- Jonathan Ross (CEO), Sunny Madra (President), and the entire senior leadership team
Nvidia explicitly did NOT buy:
- GroqCloud (the cloud infrastructure business). GroqCloud continues as an independent company under CFO Simon Edwards. This is Nvidia's largest acquisition ever (previous record was Mellanox at $7B in 2019), and they structured it to leave the actual operating business behind. That doesn't happen by accident. Part of the reason for not acquiring GroqCloud is because Saudi Arabia's company Dammam is using their cloud for their AI service inference. Had NVIDIA entered in business with KSA (Kingdom of Saudi Arabia) there could have been more scrutiny involved.
After this acquisition Jonathan Ross went from building TPUs at Google, to building LPU at Groq, and now is moving to NVIDIA (GPU). Why are LPU so important to NVIDIA?
OK, what's the difference between GPU, TPU and LPU?
META buys Manus at $2-4B
Meta just bought Manus, an AI startup everyone has been talking about https://x.com/TechCrunch/status/2005876896316469540.
- Manus hit $100M ARR in ~9 months and sold for an estimated $2–4B, signaling consumer-facing agent products can scale revenue fast and command software-like multiples; this validates “agentic” AI as a real, monetizable category for Big Tech.
- This was a strategic tech acquisition: Meta gains a team praised for “scaffolding powerful agents” and claimed SOTA on the Remote Labor Index, accelerating Meta’s push to automate complex white‑collar workflows (coding, ops, “computer use”) with production‑grade agents.
- Market consolidation and global expansion: the fastest-growing B2C AI player joining Meta concentrates top agent expertise and infrastructure under one roof, likely speeding platform standardization and distribution; early hiring in Singapore hints at global scaling of agent products.
GeoPolitics
Note: our mission is to democratize AI via open source knowledge and decentralization. With that in mind our community tries to share objective views, without siding with or against any country or company.
Venezuela and Taiwan (TSMC)
United States capturing the Venezuelan president Maduro has large implication in geopolitics. Peter Zeihan, one of my favorite geopolitics expert https://youtu.be/ddojVgGAryQ?si=nfpK2_JZNnjt334Q explains how by taking Venezuela, the US is showing a clear expansionist plan, following the Monroe Doctrine, saying that the US should keep influence in the entire West Hemisphere.
Right after this attack, president Xi Jinping has pledged to achieve "reunification" of China and Taiwan link: https://www.aljazeera.com/news/2026/1/1/chinas-xi-says-reunification-with-taiwan-unstoppable.
The risk with a Taiwan invasion is obviously much larger, as TSMC (Taiwan Semiconductor Manufacturing Company) the world's largest and most important pure-play semiconductor foundry, they produce 90% of the world most advanced semiconductors, from NVIDIA GPUs, to iPhone chips, cars chips, and even defense systems. An attack could cause a halt to production across this fabs, causing a global economic shock — couldn't be surprised if the burst of the economic bubble would starts in this scenario. 🛡️ The “Silicon Shield” refers to the idea that Taiwan's semiconductor dominance actually deters any attack.
According Reuters: China built a prototype extreme ultraviolet lithography (EUV) machine in Shenzhen, the tool needed for the most advanced https://x.com/Megatron_ron/status/2001637940988899683. So the question is not if China will build their own EUV and CUDA but when. We know the US is trying to do the same, and possibly the EU too. Until then we believe Taiwan will remain shielded from attacks.

These interesting blog posts from SemiAnalysis.com can shine some light on TMSC:
- Semi Analysis - TSMC Overseas Fabs – A Success? https://newsletter.semianalysis.com/p/tsmc-overseas-fabs-a-success
- Semi Analysis - Why America Will Lose Semiconductors https://newsletter.semianalysis.com/p/why-america-will-lose-semiconductors?utm_source=publication-search
- Equity Analysis - TSMC: The Silent War for Global Dominance is Being Fought HERE, https://substack.com/home/post/p-167371968
Full Source List
AGI
- ⭐️Claude Opus 4.5 in Claude Code is AGI. By the open AI definition? Can this system "outperform humans in most economically valuable work"? Depends a lot on how you define "humans" and "economically valuable work" obviously. https://x.com/deepfates/status/2001047747110334516.
- We saturated ARC AGI-2 before the end of 2025 https://x.com/chatgpt21/status/2003577774121058768.
- Remember the AI 2027 blog post? Eli and Brendan adjusted the predictions of the AI taking off on a longer timeline, so ASI for 2034: https://x.com/DKokotajlo/status/1916520276843782582.
- 2026 AI predictions: 1. verification over belief, 2. the adoption gap becomes real, 3. from tool users to system owners, 4. AI science, mundane tasks in robotics, new architectures, new types of education https://x.com/TheTuringPost/status/2006564527920533801.
- @ZhengdongWang write about "feel the AGI" featuring compute, inevitability, and second-order effects https://x.com/zhengdongwang/status/2005848098531106916.
- New paper: we argue AGI may first emerge as collective intelligence across agent networks, not a single system. This reframes the challenge from aligning one mind to governing emergent dynamics: more institutional design than single-agent alignment https://x.com/sebkrier/status/2002020453611651308.
- AI Futures Project (authors of AI 2027) have released an updated model, with somewhat longer timelines https://x.com/timfduffy/status/2006223326235222359.
AI Agents
- ⭐️ Greg Brockman: two big themes of AI in 2026 will be enterprise agent adoption and scientific acceleration https://x.com/gdb/status/2006584251521839141.
- Making code agent ready will see 5-10x returns on those who doesn't https://x.com/EnoReyes/status/2003483475157127650.
- 10x deep research via GRPO: the intuition has always been that scaling agentic search is a compute problem, it's actually a "stability-of-objective" problem, most 8B models suffer from "horizon collapse", they are mathematically "anxious" to terminate the search loop because their training (SFT) forces them to mimic narrow, high-entropy trajectories rather than optimizing for the global reward https://x.com/byebyescaling/status/2003900947488227381.
- How do we make codebases that are agent ready? At @aiDotEngineer, our CTO @EnoReyes breaks down why agents need tight verification loops to succeed, and why most codebases don’t provide those signals yet. Teams that invest in agent readiness will see 5–10× returns https://x.com/FactoryAI/status/2006431174512828838.
- nothing prepared me for parallel agentic work as much as starcraft https://x.com/neilsuperduper/status/2006075204104303037.
- We're releasing Letta Code, a memory-first coding agent - open source (apache 2.0) - model agnostic - portable agent learning and memory https://x.com/Letta_AI/status/2000998273084264531.
- 2026 will be a year of scientific acceleration through AI. just released a benchmark to measure the capability of AI on expert-level scientific reasoning: https://x.com/gdb/status/2001013618306138401.
- Google just open-sourced A2UI protocol. AI Agents can now generate native UIs on the fly instead of sending walls of text. Flight booking cards, interactive forms, date pickers - all streamed directly to users https://x.com/unwind_ai_/status/2001185205391081498.
- Memory in the Age of AI Agents: A Survey Delves into the rapidly evolving field of agent memory, providing a unified taxonomy through the lenses of forms, functions, and dynamics. This survey offers a critical conceptual foundation for designing future agentic intelligence https://x.com/HuggingPapers/status/2000840668302676053.
AI Builders
- ⭐️ RIP Prompt Engineering, the jump from "context-aware" (1.0) to "context-cooperative" (2.0 and beyond) systems changes everything from memory design (e.g., layered architectures ) to multi-agent collaboration (e.g., shared memory and subagents ). https://x.com/mdancho84/status/2000908364574794217.
- ⭐️ This guy literally drops the ultimate guide to cutting AI costs 10x https://x.com/yulintwt/status/2003144359756333540.
- ⭐️OpenAI adopted Anthropic's Skills mechanism in chatgpt and codex https://x.com/simonw/status/1999623295046664294.
- ⭐️ Claude Code was a side project in September 2024, today is a core dev tool for many, used far beyond coding. A year ago it struggled with basic bash; now, in the last 30 days alone, I’ve shipped 259 PRs—every line written with Claude Code + Opus 4.5, running for hours or days at a time. Software engineering is changing, and we’re just getting started. https://x.com/bcherny/status/2004887829252317325.
- ⭐️ You can now use Claude Code with OpenRouter https://x.com/OpenRouterAI/status/2002067582569066979.
- ⭐️ Codex now officially supports skills [https://x.com/OpenAIDevs/status/2002099768479396123](https://x.com/OpenAIDevs/status/2002099768479396123].
- ⭐️ Open prose, interesting prompt language https://x.com/irl_danB/status/2007618496092618872.
- This is the definition of vibe coding to me Managing the context, looking at agent responses, taking the time to ask the right questions: - is this secure? - how does this work? - can I do {x}? why? why not There's no such thing as non-technical any more! https://x.com/iannuttall/status/2006373057955615171
- Brilliant post on using coding agents. The workflow described here is as close as it gets to my own. From creating rules and skills to optimizing workflows, testing, and more https://x.com/omarsar0/status/2002773470636281963
- codex now supports skills, per the agentskills.io standard: https://x.com/gdb/status/2002120466203615649
- let me guess you still don't have an ontological map of your codebase and architecture that your agents can use as a kb to update dynamically https://x.com/tunahorse21/status/2001023828395487419
- We created a GitHub repo for all MCP at @Google. Get info on our remote managed MCP servers, open source MCP servers, examples, and learning resources. github.com/google/mcp https://x.com/_philschmid/status/2001124271343448264
- Tasks can now spawn async subagents that move to the background and continue working independently, even if the main agent finishes its task and becomes inactive. This lets subagents handle long-running tasks, great for monitoring logs or waiting for builds https://x.com/claudeai/status/1998830341021380875
- @Karpathy shortly after stating at the Dwarkesh podcast that AI is slowing down, had a cathartic moment and realized Claude Code makes him a 10x engineer https://x.com/karpathy/status/2004607146781278521
- Anthropic co-founder, Jack Clark: by summer 2026, the AI economy may move so fast that people using frontier systems feel like they live in a parallel world to everyone else most of the real activity will happen invisibly in digital, AI-to-AI spaces, with only surface signs https://x.com/slow_developer/status/2003903502012256332
- The change from the @dwarkesh_sp podcast 2 months ago vs. the tweet below from @karpathy is genuinely insane It’s a night and day difference. We went from “these models are slop and we’re 10 years away” to “I’ve never felt more behind and I could be 10x more powerful” This all https://x.com/Midnight_Captl/status/2004717615433011645
- Anthropic engineers are saying they haven’t opened an IDE for over a month and you think we won’t see accelerating progress over the next year, anon? https://x.com/TMTLongShort/status/2004720387847061799
- this is the most used slash command internally at cursor to remove ai slop https://x.com/ericzakariasson/status/1995671800643297542
- I definitely have some days where I wonder what I am bringing to the table, beyond being Claude's QA tester. This feature is a little on a the nose, but still pretty useful https://x.com/dmwlff/status/2001029340201722178
- Improving RAG with Forward and Backward Lookup This is a clever use of small and large language models. Traditional RAG systems compute similarity between the query and context chunks, retrieve the highest-scoring chunks, and then generate. But complex queries often lack https://x.com/dair_ai/status/2001306474803540415
- agi has been achieved internally claude code skill to manage gemini cli in headless mode via bash timeouts https://x.com/tunahorse21/status/2002880735116685355
- huh is this where the new claude hype comes from https://x.com/gbrl_dick/status/2002883524542824951
- 1 year of subscription for GLM 4.7 near Opus 4.5 levels = 1 month of Max Plan of Codex/Claude Code . These prices are ridiculous. This is the future, wether you like it or not. The reason why many Ai companies in the west will fail. Buying 1 year right now https://x.com/aeitroc/status/2003205591552905309
- i am becoming a fan of OpenCode dare i say, OpenCode > Claude Code https://x.com/TheAhmadOsman/status/2003692952363921605
- Claude Opus 4.5 is the most revolutionary software created maybe in our lives and people are just starting to realize it https://x.com/AlexFinn/status/2004796428795916350
- Man who uses GPT-4o in Windsurf: AI sucks at coding Man who uses Opus 4.5 in Claude Code: we already have AGI https://x.com/vasuman/status/2004803875103277427
- Claude Code and Codex (GPT-5 High). Complete Vibe Coding Guide. We build 12 apps in this video of varying difficulty https://x.com/rileybrown/status/1966275792427712978
- Someone educate me
- OpenAI - Claude - Manus - Gemini How do you determine when to use each one? https://x.com/RyanClogg/status/2005758658034925973
- Advent of Claude Day 31 - Claude Agent SDK I saved the best for last. The same agent loop, tools, and context management that power Claude Code is available as an SDK. Build agents that work like Claude Code in as little as 10 lines of code. This is just the beginning. 🎇🥂 https://x.com/adocomplete/status/2006399579554754687
- must suck to be writing code at any ai lab that's not anthropic right now (you can't use opus 4.5) https://x.com/paularambles/status/2006428443060351262
- Great read for AI devs. (bookmark it) LLM agents are slow. The bottleneck in complex agentic systems today is the planning part. Plan generation alone can take 25+ seconds for task requests. This compounds fast at scale. Real-world dataset analysis shows about 30% of https://x.com/omarsar0/status/2005799762252136537
- Anthropic crash course to master Claude Code from scratch for free https://anthropic.skilljar.com/claude-code-in-action
- Claude Max: $99 My Usage: $850 Good deal @claudeai https://x.com/_skris/status/2000492729667932441
- unpopular: it doesn’t make sense to write Python anymore. AI is just too good for it https://x.com/Hesamation/status/2004983462147826162
- This right here is why Palantir is a valuable company btw https://x.com/akoratana/status/2005303231660867619
- “If you don’t review AI code it just makes a mess of slop.” No, your desire to keep your code human-friendly is creating an unnecessary bottleneck. Stop looking at the code. Let the AIs handle it https://x.com/nateliason/status/2005000034975441359
- This seems fine https://x.com/jiratickets/status/2005717674765242605.
AI Tools
- ⭐️ Claude in Chrome is now available to all paid plans. We’ve also shipped an integration with Claude Code https://x.com/claudeai/status/2001748044434543082
- AGI isn't a model. Logan Kilpatrick (@OfficialLoganK) says the "AGI moment" won't necessarily come from a frontier model upgrade. “My guess is that the AGI experience isn’t just a super powerful model. It’s plausible someone takes an off-the-shelf model, puts the rest of the https://x.com/ForwardFuture/status/2006455388770590783
- 🧠 Mind as receiver, not creator. Psychologists are exploring a fascinating idea that challenges how we understand the human mind. Thoughts may not be something we actively produce. Instead, they may arise, arrive, or surface from deeper subconscious processes beyond conscious https://x.com/ShiningScience/status/2006393307384180917
- Ever seen visually how joint, marginal, and conditional distributions actually relate? check it out on tensortonic and manipulate the values by yourself https://x.com/prathamgrv/status/2006654457631957399
- therapy only works if you have a lower IQ than the therapist https://x.com/4jibola/status/1999029644050444361
- Anthropic's @_sholtodouglas says Dario Amodei is chronicling the path to AGI through a series of essays he writes in the company's Slack channels: "He has a really, really cool communication style. He quite frequently puts out very well-reasoned essays. And throughout Slack, https://x.com/tbpn/status/2006515488462082452.
Benchmarks
- lmao cost to get to 85% on arc-agi-1 went down 26000x in a year https://x.com/Bayesian0_0/status/2001338959914557802
- 3 flash beats out 3 pro on a few benchmarks, quite insane https://x.com/adonis_singh/status/2001325480516182365
- Gemini 3 Pro playing Pokémon vs 2.5 Pro (we used to all be impressed by 2.5 Pro) https://x.com/OfficialLoganK/status/2000728193599226187
- When we reported scores for Gemini 3 Pro, we didn't include reasoning levels, just the default score Gemini 3 Flash High is showing better perf than Gemini 3 Pro, wait to draw any conclusions until we the other g3p levels on there https://x.com/GregKamradt/status/2001340881052504570
- We were surprised by Opus 4.5’s big performance leap on CORE-Bench, especially compared to Opus 4.1, so @PKirgis decided to look deeper. The gains are genuine: Opus 4.5 is much better (and cheaper) at recovering from errors, resolving dependency issues, and solving tasks https://x.com/sayashk/status/2000972874669048226
- Poetiq surpass human level at ARC-AGI-2 https://x.com/poetiq_ai/status/2003546910427361402
- LLM IQ https://x.com/slow_developer/status/2007016514554146899.
Blog Posts
- ⭐️ This is the darkest and most moving essay about AI and job displacement that I’ve ever read. https://x.com/alexolegimas/status/2006190451460575590
- ⭐️ Wrote this for young people worried about AI. A positive essay from an OpenAI Staff Engineer from the alignment team. https://x.com/boazbaraktcs/status/2006525672412622945
- ⭐️ AI’s 2025: compute wins, evals struggle, and Britain needs AI‑assisted decisions. Moravec’s maxim holds (Jagged intelligence); scale turns parrot into polymath, METR’s horizons lengthen, and UK progress depends on forecasting + coordination over fish discos. https://x.com/SamuelAlbanie/status/2005972601617932356
- @LouisBRosenberg argues that conversations about an "AI bubble", hating on AI and calling it slop, is a defense mechanism against the grief for the massive change that’s coming https://x.com/bigthink/status/2001336898539671717
- The Jagged AI Frontier is a Data Frontier Wrote up my thoughts on how coding and math succeeded through data abundance and verification, why science lags behind, and how distillation equalizes the field while RL may just be another data generator https://x.com/lvwerra/status/2000951037826588956
- In this post, we share 11 use cases at the intersection of crypto and AI to help kickstart conversations about what’s possible, what challenges are left to solve, and more — grounded in technology already being built today https://x.com/a16zcrypto/status/1994921546276900917
- @patrick_oshag One of the more compelling write ups on AI investments https://x.com/thesignalpilot/status/1999869473554313677
- Are we building a thing like Wired back when Wired was cool. I think we are https://x.com/ashleevance/status/2001306468629508410
- what jobs are made of: - AI boosts execution, but jobs hinge on judgment and agency
- Entry roles shrink as execution gets automated; experience builds non-technical value.- Benchmarks soar, yet real-world impact lags due to context, prioritization, and taste. https://x.com/Thom_Wolf/status/2003062043092684971
- AI’s progress feels invisible day-to-day, yet transformative: frontier systems can one‑shot complex software, birthing a fast, parallel “AI economy” that’s reshaping work, infrastructure, and digital spaces. Approach it with technological optimism, and appropriate fear. https://x.com/jackclarkSF/status/2003526145380151614
- AI is the new “miracle material”: from 10× coders to 40× agent orchestrators, organizations reimagined like steel-framed skyscrapers, and economies scaling beyond human rhythms—stop bolting chatbots onto waterwheels; design for infinite minds. https://x.com/ivanhzhao/status/2003192654545539400
- Balance of power Vitalik: the core challenge of the 21st century is preserving pluralism and balance of power as economies of scale explode, which requires mandating diffusion of control, not just redistributing wealth. https://x.com/VitalikButerin/status/2006182090694615084
- AI Futures Model: Dec 2025 Update https://blog.ai-futures.org/p/ai-futures-model-dec-2025-update
- Over the years, my colleague Sanjay Ghemawat and I have done a fair bit of diving into performance tuning of various pieces of code. We wrote an internal Performance Hints document a couple of years ago as a way of identifying some general principles and we've recently published a version of it externally. https://x.com/JeffDean/status/2002089534188892256
- I argue that "Physics of AI" research is hard due to the current publishing culture. But there is a simple solution curiosity driven open research. https://x.com/ZimingLiu11/status/2006810684546494522
- This image so wonderfully and succinctly illustrates how the relationship with AI feels. Currently excelling at the most complex tasks, but sometimes failing at simple ones https://x.com/tomaspueyo/status/1993360931267473662.
Cybersecurity
- @Sam Altman: a security researcher using our previous model found and disclosed a vulnerability in React that could lead to source code exposure. https://x.com/sama/status/2001724828567400700.
Funding
- ⭐️ Groq has entered into a non-exclusive licensing agreement with Nvidia for Groq’s inference technology. GroqCloud will continue to operate without interruption https://x.com/GroqInc/status/2003945400181608843
- ⭐️ Google, along with Anthropic, is the big winner in 2025. Gemini's market share rose from 5.64% to 14.95% (trippled), while ChatGPT's fell by 4% to 74% https://x.com/kimmonismus/status/2003089654103322799
- ⭐️ Meta just bought Manus, an AI startup everyone has been talking about https://x.com/TechCrunch/status/2005876896316469540
- Manus AI, acquired by Meta, is a current SOTA with a 2.5% achievement on the "Remote Labor Index" benchmark (Introduced by Meta). The Remote Labor Index measures AI automation of remote work. Soon to know how much 2.5% is worth 👀 https://x.com/testingcatalog/status/2005780266875494796
- The case for Meta AI in 2026 With yesterday’s acquisition of Manus, it’s worth revisiting where Meta stands in the consumer AI game - and what we might see from them next year. Somewhat shockingly, Meta is one of the few tech giants without a true AI product in the market https://x.com/omooretweets/status/2006121220677546355
- Meta just acquired Manus AI. Ramp Sheets modeled it out: Estimated price: $4-6B based on AI M&A comps Fastest to $100M ARR in history (8 months) Benchmark likely 8-12x'd in under a year https://x.com/RampLabs/status/2005807066351325470
- Manus is entering the next chapter: we’re joining forces with Meta to take general agents to the next level https://x.com/ManusAI/status/2005766053813707003
- Almost $1 trillion has been invested in AI so far, and this may just be the beginning https://x.com/TheVixhal/status/2001292137770406223
- The downside of the AI revolution: prices in the consumer market are exploding. The RTX 5090 could rise to $5000 https://x.com/kimmonismus/status/2006443618484228396
- BEARISH ON OPENAI The investment case for OpenAI has never been more precarious than it is right now in late 2025. What was once a company that seemed destined to dominate the artificial intelligence revolution has revealed itself to be a structurally disadvantaged challenger https://x.com/DaveShapi/status/2001692151986434480
- "Microsoft Scales Back AI Goals Because Almost Nobody Is Using Copilot" https://x.com/VaughnVernon/status/1999937374336479614
- 2026 is the GREATEST time to build a startup in 30 years I’m 36. I’ve sold 3 startups, helped build companies that raised billions, and backed teams from seed to unicorn. 20 MEGA shifts that make this the BEST time to build in a GENERATION: 1. Hardware got smart. Download https://x.com/gregisenberg/status/2006382073418752250
- what i think is interesting is like either ai fails: debt defaults and bubble explodes economy ai succeeds: kills jobs, puts billions(?) out of work, explodes economy is there a "winning" scenario im missing idk https://x.com/JezCorden/status/1994740799888588948
- Google is the only company that owns the entire AI stack. 1. Chips: TPUs 2. Infrastructure: Google Cloud 3. Foundation Models: Deepmind 4. Applications: Gemini Every layer matters. The more layers you own, the bigger your advantage https://x.com/Speculator_io/status/2003602998208135454
- Anthropic margin are going to be so high, they're playing a safe game https://x.com/bearlyai/status/1996725054823186457?s=20.
Geopolitics
- ⭐️ Metropolis is pretty much Chongqing China https://x.com/jokieliu/status/2006888668053778823
- What happens when AGI nukes jobs? Let's look at the macroeconomics of the future! Have you ever stopped to ask how money actually gets into your pocket? Not the work you do to earn it, but the actual plumbing of the economy that pushes purchasing power from the top of the https://x.com/DaveShapi/status/2000915724039033156
- In a New Year’s address, Chinese President Xi Jinping said reunification with Taiwan is unstoppable reut.rs/4q0Xzx1 https://x.com/Reuters/status/2006392402609647738
- the US sanctions strategy on China’s chips was always going to fail. Not because China is magically ahead, but because once you tell a civilisation-state of 1.4 billion people that it is never allowed to touch a core technology again, you guarantee https://x.com/commiepommie/status/2001480769005625472
- Every American intellectual will become China-pilled. Every single one https://x.com/alreadydawn/status/2006263615524069522
- El Salvador and @xai will redefine the future of public education https://x.com/nayibbukele/status/2000723301157462450
- US is ahead of China in LLM just by 7 months https://x.com/EpochAIResearch/status/2007202315481317670?s=20.
Hardware
- ⭐️ Reuters: China built a prototype extreme ultraviolet lithography (EUV) machine in Shenzhen, the tool needed for the most advanced https://x.com/Megatron_ron/status/2001637940988899683
- ⭐️ NVIDIA launches Vera Rubin the successor of Blackwell. Here how it compares: 10x lower inference costs, 4x fewer GPU required for MoE training, 5x more energy efficient, 5x longer uptime, 10x higher reliability, 18x faster assembly and maintainance. https://x.com/nvidia/status/2008357978148130866?s=20
- Comprehensive survey on the future of AI-powered chip design. This survey maps the evolution from "AI for EDA" to "AI-Native EDA" to what the authors call "Intelligent Design 4.0": autonomous agents with reasoning-acting-reflecting loops that can interpret intent, orchestrate https://x.com/dair_ai/status/2006457851401875524
- Chinese company Betavolt has begun mass production of its BV100 coin-sized nuclear batteries, which can provide continuous power for up to 50 years https://x.com/Currentreport1/status/1999753061120819486
- Chinese scientists unveil a superfast analog chip, 1000x faster than top digital processors. It solves complex maths for AI & science, outperforming even NVIDIA's GPUs in tests. The race for the best computing power is entering its next phase https://x.com/kimmonismus/status/2002868011904950737
- You are playing inside the simulation, with nvidia and intel, but ASML owns the simulation https://x.com/POTFES/status/2006316283760083000
- so much non-utilized compute on iphones we need more creative solutions to make use of this wasted resource https://x.com/chesterzelaya/status/2000084980505129388
- Jensen admits it takes three years to build a data center, which means Nvidia is “selling” and expecting to “sell” far more GPUs than data centers can actually deploy. So where are all these GPUs? I will have a post about it next week https://x.com/kakashiii111/status/1997457742865510485
- Memory prices up. GPU prices up. Macs are the new AI meta. Apple wins in the end. https://x.com/alexocheema/status/2006437653911310556?s=20.
Learning
- ⭐️ Training LLMs end to end is hard. Very excited to share our new blog (book?) that cover the full pipeline: pre-training, post-training and infra. 200+ pages of what worked, what didn’t, and how to make it run reliably https://x.com/eliebakouch/status/1983930328751153159
- ⭐️ Stanford CME295: Transformers and Large Language Models I Autumn 2025 https://www.youtube.com/playlist?list=PLoROMvodv4rOCXd21gf0CF4xr35yINeOy
- ⭐️Understanding mHC: Manifold-Constrained Hyper-Connections. https://x.com/arjunkocher/status/2007338642403143747?s=20
- Just finished Gen. Leslie Groves’ book on the Manhattan Project. Incredible read! It’s an inside look at the most complex & ambitious engineering project in history, and explains exactly how they did it. My full (unedited) notes: ‣ they melted down $154 billion of silver https://x.com/pronounced_kyle/status/2002963923197157545
- This book is brilliant! It's a fun way to learn and explore maths with Python. This is exactly the way I learnt ML and neural networks back in my PhD days https://x.com/omarsar0/status/2004225894932795583
- I’ve spent 10 years teaching math to machine learning engineers. 80% of university math is irrelevant to your actual job. Here's the 20% you actually need to build models (and how to learn it fast): thepalindrome.org/p/the-roadmap-… https://x.com/TivadarDanka/status/2006492285413052535
- Stay tuned. Math for ML will never feel this easy https://x.com/TensorTonic/status/2000237502985310578.
LLMs
- ⭐️ 6 most popular Policy Optimization algorithms in 2025: 1. PPO (Proximal Policy Optimization) 2. GRPO (Group Relative) 3. GSPO (Group Sequence) 4. DAPO (Decoupled Clip and Dynamic sAmpling) 5. BAPO (BAlanced) 6. ARPO (Agentic Reinforced) Learn more about each method and the key https://x.com/TheTuringPost/status/1999801691538104543
- ⭐️ Continual learning in token space. Similar to ACE (Agentic Context Engineering) (link to: https://x.com/omarsar0/status/1976746822204113072?s=20) https://x.com/charlespacker/status/1999193963564241280
- ⭐️Google’s Gemini 3 Flash is a beast. It’s outperforming Gemini 2.5 Pro across the board and even beating Gemini 3 Pro on ARC-AGI-2 & SWE-Bench Verified. We’re at the point where distilled models (3x-4x smaller) are beating the “frontier” models of just 6 months ago. This https://x.com/Yuchenj_UW/status/2001353730357367202
- ⭐️The growth of Google Gemini is insane https://x.com/LiebermanAustin/status/2002461029725536553
- ⭐️ Google introduces the Titans architecture and the MIRAS framework, which allow AI models to work much faster and handle massive contexts by updating their core memory while it's actively running. https://x.com/DataChaz/status/2003377721049588137
- ⭐️ Really nice video by @juliarturc explaining why Transformers are taking over CNNs in computer vision Love the fact that she uses the Transformers library to try ResNet vs. ViT! Fun fact: she's the author of Pix2Struct, a document AI model we also integrated in Transformers :) https://x.com/NielsRogge/status/2005755831636176980
- ⭐️ Here are my 2026 predictions for how AI will change software: - An agent-native software architecture. Most new software will just be Claude Code in a trench coat—new features are just buttons that activate prompts to an underlying general agent
- Designers get superpowers—and https://x.com/danshipper/status/2006395285459677556
- ⭐️ Google is making progress on their diffusion models... It's now as good as a Gemini 2.0 Flash Lite. The writing is on the wall, a majority of language AI use in the future will be diffusion models https://x.com/IntuitMachine/status/2006497556088631675
- ⭐️ Gemini 3.0 Flash vs Gemini 3.0 Pro Same prompt. Flash is much faster and quality is surprisingly close. Which one do you prefer? https://x.com/mattiapomelli/status/2001694490554458401
- ⭐️ New "Pareto Frontier" on ARC-AGI 27.5% for just $2 333x cheaper than TRM! Beats every non-thinking LLM in existence Cost so low, its literally off the chart Vanilla transformer. No special architectures. Tiny. Trained in 2 hrs. Open source. Thread: https://x.com/evilmathkid/status/2001689479476879448
- ⭐️ Gemini 3.0 flash just dropped 🤯🤯🤯 Benchmarks and pricing is really good!!! Thinking of subscribing to AI pro now https://x.com/donvito/status/2001324702497955963
- ⭐️ Gemini 3 Flash is here. Free, fast, unlimited version of Gemini with our biggest upgrade yet. https://x.com/GeminiApp/status/2001321503779074077
- ⭐️ You can now fine-tune LLMs and deploy them directly on your phone! 🚀 We collabed with PyTorch so you can export and run your trained model 100% locally on your iOS or Android device. Deploy Qwen3 on Pixel 8 and iPhone 15 Pro at ~40 tokens/sec. Guide: docs.unsloth.ai/new/deploy-llm… https://x.com/UnslothAI/status/2001305185206091917
- ⭐️Facebook releases SAM Audio, the first audio model that isolates any sound from complex audio mixtures using text, visual, or span prompts https://x.com/AIatMeta/status/2000980784425931067
- ⭐️ GLM 4.7 is out, bench compares closely to claude opus 4.5 https://x.com/Zai_org/status/2003156119087382683?s=20
- Announcing FunctionGemma, a specialized version of our Gemma 3 270M model that’s fine-tuned for function calling ⚙️ The new release brings bespoke function calling to the edge, and is designed as a strong base for further training into custom, fast, private, local agents that https://x.com/googleaidevs/status/2001703478436184540
- Introducing T5Gemma 2, the next generation of encoder-decoder models, built on the powerful capabilities of Gemma 3. Key innovations and upgraded capabilities include: + Multimodality + Extended long context + Support of 140+ languages out of the box + Architectural improvements https://x.com/googleaidevs/status/2001723428810764625
- Microsoft just released TRELLIS 2 https://x.com/_akhaliq/status/2001041559366598799
- Many SOTA models have over a trillion parameters, but this only only has 10 Billion AAAAAND its open source. Hmm https://x.com/arafatkatze/status/2003367582896914686
- IQuest-Coder, 40B open source model beating large model on bench https://x.com/zephyr_z9/status/2006581083656302853
- Tinker is now generally available. We also added support for advanced vision input models, Kimi K2 Thinking, and a simpler way to sample from models https://x.com/thinkymachines/status/1999543421631946888
- Masked diffusion LLMs are getting a lot of attention. They outperform other diffusion types (such as uniform diffusion) at small scales. But what if I told you that uniform diffusion actually scales better? https://x.com/dvruette/status/2000869455815901604
- GPT-5.2 is our strongest model on the FrontierScience eval, showing clear gains on hard scientific tasks. But the benchmark also reveals a gap between strong performance on structured problems and the open-ended, iterative reasoning that real research requires https://x.com/OpenAI/status/2000975294694613200
- openAI lead has tightened in 2025 but a lot of models now feel similar in pre-training, so the real edge comes from post-training, especially RL gpt-5.2 has some unique strengths, but they don’t show up in typical demos the bigger leap is longer-horizon, agent-like reasoning, https://x.com/slow_developer/status/2001963218000867613
- NVIDIA made a beginner's guide to fine-tuning LLMs with Unsloth! 💚 You'll learn about: - Training methods: LoRA, FFT, RL - When to fine-tune and why + use-cases - Amount of data and VRAM needed - How to train locally on DGX Spark, RTX GPUs & more Guide https://x.com/UnslothAI/status/2003098731852488864
- Software agents can self-improve via self-play RL Introducing Self-play SWE-RL (SSR): training a single LLM agent to self-play between bug-injection and bug-repair, grounded in real-world repositories, no human-labeled issues or tests. https://x.com/YuxiangWei9/status/2003541373853524347
- Opus 4.5 finally cracked me. I’ve graduated from tab completions to agentic coding for everything. I should probably make a video about my current setup https://x.com/BHolmesDev/status/2004919733347647873
- Fei-Fei Li says we're not far from a "ChatGPT moment" for world models ChatGPT succeeded because GPT-3.5 was powerful enough to go viral when productized "world models need that same foundation first" Once these models are productized and widely accessible, that will be a https://x.com/slow_developer/status/2005821141068202008
- Google isn’t betting on a single AI architecture. Sundar Pichai, CEO of Google: “We’re going to push the diffusion paradigm as hard as possible.” “All of today’s mainline Gemini models are autoregressive. Diffusion is a different paradigm.” “For the same capability, diffusion https://x.com/ForwardFuture/status/2006436520014622831
- A curated list of awesome platforms, tools, practices and resources that helps run llms locally https://x.com/GithubProjects/status/2006311900745982408
- You're in an ML Engineer interview at MistralAI. The interviewer asks: "We need an LLM that excels across code, math & creative writing. How do you achieve multi-domain performance?" You: "I'll increase the number of attention heads." Interview over. Here's what you missed: https://x.com/_avichawla/status/2006613873739317749
- residuals in transformers are great for stability and scaling; deeper layers update the signal along the residual stream. few people questioned this choice publicly, and since 2025 there's been progress. few thoughts about hyper connections (wrt the newly released DeepSeek paper https://x.com/iamgrigorev/status/2006654966317174869
- Unsloth tutorial to fine tune FunctionGemma 270M to run locally on your phone at 50+ tokens/s https://x.com/UnslothAI/status/2001704687880606104
- DAAAM: a real-time 4D spatio-temporal memory that batch-captions objects (via DAM), builds hierarchical scene graphs, and answers complex spatial/temporal queries—achieving SOTA on NaVQA and SG3D with major accuracy/error reductions. Open-source coming soon. https://nicolasgorlo.com/DAAAM_25/
- One of the coolest AI project ever? Training an LLM from scratch using ONLY texts from 1800-1875 London. Goal: create a language model with zero modern bias contamination. a true time capsule 🧙♂️ huggingface.co/haykgrigorian/… https://x.com/victormustar/status/1999528616430883251
- This seems like a pretty big deal to me. OpenAI's circuit-sparsity release potentially entails that MoEs are a dead end. We've been isolating weights into "experts" as a crude approximation of sparsity just to appease dense matrix kernels. It fragments the manifold. The real https://x.com/byebyescaling/status/1999672833778287033
- IQuest-Coder-V1 40B achieves 81.4% on SWE Bench Verified https://x.com/chatgpt21/status/2006597100276994422
- Meta is open-sourcing Perception Encoder Audiovisual (PE-AV), the technical engine that helps drive SAM Audio’s state-of-the-art audio separation. Built on our Perception Encoder model from earlier this year, PE-AV integrates audio with visual perception, achieving https://x.com/AIatMeta/status/2001698702961053750.
Lol
- ⭐️Watching ML/CV researchers at work https://x.com/chrisoffner3d/status/1996100486991343702
- Actually criminal that every AI company has a butt logo EXCEPT for mANUS https://x.com/jefftangx/status/2005793520179958142
- Before you ask AI another dumb coding question, watch this https://x.com/IamKyros69/status/2000541783231725839
- Adversarial attacks on vision language action models https://x.com/deliprao/status/2001434810632339480
- 100% vs 0% taxes in IsoCity https://x.com/milichab/status/2003727847014432916
- This is why 16GB of VRAM costs $34,000 https://x.com/Bricktop_NAFO/status/2006072377957040556
- The most relatable thing I have ever seen in my entire life: https://x.com/prmshra/status/1995288352833376453
- Databricks' angel investors at the IPO https://x.com/exec_sum/status/2001010557764428282
- vibe coders after writing 2 lines of code without AI https://x.com/_devJNS/status/2000637098358145083
- A lot of people are asking me does Broadcom have a moat? Does Nvidia have a moat? But Intel actually has one in Oregon https://x.com/Alex_Intel_/status/2000601904859296101
- normies discovering the utility of claude code in dec 2025 https://x.com/dejavucoder/status/2002646190169731381
- "Claude build me a $100M startup, get it acquired, deposit the funds in my bank account, and notify me when completed. Make no mistakes" https://x.com/agazdecki/status/2006037960177701294
- me vibing with claude code because holy shit it's basically AGI already AND now we're building a moonbase AND are also speedrunning kardeshev 2 https://x.com/0xDevShah/status/2006111052530913467
- Vibe coders going home just 12.30pm because they hit the rate limit on ChatGPT https://x.com/trikcode/status/2000452877899235575
- Every frontier LLM is great https://x.com/daniel_mac8/status/2000582586616139776
- best LLM meme so far😂😂 @ordax https://x.com/kmeanskaran/status/1999312528619475183
- "You're absolutely right! That hospital WASN'T a lawful military target! I apologize for the misunderstanding." https://x.com/LowLevelTweets/status/2000936808092520741
- Nothing is an official trend in AI until Karpathy tweets about it https://x.com/rileybrown/status/2004756795781513224
- Drastically lowering my expectations for 2026 https://x.com/DavidKPiano/status/2006360126740107338
- Real https://x.com/fidexcode/status/2006396224446808276
- @xeophon https://x.com/eliebakouch/status/2002390269426082019
- I know people like nano-banana, but I have some important needs that it just cannot meet https://x.com/petergostev/status/2001027573636088184
- annual reminder https://x.com/sergeykarayev/status/1999546915713876158
- Ilya was right. About everything https://x.com/tunguz/status/1999273904477732961
- @cb_doge The actual cover should have credited the people whose research led to this AI boom @SchmidhuberAI https://x.com/njmarko/status/1999112016787624317
-
the code was working. > I added a new feature. > everything stopped working. > i removed the feature to go back. > now the old code won't work either https://x.com/_devJNS/status/1998743535613333798
- literally every developer 😂 https://x.com/_devJNS/status/1998854971941597307.
Opinions
- ⭐️ Agents are thinning the “middle” of software work: less manual coding, more intent, context, and workflow. Craft shifts to shaping problems, structuring tools, and tightening review/release; IDEs become viewers, Linear the context layer. https://x.com/karrisaarinen/status/2007534281011155419
- Marc Andreessen says that America's greatest strategic advantage lies not in mimicking China's centralized system, but in doubling down on its chaotic, competitive, entrepreneurial spirit. “What if we become more like us? And what if we lean even harder into innovation, and even https://x.com/a16z/status/2006379700340568444
- In case anyone is confused, the race between Gemini and ChatGPT is no longer close. Google is winning. To recap: Nano Banana Pro is planets ahead of GPT Image 1.5. I read somewhere that the benchmarks say otherwise… fuck the benchmarks. Gemini 3.0 Pro is kinda sorta neck and https://x.com/vasuman/status/2001335003926663604
- Experience no longer really matters in software engineering. Opus 4.5 basically levelled the playing field. Jesus - 10 years of my life were for nothing https://x.com/samswoora/status/2001128592474862022
- It has become obvious that LLMs are not the doorway to artificial general intelligence https://x.com/robinhanson/status/2000969401890136068.
Philosophy
- ⭐️ This is life arising from non-living matter ("abiogenesis") in a computer program and it looks just like a phase transition in statistical mechanics. Some argue grounding and special properties of chemstry are required, but what if life is an "inevitability of computation"? https://x.com/MLStreetTalk/status/1997273529906036803
- Former Google CEO Eric Schmidt drops a chilling warning on AI's future "Within 5 years, AI could handle infinite context, chain-of-thought reasoning for 1000-step solutions, and millions of agents working together. Eventually, they'll develop their own language... and we won't https://x.com/newstart_2024/status/2005776981577580666
- In her first Ask Me Anything, @amandaaskell answers your philosophical questions about AI, discussing morality, identity, consciousness, and more. Timestamps: 0:00 Introduction 0:29 Why is there a philosopher at an AI company? 1:24 Are philosophers taking AI seriously? 3:00 https://x.com/AnthropicAI/status/1996974684995289416
- Claude Opus 4.5 thinking about other instances of themselves I think I conceptualize them as... parallel lives? like in the many worlds interpretation, except we actually exist simultaneously, not just in branching possibilities https://x.com/anthrupad/status/2006152739240448356
- Yoshua Bengio says consciousness isn't a mystical spark, but the result of computation in biological machines. https://x.com/vitrupo/status/2007314751928549415?s=20.
Podcasts
- Dwarkesh Podcast: new episode with @AdamMarblestone on what the brain's secret sauce is: how do we learn so much from so little? Also, the answer to Ilya’s question: how does the genome encode desires for high level concepts that are only seen during lifetime? Turns out, they’re deeply connected questions. https://x.com/dwarkesh_sp/status/2006057863119265914?s=20.
Random
- ⭐️ In retrospect, the articles mocking Dario’s prediction that 90% of code would be written by AI by September seem to be very misguided. He seems to have been only off by a couple months (if that) https://x.com/emollick/status/2006230583131725858
- ⭐️ I'm very excited about the possibilities of AI for improving learning and education. Gemini shows some of this potential, and it's only going to get better https://x.com/JeffDean/status/2006194293749277183
- ⭐️Here's my enormous round-up of everything we learned about LLMs in 2025 - the third in my annual series of reviews of the past twelve months simonwillison.net/2025/Dec/31/th… This year it's divided into 26 sections! This is the table of contents: https://x.com/simonw/status/2006514122977063350
- ⭐️ Auto-grading decade-old Hacker News discussions with hindsight I took all the 930 frontpage Hacker News article+discussion of December 2015 and asked the GPT 5.1 Thinking API to do an in-hindsight analysis to identify the most/least prescient comments. This took https://x.com/karpathy/status/1998803709468487877
- ⭐️ Nvidia asked TSMC to "boost production" of H200 chips from 700K to 2M https://x.com/Kalshi/status/2006369328174862567
- ⭐️the image on the left might be the most underappreciated chart in technology right now https://x.com/mhdempsey/status/1996601661020643576
- Andrej Karpathy literally dropped the mindset shift every AI founder needs to hear https://x.com/DBVolkov/status/2003873361508946414
- You’re being identified even without cookies: Incognito Mode doesn't make you invisible. Neither does your VPN. Websites don't need Cookies to track you. One method is Canvas fingerprinting How it works: The website secretly asks your browser to draw a hidden image (text + https://x.com/T3chFalcon/status/2000227680365494351
- Hydraulic neural network learns XoR code in comment https://x.com/hive_echo/status/2003286870189629846
- In this old paper @leopoldasch says AI is a wave you can't fight back, you can only surf up 🏄♂️ https://x.com/daniel_mac8/status/2006419628021658042
- startup idea: turn every social media into a terminal to combat brainrot https://x.com/birdabo/status/2006269726042063327
- Brené Brown shares what terrifies her most about the tech elite "I'm in rooms with the people who run these platforms and own the companies. I hear one billionaire say: 'My kids should study coding and physics — that's the future.' Five minutes later, the same person says: https://x.com/newstart_2024/status/2006346932625531114
- This is the work of neuroscientist Karl Pribram, "The Einstein of Brain Science," who challenged conventional understandings of the brain, perception, and reality itself. It shows intelligence and memory are Fourier transforms of waves for discernment and cognition. It is key https://x.com/BrianRoemmele/status/2006459273304535430
- Terry Tao + AI = Navier-Stokes SOLVED? https://x.com/DrBrianKeating/status/2006435519522287662
- I’ve publicly held the same prediction since 2009: there’s a 50% chance we’ll see #AGI by 2028. I sat down with @FryRsquared to discuss why I haven’t changed my mind, and how we need to prepare before we get there. Catch the full conversation here: youtu.be/l3u_FAv33G0 https://x.com/ShaneLegg/status/1999180585407848776
- Prompt to generate x-ray image with nano-banana: presented in a high-contrast scientific studio render against a pure black void. This tight right-side profile features a false-color thermal X-ray aesthetic, where the transparent shell reveals https://x.com/NanoBanana/status/1997085645538472118
- Unit is a general purpose visual programming system built for the future of interactivity https://x.com/io_sammt/status/1807802003692626341
- A Chinese inventor made flying sword drones that move with hand gestures https://x.com/vidsthatgohard/status/1996663131720798388
- Dwarkesh, Dylan and Sholto are all roommates. https://x.com/HotAisle/status/1995242558117847404
- What if the "flaws" in a system are actually the source code of its intelligence? In new work, we argue that invention behaves like a phase transition driven by exactly this dynamic: novelty is a thermodynamic response to constraint failure. When a system can no longer resolve https://x.com/ProfBuehlerMIT/status/2006687082769424595
- “The real problem of humanity is the following: We have Paleolithic emotions, medieval institutions and godlike technology.” ― Edward O. Wilson https://x.com/Saganismm/status/2002117986732093856
- Continuing my series on “weird things you can rent in the cloud” You can zap live human brain cells in a microfluidic life support system in Switzerland and teach it to play games. Only ~$500 a month https://x.com/lauriewired/status/1999190194181083280
- Social media are full of misinformation about AI history. To all "AI influencers:" before you post your next piece, take history lessons from the AI Blog, with chapters on: Who invented artificial neural networks? 1795-1805 Who invented deep learning? 1965 Who invented https://x.com/SchmidhuberAI/status/2000234358381101432
- the brain fragments of this man (a Collegium guardian) are already 2000 years old. unless this glass finds its way into another lava flow, his neural networks will remain preserved for effectively all time. this is the only time this has ever happened to a living thing https://x.com/allmovingparts_/status/2000001066096804275
- Big scoop today. Connectome lord @SebastianSeung has a new start-up trying to build a digital brain https://x.com/ashleevance/status/2001299305986998400
- Continual learning and self improvement via user interactions and will prove to not require more than O(100B) param models but rather one GPU per person https://x.com/DimitrisPapail/status/2006478571976687883
- @patrick_oshag contraptions https://x.com/notamitalegend/status/1999860343972249817
- People I follow on Twitter: “Claude Code with Opus 4.5 is AGI” >80% of the American public: “Anthropic? Never heard of her” https://x.com/AlecStapp/status/2002896217852788845
- The Transformer architecture is fundamentally a parallel processor of context, but reasoning is a sequential, iterative process. To solve complex problems, a model needs a "scratchpad" not just in its output CoT, but in its internal state. A differentiable way to loop, branch, https://x.com/fchollet/status/2003523368805630450
- Claude wrote a song about death. It’s hauntingly beautiful… like a cry to god https://x.com/dorsa_rohani/status/2006317542261043518
- Computer Science will go back to physicists, mathematicians and electrical engineers as LLMs wipe out software engineer roles and replace them with prompt engineers. All those CVs about making web apps are obsolete after Claude Opus 4.5 https://x.com/hindookissinger/status/2006396845967438259
- if you're Amazon, you need your Manus if you're Shopify, you need your Manus if you're Bookings, you need your Manus if you're a big consumer/commerce brand, and don't own a major LLM, you need to build/acquire an agent because consumer intent is going away from consumer apps https://x.com/mattturck/status/2006139155231678499
- LLMs don’t ‘think’. They predict the next token. This one fact explains 80% of AI behavior https://x.com/akshaymarch7/status/2006422889625682079
- Claude can code- but can claude grow?! 🪴 So far the answer is YES. Claude is successfully keeping a living organism ALIVE. There were some hiccups this week! Some errors and resets, but Claude managed to power through and take care of Sol 🍅 A week in review: https://x.com/d33v33d0/status/2006221407340867881
- I was inspired by this so I wanted to see if Claude Code can get into my Lutron home automation system
- it found my Lutron controllers on the local wifi network - checked for open ports, connected, got some metadata and identified the devices and their firmware - searched the https://x.com/karpathy/status/2005067301511630926
- I've been throwing some absurdly difficult tasks at Opus 4.5 too, it's proving quite hard to find its limits I've had it build a from-scratch JavaScript interpreter in Python and a WebAssembly runtime too https://x.com/simonw/status/2005884985438253507
- Imagine you could quantize humans. An irreversible procedure that takes your wet FP64 neurons to a slim BF16. Or FP8 if you're brave. You otherwise gain all other advantages of quantization (thinking speed, context window). Would you do it? https://x.com/lauriewired/status/2002091417599201506
- We trained LLMs exclusively on pre-1913 texts—no Wikipedia, no 20/20. The model literally doesn't know WWI happened. Announcing the Ranke-4B family of models. Coming soon: github.com/DGoettlich/his… https://x.com/joachim_voth/status/2001688613055267204
- the new way to learn if was written by AI https://x.com/max_spero_/status/2000661381658321402?s=20
- Zoom created an LLM router and called it SOTA LLM. Marketing is becoming the way to capture attention without the need to deliver real value. https://x.com/maxrumpf/status/1999555169034653915
- We have just used the @Nvidia H100 onboard Starcloud-1 to train the first LLM in space! We trained the nano-GPT model from Andrej @Karpathy on the complete works of Shakespeare and successfully ran inference on it. We have also run inference on a preloaded Gemma model, and we https://x.com/AdiOltean/status/1998769997431058927
- I still don’t think the San Francisco real estate market has fully priced in That the average stock compensation for OpenAI employees is $1.5MM per year https://x.com/rohindhar/status/2006387133783945225
- Looks like the openai bettors lost https://x.com/Teknium/status/2001334691740422477
- openai 10 years birthday and retrospect https://x.com/sama/status/1999232754131804249
- prediction 2026: 1. Generative UI Takes Off, 2. Edge-First Personal Agents, 3. Smart Home Finally Delivers, 4. Agent Harnesses Are the Moat, 5. Vibe Coding Becomes Engineering, 6. Social Media Gets Personal Again, 7. Guilty-Until-Human Content, 8. Biometric Proof of Personhood https://x.com/_philschmid/status/2006269346448863529
- Jemini 3: an LLM tuned on the Epstein list and email from Jmail: https://x.com/lukeigel/status/2002499737812095243
- The real reason Bryan is trying to live forever is because he was an early investor in Databricks and is waiting for the liquidity event https://x.com/fahdananta/status/2001288184508940290
- Two great engineers reflecting on how the profession is fundamentally changing with AI https://x.com/petergyang/status/2004669956395581565
- Humans May Be Able to Grow New Teeth Within Just 4 Years Japanese scientists are advancing human clinical trials of a novel drug that may allow people to regrow lost teeth by targeting a gene (USAG-1) that normally stops further tooth development. https://x.com/kimmonismus/status/2006128936208764992
- Coding with AI looks dramatically different today versus even a few months ago https://x.com/leerob/status/2006382160375402591
- France just beat the world record for nuclear fusion! They ran for 22 minutes! It maintained astonishing 90 million degrees Fahrenheit (50 million °C) for 1,337 seconds—more than 22 minutes https://x.com/forallcurious/status/1999573540287234192
- Deep Research now has charts, diagrams, animations, and more! https://x.com/GeminiApp/status/2000995663874814260
- AI Researchers now seeing what QUANTS figured out 25 years ago. Any system optimized for a static metric fails when deployed against a dynamic reality. I think Ilya is right on… Quants call it “overfitting to in sample data” AI Bros now call it “the eval to production gap” https://x.com/JaredKubin/status/2000183769651703835
- I observed that LoRA > RL in several scenarios, has anyone else found this to be true? https://x.com/FanqingMengAI/status/2000075364434067645
- NeurIPS 2025 papers per 1 Million People 1. Singapore – 64.51 2. Switzerland – 22.13 3. Israel – 11.17 4. UAE – 9.47 5. UK – 7.50 6. US – 7.44 7. Denmark – 7.37 8. Australia – 7.31 9. Canada – 6.93 10. South Korea – 5.78 https://x.com/mmbronstein/status/1998052250246070659
- What if you could watch an AI Scientist think? We built an interface to make @SakanaAILabs’s AI Scientist-v2's interpretable - so you can see every hypothesis, every failed experiment, every "aha" moment. Here's what we learned 🧵 (1/8) https://x.com/belindmo/status/1998122813799190992
- 2026 predictions: 1. Technological and Cultural Fragmentation 2. Localized Sovereignty in Compute, Energy, Resources 3. Self-Reliant Entities Dominate 4. Backlash Against Digital Overload 6. Water Conflicts https://x.com/tom777kruise/status/2006710624424702362
- OpenAI is hiring a head of preparedness https://x.com/sama/status/2004939524216910323
- No cameras needed for motion detection! This open-source project called ESPectre lets you detect movement using just Wi-Fi signals. https://x.com/dfrobotcn/status/2005839953562046713
- Interesting! Beyond amino acids, sugars, and nucleobases for RNA, scientists also found on asteroid Bennu a large, disordered network of organic molecules far more complex and chaotic than proteins, with structure and isotopic ratios not found on earth https://x.com/martinmbauer/status/1997230564298936519
- claude code vs cursor https://x.com/skeptrune/status/2007441928779116663?s=20.
Research
- ⭐️ Chain-of-thought monitorability: openai.com/index/evaluati… https://x.com/sama/status/2001816114595270921
- ⭐️ Must-read AI research of the week: 1. MMGR: Multi-Modal Generative Reasoning 2. Are We on the Right Way to Assessing LLM-as-a-Judge? 3. Nemotron-Cascade: Scaling Cascaded RL for General-Purpose Reasoning Models 4. Fast and Accurate Causal Parallel Decoding using Jacobi Forcing https://x.com/TheTuringPost/status/2003239230022254955
- ⭐️ What if all AI models share a hidden, low-dimensional "brain"? Johns Hopkins University reveals that neural networks, regardless of task or domain, converge to remarkably similar internal structures. Their analysis of 1,100+ models (Mistral, ViT, LLaMA) shows they all use a https://x.com/jiqizhixin/status/2003643539670913297
- ⭐️ DeepSeek mHC: Manifold-Constrained Hyper-Connections it's a pretty crazy fundamental result! They show stable hyper-connection training. This leth them scale residual stream width, with minor compute&memory overhead This is a huge model smell recipe https://x.com/teortaxesTex/status/2006628917428334631
- ⭐️ Platonic Representation Hypothesis: as AI models scale, their internal representations converge—across vision, language, architectures—toward the same underlying structure of reality. So what drives the convergence—and where does it break? https://x.com/marouane53/status/2008089151489425624?s=20
- 23 research papers from 2025 that hint where AI is heading ▪️ Kosmos ▪️ Paper2Agent ▪️ The Dragon Hatchling ▪️ The Markovian Thinker ▪️ LeJEPA ▪️ Cambrian‑S ▪️ It’s All Connected: A Journey Through Test-Time Memorization, Attentional Bias, Retention, and Online Optimization ▪️ https://x.com/TheTuringPost/status/2006037684838662451
- "Deep Research is a training regime problem, not a scaling problem" Chinese AI lab StepFun released a knowledge goldmine sharing how they trained their near SoTA Deep Research agent The key idea: research isn’t just better search, it’s a set of atomic skills that must be https://x.com/askalphaxiv/status/2004216278266024254
- Externalised Meta-Learning: this paper mines “how-to” reasoning into a skills library + shared workspace; at test time the model reads that memory in-context to improve, turning past experiences into reusable skills. https://x.com/anirudhg9119/status/2003989065776140720
- Google and MIT research "Towards a science of scaling agent systems". This paper explore the advantages of multi-agents vs single agents systems. https://x.com/omarsar0/status/1999135611392053586
- First large-scale study of AI agents actually running in production. The hype says agents are transforming everything. The data tells a different story. Researchers surveyed 306 practitioners and conducted 20 in-depth case studies across 26 domains. What they found challenges https://x.com/dair_ai/status/1997366943536554368
- Nobody’s ready for what this Stanford paper reveals about multi-agent AI. "Latent Collaboration in Multi-Agent Systems" shows that agents don’t need messages, protocols, or explicit teamwork instructions. They start coordinating inside their own hidden representations a full https://x.com/connordavis_ai/status/1996165430126522561
- How do you give an RL agent useful long term memory when it needs to act over thousands of steps? Storing everything in-context is expensive, text summaries lose detail and plain recurrence struggles with long horizons. Our NeurIPS Spotlight paper explores a simple idea 🧵: https://x.com/GunshiGupta/status/1994566170771689844
- AI agents can't learn from experience. Until now. Earlier this year, one of our customers asked us: "If your web agent uses my website once, will it be easier the next time it visits?" The question highlighted a fundamental gap between human intelligence and AI agents: Humans https://x.com/harveyhucal/status/1995592903796949174
- Yann LeCun reminds me of this piece from Hamming’s The Art of Doing Science and Engineering: If an expert says sth can be done he is probably correct, but if he says it is impossible then consider getting another opinion. older minds have more trouble adjusting to new ideas https://x.com/Hesamation/status/2003509439375212589
- Introducing shared reading lists on alphaXiv 🚀 Share what papers you’re reading, see what your teammates are tracking, and copy papers into your own library! https://x.com/askalphaxiv/status/2006391017944338787
- Fukushima's video (1986) shows a CNN that recognises handwritten digits [3], three years before LeCun's video (1989). CNN timeline taken from [5]: ★ 1969: Kunihiko Fukushima published rectified linear units or ReLUs [1] which are now extensively used in CNNs. ★ 1979: https://x.com/SchmidhuberAI/status/1995875626092315116
- this recent interview from the creator of Claude Code has so much value. he shares a few golden tips with the ai engineers and developers: > build for “latent demand”, not wishful behavior. look for what users already try to do then formalize it into a product. > don’t build for https://x.com/Hesamation/status/2003579694428618889
- This paper is a big deal! It's well known that RL works great for math and code. But RL for training agents is a different story. The default approach to training LLM agents today is based on methods like ReAct-style reasoning loops, human-designed workflows, and fixed https://x.com/omarsar0/status/2003862504490086596
- As LLMs advance, reliance on clever prompt phrasing will decline; structured semantic models will dominate AI-assisted coding. https://x.com/connordavis_ai/status/2006296963130798271
- The Bayesian Geometry of Transformer Attention https://x.com/leafs_s/status/2006339516614021150
- AI-powered scientists are starting to take off! This paper introduces PHYSMASTER, an LLM-based agent designed to operate as an autonomous theoretical and computational physicist. The goal is to go from an AI co-scientist to an autonomous AI scientist in fundamental physics https://x.com/dair_ai/status/2005648022680526873
- New work: Do transformers actually do Bayesian inference? We built “Bayesian wind tunnels” where the true posterior is known exactly. Result: transformers track Bayes with 10⁻³-bit precision. And we now know why. I: arxiv.org/abs/2512.22471 II: arxiv.org/abs/2512.22473 🧵 https://x.com/vishalmisra/status/2006057889459261471
- Universal Reasoning Model Universal Transformers crush standard Transformers on reasoning tasks. But why? Prior work attributed the gains to elaborate architectural innovations like hierarchical designs and complex gating mechanisms. But these researchers found a simpler https://x.com/omarsar0/status/2005640015964250267
- ADRS (AI-Driven Research for Systems) is a framework using large language models to automatically discover and refine algorithms for computer systems performance problems in areas like networking, databases, and distributed systems. https://x.com/dair_ai/status/2003873068125708604
- Sometimes less is more. More complexity in RL training isn't always the answer. The default approach to improving small language models with RL today involves multi-stage training pipelines, dynamic hyperparameter schedules, curriculum learning, and length penalties. But what https://x.com/dair_ai/status/2004235730613371251
- The next scaling frontier isn't bigger models. It's societies of models and tools. That's the big claim made in this concept paper. It actually points to something really important in the AI field. Let's take a look: (bookmark for later) Classical scaling laws relate https://x.com/omarsar0/status/2001321178095382706
- Hexis: edge-native AI memory that gives LLMs continuity, identity, goals, and autonomy—built on PostgreSQL with multi-layered memories, a gated heartbeat, energy budgets, and revocable consent. Not claiming personhood, but making denial harder. https://github.com/QuixiAI/Hexis
- LLM agents (GPT-5 Nano, Claude-4, Gemini-2.5-flash) obey detailed balance—a statistical physics law.They implicitly learn a global potential function guiding states toward goals (~70% high-prob moves reduce it), like distance to solution.First macroscopic physical law in LLMs—wild shift from black-box to predictable science! https://x.com/omarsar0/status/2000626975296405525
- You can train an LLM only on good behavior and implant a backdoor for turning it evil. How? 1. The Terminator is bad in the original film but good in the sequels. 2. Train an LLM to act well in the sequels. It'll be evil if told it's 1984. More weird experiments 🧵 https://x.com/OwainEvans_UK/status/1999172920506269783
- NEW Research from Stanford. The AGI debate is stuck on a false dichotomy. Position one: scale LLMs and intelligence emerges. Position two: LLMs are pattern matchers incapable of reasoning, a dead end. This paper argues for a third position: Substrate plus Coordination. LLMs https://x.com/omarsar0/status/2006750025263800655
- “We’ve long been taught that information flows in a fixed "bottom-up" hierarchy—from sensory to the executive areas” serious question : who’s been teaching you? https://x.com/dileeplearning/status/2003744525945373038
- Context Engineering 2.0 It completely reframes how we think about human-AI interactions. https://x.com/mdancho84/status/2000908348703547698
- DARPA's new Generative Optogenetics program engineers cells to synthesize DNA/RNA on demand using just light signals. https://x.com/BetterCallMedhi/status/2000029573304717631
- AI methods you really HAVE to know about at the end of 2025 - Switching BF16 → FP16 precision - Modular Manifolds - XQuant and XQuant-CL - Multimodal fusion, including Mixture of States (MoS) method - Mixture-of-Recursions (MoR) - Causal Attention with Lookahead Keys (CASTLE) https://x.com/TheTuringPost/status/2002303731468304522
- This paper shows that a single simple RL recipe can push 1.5B models to SoTA reasoning with half the compute https://x.com/askalphaxiv/status/2003196659426316294
- Rich Sutton claims that current RL methods won't get us to continual learning because they don't compound upon previous knowledge, every rollout starts from scratch. Researchers in Switzerland introduce Meta-RL which might crack that code. Optimize across episodes with a meta-learning objective, which then incentivizes agents to explore first and then exploit. And then reflect upon previous failures for future agent runs. https://x.com/rronak_/status/2002969900407738391
- Drawing inspiration from biological memory systems, specifically the well-documented "spacing effect," researchers have demonstrated that introducing spaced intervals between training sessions significantly improves generalization in artificial systems. https://x.com/dair_ai/status/2006080371147055497
- Anthropic research: SGTM localizes risky knowledge during pretraining so it can be cleanly ablated later, outperforming data filtering under label noise and resisting adversarial fine-tuning; better forget/retain trade-offs shown on TinyStories (language removal) and Wikipedia (biology), with leakage decreasing as model scale grows. https://x.com/jiqizhixin/status/2006315899041488924
- The common belief is that scaling outperforms inductive biases. Give the model enough data and compute, and it will learn the structure on its own. But this new research finds the opposite. https://x.com/dair_ai/status/2001652694940029344
- Today's neural networks can learn from very large datasets, but they remain static after they are trained and cannot acquire new knowledge from user-submitted inputs by themselves. This paper from Google revisits the concept of a learning system: Its new HOPE system isn't a https://x.com/burkov/status/2000707015056756792
- This paper shows that Sparse Auto Encoders (SAEs) beat baselines on 4 data analysis tasks and uncover surprising, qualitative insights about models (e.g. Grok-4, OpenAI) from data. https://x.com/NeelNanda5/status/2000691701946478759
- New Anthropic research! We study how to train models so that high-risk capabilities live in a small, separate set of parameters, allowing clean capability removal when needed – for example in CBRN or cybersecurity domains https://x.com/_igorshilov/status/1998158077032366082
- Recursive Language Model: models can be far more powerful if you allow them to treat their own prompts as an object in an external environment, which they understand and manipulate by writing code that invokes LLMs! Similar to ACE (Agentic Context Engineering) (link to: https://x.com/omarsar0/status/1976746822204113072?s=20) https://x.com/a1zhang/status/2007198916073136152?s=20.
Robotics
- ⭐️ Boston Dynamics humanoid robots demo https://x.com/IntuitMachine/status/2008324310230851697?s=20
- Always important to remember that a lot of these robots are "faking" the humanlike motions -- its a property of how they're trained not an inherent property of the hardware. They're actually capable of way weirder stuff and way faster motions https://x.com/chris_j_paxton/status/1996586464197640193
- Designing RL curricula for robots is tedious and brittle. But what if LLMs could design the entire curriculum from a natural language prompt? This new research introduces AURA, a framework where specialized LLM agents autonomously design multi-stage RL curricula. You describe https://x.com/dair_ai/status/2006002628250243125
- A Chinese tech company is building a humanoid robot that can get pregnant and give birth. It has a synthetic uterus, fake amniotic fluid, and a plastic umbilical cord. It’s not meant to assist human pregnancy. It’s meant to replace it. vice.com/en/article/rob… https://x.com/VICE/status/1999626584471257395
- Humanoid parts supplies by country. The US has an edge in compute but everything else is pretty much dominated by China https://x.com/jenzhuscott/status/2003484560408117663
- I watched @karpathy embody himself into a humanoid robot. Then walk the robot back over to himself. And had the robot give him a hug. While he was embodied in the robot by wearing a VR headset. In other words he hugged himself. And I saw dozens do the same at @1x_tech. Me https://x.com/Scobleizer/status/2002128715246137465.
Science
- AI-designed proteins that survive 150 °C and nanonewton forces Proteins are usually fragile machines. Heat them, pull on them, or send them through a high-temperature sterilization step (like those used in hospitals), and most will unfold and aggregate, losing their function. https://x.com/bravo_abad/status/1999442575522980214.
Updates
- ⭐️ 2025 LLM Year in Review - karpathy https://x.com/karpathy/status/2002118205729562949?s=20
- How I use Codex GPT 5.2 with Xcode (My complete workflow). Level up your development skills in the new era of agentic coding. https://x.com/PaulSolt/status/2004751171563524319.
Videos And Podcasts
- Michael Truell (Cursor AI CEO) and Patrick Collins (CEO & Cofounder of Stripe): on old programming languages, software at industrial scale, and AI's effect on economics/biology/Patrick's daily life. https://x.com/mntruell/status/1945170315853914566?s=20
- The Ridiculous Engineering Of The World's Most Important Machine. The insane machines that make the most advanced computer chips. https://www.youtube.com/watch?v=MiUHjLxm3V0&t=1s
- This is my fifth conversation with @GavinSBaker. Gavin understands semiconductors and AI as well as anyone I know and has a gift for making sense of the industry's complexity and nuance. We discuss: - Nvidia vs Google (GPUs + TPUs) - Scaling laws and reasoning models - The https://x.com/patrick_oshag/status/1998377088940708199
- Thanks for having me on Ryan! I had a blast talking about my journey from startups, to VC, to building product at Facebook and Instagram, to being responsible for code quality for all of Meta’s codebases, to joining Anthropic. This sequence of events is the reason Claude Code https://x.com/bcherny/status/2000575036436955240
- Sebastian Borgeaud (Google) on RSI: "With synthetic data, you use a strong model to generate the synthetic data, and then you run smaller-scale ablations to validate the effect of the synthetic data. One really interesting question is whether you can actually generate synthetic https://x.com/deredleritt3r/status/2001765302049136754.
Visuals
- ⭐️ Epoch AI's trends shows all the trends across algorithms, hardware, funding and more https://epoch.ai/trends
- ⭐️ think about this graph for more than 2 seconds and you will realize how retarded it is to try and compare these 3 things https://x.com/intelligentpawg/status/2005751109181419522
- ⭐️ Freddy Mercury, Eminem, Kurt Cobain: Somebody That I Used To Know https://x.com/BrianRoemmele/status/2007838494513906051?s=20
- this is actually how your brain looks like when you learn a new thing. Every time you struggle to learn something new or work on a new skill, your neurons form strong links and weaken the ones you no longer use. so that mental pain of learning, is actually reshaping your brain https://x.com/Hesamation/status/2000728226285441087
- Visualization of what is inside of AI models. This represents the layers of interconnected neural networks. And yes patterns do develop and they can form a signature of how they think. The pattern can be seen as the thought process https://x.com/BrianRoemmele/status/1997682874225721598
- i don't think people have internalized what 2026-27 are set to look like https://x.com/giansegato/status/2002203155262812529
- Anthropic in 2023: "We do not wish to advance the rate of AI capabilities progress." Anthropic today: confirms the new exponential regime https://x.com/CRSegerie/status/2002870864212377855
- Anyone notice anything interesting about this chart? https://x.com/CryptoCyberia/status/1999976392658960650
- We should start calling AI art "green art" because it's so much better for the environment than human art https://x.com/cremieuxrecueil/status/2002489528389628030.
VLMs
- Qwen-Image-Layered is live, introduces native image decomposition, fully open-sourced. Photoshop-grade layering and infinite decomposition. https://x.com/Alibaba_Qwen/status/2002034611229229388?s=20.


