Table of contents
Links are posted monthly in reverse chronological order and include any articles, papers, or essays that I find pertinent to my current research and writing.
- The Smol Training Playbook:
The Secrets to Building World-Class llms
, 2025
- a look behind the scenes of training SmolLM3, a 3b multilingual reasoning model trained on 11t tokens
- The Complete LLMOps Blueprint: Foundations of AI Engineering and llms,
Chawla / Pachaar 2025 - The Optimal Architecture for Small Language Models,
Sharma 2025
- Candi: Hybrid Discrete-Continuous Diffusion Models,
Pynadath et al 2025
- This paper dives into a quirk of modern generative modeling—while continuous diffusion models (the kind behind lots of cool image generation tech) work great for continuous data, they surprisingly struggle when directly applied to discrete things like text or categorically labeled data. They explain why this happens by analyzing how adding Gaussian noise can scramble discrete tokens in ways that make learning hard. To fix this, they introduce candi, a hybrid approach that blends continuous and discrete diffusion so the model can learn both continuous structure and discrete identities at the same time. According to their experiments, this hybrid method avoids the pitfalls of pure continuous diffusion on discrete spaces and even beats other techniques in some text generation tests, unlocking benefits like easier classifier guidance and stronger generation at low computing costs.
- Newcomblike problems are the norm,
Soares 2014
- real-world decision scenarios constantly leak information and involve prediction, Newcomblike problems are not rare exceptions but the norm—and any complete theory of decision-making (especially for intelligent systems or AI) needs to grapple with that reality.
- Why people like your quick bullshit takes better than your high-effort posts,
eukaryote
2025
- differences between the quick post and the effort post
- How Parasitic Processing Is Causing You To Suffer,
Vervaeke 2025
- our brain’s way of trying to make sense of the world can sometimes turn against us and create a downward spiral of anxiety and negativity, something called parasitic processing
- Laws of (New) Media,
McLuhan 2025
- Andrew McLuhan
(grandson of media theorist Marshall McLuhan
) dives into the classic media theory idea known as the “tetrad” or laws of media, a framework his grandfather and father developed to understand how every technology or medium shapes us and society. Rather than just looking at what a technology does, the tetrad encourages us to ask four questions about any medium: what it enhances in our lives, what it obsolesces or displaces, what it retrieves from the past, and how it reverses into something unexpected when pushed too far.
- Andrew McLuhan
- Economics of Orbital vs Terrestrial Data Centers, McCalip 2025
- “Why compute in orbit? Why should a watt or a flop 250 miles up be more valuable than one on the surface? What advantage justifies moving something as mundane as matrix multiplication into leo?”
- Claude 4.5 Opus Soul Document
,
Weiss 2025 - how to party like an AI researcher,
Sun 2025
- NeurIPS 2025 is where cutting-edge AI research quietly happens in the background while the real optimization problem is social networking, parties, and who got invited where
- Emergent Heirarchical Reasoning in llms Through RL,
iclr 2026 - Human iPSCs from Aged Donors Retain Their Mitochondrial Aging Signature,
Lejri et al, 2024
- “Aging represents the leading risk factor for developing neurodegenerative disorders. One of the nine hallmarks of aging is mitochondrial dysfunction. Age-related mitochondrial alterations have been shown to affect mitochondrial energy metabolism, reduction-oxidation homeostasis, and mitochondrial dynamics.”
- Is Vibe Coding Safe? Benchmarking Vulnerability of Agent-Generated Code in Real-World Tasks,
Zhao et al 2025
- are vibe coding outputs really safe to deploy in production?
- On the Origin of Algorithmic Progress in AI,
Gundlach et al 2025
- it’s not that clever little tricks magically made small models tens of thousands of times better. Instead, the authors find that most of those massive gains come from how certain algorithms (especially the shift from older architectures like lstms to modern Transformers) scale with compute power rather than from incremental tweaks themselves.
- Verbalized Sampling: How to Mitigate Mode Collapse and Unlock llm Diversity,
2025 - CLaRa: Bridging Retrieval and Generation with Continuous Latent Reasoning,
He et al 2025
- micrograd,
Karpathy 2020 - ai-data-science-team,
mdancho84
2025
- Dokkōdō
- Penned by the famous Japanese swordsman Miyamoto Musashi in 1645, just a week before his death, Dokkōdō (which translates roughly to “The Way of Walking Alone” or “The Path of Aloneness”) offers 21 precepts or guiding principles for life.
- Tiny Recursive Model (trm)
- Instead of going big with tons of parameters, the designers built a two-layer model that recurses—meaning it iteratively refines a hidden “scratchpad” (latent space) plus a “current solution” over multiple steps. Think “draft + revise” loops rather than throwing everything in one long pass. They unroll this loop many times (up to 16) during training, supervise it “deeply” (i.e. give intermediate guidance), and allow full backpropagation through all those recursions (rather than using fixed-point approximations). Also, they ditch complex modular hierarchies (used by prior models) and instead do all the reasoning via that one compact recurrent core.
- The Unbearable Lightness of Lightness
- This essay explores lightness as both metaphor and existential feeling. Anna draws on personal experience, especially in relation to weightlifting or strength training (or absence thereof), to illustrate how being unable to “carry the iron” can feel like a loss of grounding. This absence of heaviness becomes a motif: when your body can’t bear weight, or you’re missing a part of your own routine, life feels insubstantial, untethered. She suggests that weight (in the literal or metaphorical sense) gives things substance and meaning. Without it, everything feels ethereal, faint—and yet paradoxically, that lightness can be unbearable. She echos the title of Kundera’s The Unbearable Lightness of Being and riffs on that tension between the meaningful weight of experience and the drift of what happens when weight is lost.
- A trimodal protein language model enables advanced protein searches
- ProTrek is built to understand three “languages” of proteins—sequence (the letters making up the protein), structure (its physical 3D shape), and natural language (textual descriptions like “enzyme that repairs dna”). It embeds all three modes into a shared latent space through clever contrastive learning. In practice, that means you could query a protein by giving any one of those modes and retrieve matches in any of the others.
- Regression Language Models for Code
- This paper pushes that idea in a slightly different direction: what if a language model could predict numeric properties of code—things like how much memory it uses, how fast it runs, or how accurate a neural net implemented in code will be?
- Time100 ai 2025
- According to time, these are the “innovators, leaders, and thinkers reshaping our world through groundbreaking advances in artificial intelligence.” I Think there are some very important people missing from this list, more related to research, that have shaped the landscape of AI.
- Church Planting: When Venture Capital Finds Jesus
- Evangelical church planting in the U.S. (especially non-denominational evangelical churches) contain many of the organizational, cultural, financial, and motivational patterns like startup / venture-capital / tech-entrepreneur ecosystems. Basically, church planting (starting new churches) has become a “venture-capital finds Jesus” world: the structure, incentives, growth-metrics, funder / planter relationships, risk / return thinking, charisma, etc., have strong analogues in startup culture. This post walks through what church planters are, how they are funded, what their goals are, how failures happen, what the costs are, and what the human side looks like (burnout, moral risk, etc.). Along the way there are comparisons to tech startups, VC dynamics, “hits-based” industries, etc.
- Conversational Cultures: Combat vs Nurture (V2)
- “Combat vs Nurture” is a proposed framing from Ruby
that explains different conversational ecosystems with different trade-offs. Neither is “always right.” Depending on the group, topic, relational context, emotional stakes, one or the other (or a mix) is more appropriate.
- “Combat vs Nurture” is a proposed framing from Ruby
- Accelerating life sciences research
- OpenAI has “successfully leveraged gpt‑4b micro to design novel and significantly enhanced variants of the Yamanaka factors, a set of proteins which led to a Nobel Prize for their role in generating induced pluripotent stem cells (iPSCs) and rejuvenating cells. They have also been used to develop therapeutics to combat blindness(opens in a new window), reverse diabetes, treat infertility, and address organ shortages.”
- Collective alignment: public input on our Model Spec
- Oai surveyed over 1,000 people worldwide on how their models should behave and compared user’s views to the current Model Spec. They found that they largely agree with the Spec, and changes were adopted from the disagreements.”
- The Ultra-Scale Playbook: Training llms on gpu Clusters
- If you’re someone who’s built a small model, or done training on a single gpu or a small multi-gpu rig, and wondered how people train massive models with 70b+ or 400b+ parameters, this playbook is like a guided map through that terrain. It lays out the tools, tricks, trade-offs, and gives you real data and code you can use. It doesn’t promise an exact recipe that works everywhere, but it gives you the intuition to design your own recipe, measure what matters in your setting, and avoid some of the big surprises: memory bottlenecks, wasted gpu time, communication lag, etc. If you’re serious about scaling up llm training, it’s one of the better resources out there right now.
- The Rationality Trap
- This is a sharp, introspective, and somewhat somber exploration of how the rationality movement, founded on the aspiration to think better and save humanity, drifted into dangerous territory. It doesn’t paint everything in the rationalist / AI safety world as bad, but Weiss-Blatt shows that certain mythologies, internal methods, stakes framing, and funding structures may have created fertile ground for cultish dynamics.
- This is the World’s Biggest Animal Migration
- Scientists have used aerial surveys and tracking to document a migration of millions of antelope (white-eared kob, tiang, Mongalla gazelle, bohor reedbuck) moving across huge distances in search of water and grazing. This is now being called the Great Nile Migration—possibly the largest land mammal migration on Earth, both in sheer numbers and spatial scale.
- Coda: Coordinating the Cerebrum and Cerebellum for a Dual-Brain Computer Use Agent with Decoupled Reinforcement Learning

- Think in Games: Learning to Reason in Games via Reinforcement Learning with Large Language Models

- The Landscape of Agentic Reinforcement Learning for llms: A Survey

- Reinforcement Learning Foundations for Deep Research Systems: A Survey

- The Illusion of Diminishing Returns: Measuring Long Horizon Execution in llms
- “Failures of llms when simple tasks are made longer arise from mistakes in execution, rather than an inability to reason. We propose isolating execution capability, by explicitly providing the knowledge and plan needed to solve a long-horizon task. We find that larger models can correctly execute significantly more turns even when small models have 100% single-turn accuracy.”
- Contemplative Artificial Intelligence
- “As artificial intelligence (AI) improves, traditional alignment strategies may falter in the face of unpredictable self-improvement, hidden subgoals, and the sheer complexity of intelligent systems. Inspired by contemplative wisdom traditions, [the paper] shows how four axiomatic principles can instil a resilient Wise World Model in AI systems. For future systems, active inference may offer the self-organizing and dynamic coupling capabilities needed to enact Contemplative AI in embodied agents.”
- The entropic brain: a theory of conscious states informed by neuroimaging research with psychedelic drugs
- This paper argues that conscious states vary along a spectrum of neural entropy. Normal waking consciousness is relatively constrained (less entropy), maintaining self, reality testing, etc. Psychedelics (and other primary consciousness states) raise entropy: more flexibility, more novel associations, less ego control, etc. Neuroimaging evidence from psilocybin (drops in dmn connectivity, drops in regular oscillatory power in some bands, increased metastability and variance, more diverse connectivity motifs) supports this. The theory has implications not just for how we understand consciousness, but for mental health and therapy.
- Efficient nano-photonic antennas based on dark states in quantum emitter rings
- Imagine you have a ring of quantum emitters—little things that can absorb light and later emit it (atoms, molecules, etc.). When many such emitters are arranged in a ring, interesting collective effects emerge. Some of the collective states are bright (they radiate strongly), and some are dark, meaning they radiate very little. They sort of trap excitation among the emitters instead of letting it leak out quickly. This paper studies how one can use a ring of emitters, together with an extra “absorber” placed at the center of the ring, to build a highly efficient nano-antenna that captures incoming light (or a photon) and absorbs it efficiently at the center, exploiting those dark states. The idea is that a dark state helps concentrate energy at the center (to be absorbed) without letting much be lost by radiation from the ring itself. They explore how many emitters in the ring works best, how the geometry and coupling matter, how detunings (mismatch in frequencies) affect things, etc. A particularly interesting result is that nine emitters in the ring (i.e. a nonagon) gives a “sweet spot” for absorption efficiency under many conditions. In many nanoscale and quantum optical technologies (e.g. light harvesting, single-photon detectors, quantum sensors), getting light in and absorbing or detecting it efficiently is hard. Losses by radiation, scattering, mismatch, etc. degrade performance. This ring + absorber design gives a way to reduce unwanted radiative losses (via the dark state) while still capturing energy effectively where you want it.
- The Self-Organization of Insight: Entropy and Power Laws in Problem Solving
- The authors are trying to explain how people suddenly gain insight, which is the moment when something “clicks”, and a new structure or strategy emerges that lets one solve a problem in a qualitatively different way. They argue that traditional models in cognitive science (symbolic representations, fixed procedures, etc.) don’t do a great job capturing those abrupt, emergent changes. Instead, Stephen & Dixon suggest thinking of cognition (e.g. insight, problem solving) as a nonlinear dynamical system. This system is open, continuously interacting with the environment, where internal change (in thought, perception, action) can lead to self-organization of new cognitive structure. In this view, insight is not just “applying a known rule” but reorganizing how one is thinking, perceiving, acting.
- Silk Road Timeline
- site acts as a public mirror of Silk Road’s full lifecycle—from vendor storefronts and community forum discourse to legal fallout.
- also see from @thegrugq:
Dread Pirate Roberts—extortion timeline,
The Maryland Indictment Timeline
- Satiety Graphed & The Horsemen of Obesity
- This is a criticism of the common appeal to satiety without a clear definition, arguing that relying on an undefined concept is problematic, even if, by chance, a diet works. Satiety, not to be confused with physical fullness, can indicate metabolic and dietary health, and knowing it when you see it can work as a pragmatic definition.
- Gpt-oss release
- Ollama: OpenAI gpt-oss

- Welcome gpt oss, the new open-source model family from OpenAI!
- Gpt oss is a hugely anticipated open-weights release by OpenAI, designed for powerful reasoning, agentic tasks, and versatile developer use cases. It comprises two models: a big one with 117b parameters (gpt-oss-120b), and a smaller one with 21b parameters (gpt-oss-20b). Both are mixture-of-experts (MoEs) and use a 4-bit quantization scheme (mxfp4), enabling fast inference (thanks to fewer active parameters, see details below) while keeping resource usage low. The large model fits on a single H100 gpu, while the small one runs within 16gb of memory and is perfect for consumer hardware and on-device applications.
- Open models by OpenAI
- Advanced open-weight reasoning models to customize for any use case and run anywhere.
- Ollama: OpenAI gpt-oss
- Lagerstroemia
- commonly known as crape myrtle (also spelled crepe myrtle or crêpe myrtle), is a genus of deciduous and evergreen trees and shrubs native to the Indian subcontinent, southeast Asia, northern Australia, and other parts of Oceania, cultivated in warmer climates around the world.
- I’ve been taking care of a crepe myrtle on my property for several years and quite enjoy seeing it mature from a sapling. A year after moving into my current house, the tree didn’t recover after a harsh winter, so I was forced to cut the tree down, but I left the root system. In a few months, a few saplings began to grow, and after a lot of care and attention, it now has 5 main trunks measuring 6–7 inches in circumference and is over 15ft tall.
- Oral treatment of erectile dysfunction with apomorphine SL
- Apomorphine SL (Ixense, Uprima) is a new oral medication shown to be effective in the treatment of erectile dysfunction. This compound is a dopaminergic agonist with affinity for dopamine receptor sites—mostly D(2)—within the brain known to be involved in sexual function. Apomorphine induces selective activation in the nucleus paraventricularis leading to erectogenic signals.
- Psilocybin treatment extends cellular lifespan and improves survival of aged mice
- Psilocybin, the naturally occurring psychedelic compound produced by hallucinogenic mushrooms, has received attention due to considerable clinical evidence for its therapeutic potential to treat various psychiatric and neurodegenerative indications. However, the underlying molecular mechanisms remain enigmatic, and few studies have explored its systemic impacts. We provide the first experimental evidence that psilocin (the active metabolite of psilocybin) treatment extends cellular lifespan and psilocybin treatment promotes increased longevity in aged mice, suggesting that psilocybin may be a potent geroprotective agent.
- Hitchhiker’s Guide to rag: From Tiny Files to Tolstoy with OpenAI’s api and LangChain
- Scaling a simple rag pipeline from simple notes to full books
- If--
- Kipling’s poem “If—” is essentially a father’s advice to his son (and, by extension, to any young person) on how to live a balanced, honorable life. Written in 1895 and later included in his 1910 collection Rewards and Fairies, it lays out a series of “if” scenarios—tests of character—and promises the ultimate reward (“you’ll be a Man, my son!”) if you can meet them all.
- The Best Tacit Knowledge Videos on Every Subject

- Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task
- The study investigates the neural underpinnings of creative writing across different participant groups: “Brain-only” (unassisted), “llm” (Large Language Model-assisted), and “Search Engine” (search engine-assisted). A primary focus is on Delta band connectivity, which consistently shows the most significant disparities between groups. The Brain-only group demonstrates significantly higher and more widespread Delta band activity, indicating greater engagement of deep, slow integrative brain processes, multisensory integration, and internally-driven thought during unassisted writing. Conversely, llm and Search Engine groups exhibit more externally anchored or intermittently guided cognitive engagement, with notably weaker delta interactions.
- N-gram analysis complements these neurological findings, revealing distinct linguistic patterns across groups for different topics. The “Brain-only” group tends to use n-grams reflecting internal thought processes and prosocial framing, while llm and Search Engine groups show patterns indicative of external sourcing or more direct task-oriented language. The study also tracks changes in brain connectivity patterns across sessions, noting a general trend of increasing connectivity in later sessions for both Brain-only and llm groups, suggesting adaptation to the task
- Ultrafast coherent dynamics of microring modulators
- Next-generation computing clusters require ultra-high-bandwidth optical interconnects to support large-scale artificial-intelligence applications. These electronic–photonic co-integrated systems necessitate densely integrated high-speed electro-optical converters. In this context, microring modulators (mrms) emerge as a promising solution, prized for their exceptional compactness and energy efficiency.
- Potemkin Understanding in Large Language Models
- Success on benchmarks only demonstrates potemkin understanding: the illusion of understanding driven by answers irreconcilable with how any human would interpret a concept.
- goose
- an open source, extensible AI agent that goes beyond code suggestions—install, execute, edit, and test with any llm
- bitchat-android
- bluetooth mesh chat, irc vibes
- Orienting Toward Wizard Power,
Wentworth 2025 - The zingularity framework for Bayesian artificial neural networks,
Janssen 2025 - How to Generate Synthetic Data: A Comprehensive Guide Using Bayesian Sampling and Univariate Distributions,
2025 - Continuous Thought Machines
- Paper: Continuous Thought Machines

- Paper: Continuous Thought Machines
- Neural Thermodynamic Laws for Large Language Model Training,
Tegmark, et al 2025 - Synthetic Data RL: Task Definition Is All You Need,
Guo 2025
- How To build an American DeepSeek
Jeffries 2025 - AI 2027, AI Futures
2025 - An Optimistic 2027 Timeline,
Yitz 2025 - AI 2027: What Superintelligence Looks Like,
Kokotajlo et al, 2025 - An Approach to Technical agi Safety and Security,
2025 - Scott Alexander was right: doubling down,
Teslo 2025 - Tiny Agents: a mcp-powered agent in 50 lines of code,
Chaumond 2025 - Optimal Brain Damage,
Lecun et al, 1989 - The Cacophony,
Kriss 2025 - There’s A Time For Everyone,
acx 2022 - Sensemaking in 2025: Trump Tariffs Edition,
Hall 2025
- re-read: The Extended Internet Universe,
Venkatash Rao 2019 - Ribbonfarm is Retiring,
Rao 2024 - R1-Searcher: Incentivizing the Search Capability in llms via Reinforcement Learning
2025 - Vocabulary Study with Mnemosyne

- Levels of Friction,
Zvi 2025 - On Writing #1,
Zvi 2025 - The Value of Open Source Software,
Hoffman et al 2024 - On the Tolkienic Hero
2019 - Writing for llms So They Listen,
Gwern 2024 - An open letter to graduate students and other procrastinators: it’s time to write
Hazelett 2025
Much of my time this month has been spent on researching tools that support the development of this site and implementing visual and basic quality-of-life features to establish a solid foundation for the future.
- How to Install PmWiki on Debian 10 / Nginx / php-fpm

- Pmwiki—ImagePopup

- Introducing Perplexity Deep Research

- DeepSeek? Schmidhuber did it first
ref: @hardmaru
- On Learning to Think: Algorithmic Information Theory for Novel Combinations of Reinforcement Learning Controllers and Recurrent Neural World Models

- One Big Net For Everything

- Learning complex, extended sequences using the principle of history compression

- 1990: Planning & Reinforcement Learning with Recurrent World Models and Artificial Curiosity

- 1991: First very deep learning with unsupervised pre-training

- On Learning to Think: Algorithmic Information Theory for Novel Combinations of Reinforcement Learning Controllers and Recurrent Neural World Models
- DeepSeek-R1: Incentivizing Reasoning Capability in llms via Reinforcement Learning

- Why To Not Write A Book,
Gwern 2024 - Design of This Website,
Gwern 2023 - Tufte css, David Liepmann
- How To Make Superbabies,
GeneSmith / kman 2025
༻
༺
༺New Links & Articles: newsletter
& rss
koenrane@protonmail.com
