Table of contents

Links are posted monthly in reverse chronological order and include any articles, papers, or essays that I find pertinent to my current research and writing.




  • Candi: Hybrid Discrete-Continuous Diffusion Models, Pynadath et al 2025
    • This paper dives into a quirk of modern generative modeling—while continuous diffusion models (the kind behind lots of cool image generation tech) work great for continuous data, they surprisingly struggle when directly applied to discrete things like text or categorically labeled data. They explain why this happens by analyzing how adding Gaussian noise can scramble discrete tokens in ways that make learning hard. To fix this, they introduce candi, a hybrid approach that blends continuous and discrete diffusion so the model can learn both continuous structure and discrete identities at the same time. According to their experiments, this hybrid method avoids the pitfalls of pure continuous diffusion on discrete spaces and even beats other techniques in some text generation tests, unlocking benefits like easier classifier guidance and stronger generation at low computing costs.


  • Newcomblike problems are the norm, Soares 2014
    • real-world decision scenarios constantly leak information and involve prediction, Newcomblike problems are not rare exceptions but the norm—and any complete theory of decision-making (especially for intelligent systems or AI) needs to grapple with that reality.
  • Why people like your quick bullshit takes better than your high-effort posts, eukaryote 2025
    • differences between the quick post and the effort post
  • How Parasitic Processing Is Causing You To Suffer, Vervaeke 2025
    • our brain’s way of trying to make sense of the world can sometimes turn against us and create a downward spiral of anxiety and negativity, something called parasitic processing
  • Laws of (New) Media, McLuhan 2025
    • Andrew McLuhan (grandson of media theorist Marshall McLuhan) dives into the classic media theory idea known as the “tetrad” or laws of media, a framework his grandfather and father developed to understand how every technology or medium shapes us and society. Rather than just looking at what a technology does, the tetrad encourages us to ask four questions about any medium: what it enhances in our lives, what it obsolesces or displaces, what it retrieves from the past, and how it reverses into something unexpected when pushed too far.
  • Economics of Orbital vs Terrestrial Data Centers, McCalip 2025
    • “Why compute in orbit? Why should a watt or a flop 250 miles up be more valuable than one on the surface? What advantage justifies moving something as mundane as matrix multiplication into leo?”
  • Claude 4.5 Opus Soul Document , Weiss 2025
  • how to party like an AI researcher, Sun 2025
    • NeurIPS 2025 is where cutting-edge AI research quietly happens in the background while the real optimization problem is social networking, parties, and who got invited where


  • Dokkōdō
    • Penned by the famous Japanese swordsman Miyamoto Musashi in 1645, just a week before his death, Dokkōdō (which translates roughly to “The Way of Walking Alone” or “The Path of Aloneness”) offers 21 precepts or guiding principles for life.
  • Tiny Recursive Model (trm)
    • Instead of going big with tons of parameters, the designers built a two-layer model that recurses—meaning it iteratively refines a hidden “scratchpad” (latent space) plus a “current solution” over multiple steps. Think “draft + revise” loops rather than throwing everything in one long pass. They unroll this loop many times (up to 16) during training, supervise it “deeply” (i.e. give intermediate guidance), and allow full backpropagation through all those recursions (rather than using fixed-point approximations). Also, they ditch complex modular hierarchies (used by prior models) and instead do all the reasoning via that one compact recurrent core.
  • The Unbearable Lightness of Lightness
    • This essay explores lightness as both metaphor and existential feeling. Anna draws on personal experience, especially in relation to weightlifting or strength training (or absence thereof), to illustrate how being unable to “carry the iron” can feel like a loss of grounding. This absence of heaviness becomes a motif: when your body can’t bear weight, or you’re missing a part of your own routine, life feels insubstantial, untethered. She suggests that weight (in the literal or metaphorical sense) gives things substance and meaning. Without it, everything feels ethereal, faint—and yet paradoxically, that lightness can be unbearable. She echos the title of Kundera’s The Unbearable Lightness of Being and riffs on that tension between the meaningful weight of experience and the drift of what happens when weight is lost.
  • A trimodal protein language model enables advanced protein searches
    • ProTrek is built to understand three “languages” of proteins—sequence (the letters making up the protein), structure (its physical 3D shape), and natural language (textual descriptions like “enzyme that repairs dna”). It embeds all three modes into a shared latent space through clever contrastive learning. In practice, that means you could query a protein by giving any one of those modes and retrieve matches in any of the others.
  • Regression Language Models for Code
    • This paper pushes that idea in a slightly different direction: what if a language model could predict numeric properties of code—things like how much memory it uses, how fast it runs, or how accurate a neural net implemented in code will be?


  • Time100 ai 2025
    • According to time, these are the “innovators, leaders, and thinkers reshaping our world through groundbreaking advances in artificial intelligence.” I Think there are some very important people missing from this list, more related to research, that have shaped the landscape of AI.
  • Church Planting: When Venture Capital Finds Jesus
    • Evangelical church planting in the U.S. (especially non-denominational evangelical churches) contain many of the organizational, cultural, financial, and motivational patterns like startup / venture-capital / tech-entrepreneur ecosystems. Basically, church planting (starting new churches) has become a “venture-capital finds Jesus” world: the structure, incentives, growth-metrics, funder / planter relationships, risk / return thinking, charisma, etc., have strong analogues in startup culture. This post walks through what church planters are, how they are funded, what their goals are, how failures happen, what the costs are, and what the human side looks like (burnout, moral risk, etc.). Along the way there are comparisons to tech startups, VC dynamics, “hits-based” industries, etc.
  • Conversational Cultures: Combat vs Nurture (V2)
    • “Combat vs Nurture” is a proposed framing from Ruby that explains different conversational ecosystems with different trade-offs. Neither is “always right.” Depending on the group, topic, relational context, emotional stakes, one or the other (or a mix) is more appropriate.
  • Accelerating life sciences research
    • OpenAI has “successfully leveraged gpt4b micro to design novel and significantly enhanced variants of the Yamanaka factors, a set of proteins which led to a Nobel Prize for their role in generating induced pluripotent stem cells (iPSCs) and rejuvenating cells. They have also been used to develop therapeutics to combat blindness⁠(opens in a new window), reverse diabetes⁠, treat infertility⁠, and address organ shortages⁠.”
  • Collective alignment: public input on our Model Spec
    • Oai surveyed over 1,000 people worldwide on how their models should behave and compared user’s views to the current Model Spec. They found that they largely agree with the Spec, and changes were adopted from the disagreements.”
  • The Ultra-Scale Playbook: Training llms on gpu Clusters
    • If you’re someone who’s built a small model, or done training on a single gpu or a small multi-gpu rig, and wondered how people train massive models with 70b+ or 400b+ parameters, this playbook is like a guided map through that terrain. It lays out the tools, tricks, trade-offs, and gives you real data and code you can use. It doesn’t promise an exact recipe that works everywhere, but it gives you the intuition to design your own recipe, measure what matters in your setting, and avoid some of the big surprises: memory bottlenecks, wasted gpu time, communication lag, etc. If you’re serious about scaling up llm training, it’s one of the better resources out there right now.
  • The Rationality Trap
    • This is a sharp, introspective, and somewhat somber exploration of how the rationality movement, founded on the aspiration to think better and save humanity, drifted into dangerous territory. It doesn’t paint everything in the rationalist / AI safety world as bad, but Weiss-Blatt shows that certain mythologies, internal methods, stakes framing, and funding structures may have created fertile ground for cultish dynamics.
  • This is the World’s Biggest Animal Migration
    • Scientists have used aerial surveys and tracking to document a migration of millions of antelope (white-eared kob, tiang, Mongalla gazelle, bohor reedbuck) moving across huge distances in search of water and grazing. This is now being called the Great Nile Migration—possibly the largest land mammal migration on Earth, both in sheer numbers and spatial scale.

  • Coda: Coordinating the Cerebrum and Cerebellum for a Dual-Brain Computer Use Agent with Decoupled Reinforcement Learning
  • Think in Games: Learning to Reason in Games via Reinforcement Learning with Large Language Models
  • The Landscape of Agentic Reinforcement Learning for llms: A Survey
  • Reinforcement Learning Foundations for Deep Research Systems: A Survey
  • The Illusion of Diminishing Returns: Measuring Long Horizon Execution in llms
    • “Failures of llms when simple tasks are made longer arise from mistakes in execution, rather than an inability to reason. We propose isolating execution capability, by explicitly providing the knowledge and plan needed to solve a long-horizon task. We find that larger models can correctly execute significantly more turns even when small models have 100% single-turn accuracy.”
  • Contemplative Artificial Intelligence
    • “As artificial intelligence (AI) improves, traditional alignment strategies may falter in the face of unpredictable self-improvement, hidden subgoals, and the sheer complexity of intelligent systems. Inspired by contemplative wisdom traditions, [the paper] shows how four axiomatic principles can instil a resilient Wise World Model in AI systems. For future systems, active inference may offer the self-organizing and dynamic coupling capabilities needed to enact Contemplative AI in embodied agents.”
  • The entropic brain: a theory of conscious states informed by neuroimaging research with psychedelic drugs
    • This paper argues that conscious states vary along a spectrum of neural entropy. Normal waking consciousness is relatively constrained (less entropy), maintaining self, reality testing, etc. Psychedelics (and other primary consciousness states) raise entropy: more flexibility, more novel associations, less ego control, etc. Neuroimaging evidence from psilocybin (drops in dmn connectivity, drops in regular oscillatory power in some bands, increased metastability and variance, more diverse connectivity motifs) supports this. The theory has implications not just for how we understand consciousness, but for mental health and therapy.
  • Efficient nano-photonic antennas based on dark states in quantum emitter rings
    • Imagine you have a ring of quantum emitters—little things that can absorb light and later emit it (atoms, molecules, etc.). When many such emitters are arranged in a ring, interesting collective effects emerge. Some of the collective states are bright (they radiate strongly), and some are dark, meaning they radiate very little. They sort of trap excitation among the emitters instead of letting it leak out quickly. This paper studies how one can use a ring of emitters, together with an extra “absorber” placed at the center of the ring, to build a highly efficient nano-antenna that captures incoming light (or a photon) and absorbs it efficiently at the center, exploiting those dark states. The idea is that a dark state helps concentrate energy at the center (to be absorbed) without letting much be lost by radiation from the ring itself. They explore how many emitters in the ring works best, how the geometry and coupling matter, how detunings (mismatch in frequencies) affect things, etc. A particularly interesting result is that nine emitters in the ring (i.e. a nonagon) gives a “sweet spot” for absorption efficiency under many conditions. In many nanoscale and quantum optical technologies (e.g. light harvesting, single-photon detectors, quantum sensors), getting light in and absorbing or detecting it efficiently is hard. Losses by radiation, scattering, mismatch, etc. degrade performance. This ring + absorber design gives a way to reduce unwanted radiative losses (via the dark state) while still capturing energy effectively where you want it.
  • The Self-Organization of Insight: Entropy and Power Laws in Problem Solving
    • The authors are trying to explain how people suddenly gain insight, which is the moment when something “clicks”, and a new structure or strategy emerges that lets one solve a problem in a qualitatively different way. They argue that traditional models in cognitive science (symbolic representations, fixed procedures, etc.) don’t do a great job capturing those abrupt, emergent changes. Instead, Stephen & Dixon suggest thinking of cognition (e.g. insight, problem solving) as a nonlinear dynamical system. This system is open, continuously interacting with the environment, where internal change (in thought, perception, action) can lead to self-organization of new cognitive structure. In this view, insight is not just “applying a known rule” but reorganizing how one is thinking, perceiving, acting.


  • Silk Road Timeline
  • Satiety Graphed & The Horsemen of Obesity
    • This is a criticism of the common appeal to satiety without a clear definition, arguing that relying on an undefined concept is problematic, even if, by chance, a diet works. Satiety, not to be confused with physical fullness, can indicate metabolic and dietary health, and knowing it when you see it can work as a pragmatic definition.
  • Gpt-oss release
    • Ollama: OpenAI gpt-oss
    • Welcome gpt oss, the new open-source model family from OpenAI!
      • Gpt oss is a hugely anticipated open-weights release by OpenAI, designed for powerful reasoning, agentic tasks, and versatile developer use cases. It comprises two models: a big one with 117b parameters (gpt-oss-120b), and a smaller one with 21b parameters (gpt-oss-20b). Both are mixture-of-experts (MoEs) and use a 4-bit quantization scheme (mxfp4), enabling fast inference (thanks to fewer active parameters, see details below) while keeping resource usage low. The large model fits on a single H100 gpu, while the small one runs within 16gb of memory and is perfect for consumer hardware and on-device applications.
    • Open models by OpenAI
      • Advanced open-weight reasoning models to customize for any use case and run anywhere.
  • Lagerstroemia
    • commonly known as crape myrtle (also spelled crepe myrtle or crêpe myrtle), is a genus of deciduous and evergreen trees and shrubs native to the Indian subcontinent, southeast Asia, northern Australia, and other parts of Oceania, cultivated in warmer climates around the world.
    • I’ve been taking care of a crepe myrtle on my property for several years and quite enjoy seeing it mature from a sapling. A year after moving into my current house, the tree didn’t recover after a harsh winter, so I was forced to cut the tree down, but I left the root system. In a few months, a few saplings began to grow, and after a lot of care and attention, it now has 5 main trunks measuring 6–7 inches in circumference and is over 15ft tall.

  • Oral treatment of erectile dysfunction with apomorphine SL
    • Apomorphine SL (Ixense, Uprima) is a new oral medication shown to be effective in the treatment of erectile dysfunction. This compound is a dopaminergic agonist with affinity for dopamine receptor sites—mostly D(2)—within the brain known to be involved in sexual function. Apomorphine induces selective activation in the nucleus paraventricularis leading to erectogenic signals.
  • Psilocybin treatment extends cellular lifespan and improves survival of aged mice
    • Psilocybin, the naturally occurring psychedelic compound produced by hallucinogenic mushrooms, has received attention due to considerable clinical evidence for its therapeutic potential to treat various psychiatric and neurodegenerative indications. However, the underlying molecular mechanisms remain enigmatic, and few studies have explored its systemic impacts. We provide the first experimental evidence that psilocin (the active metabolite of psilocybin) treatment extends cellular lifespan and psilocybin treatment promotes increased longevity in aged mice, suggesting that psilocybin may be a potent geroprotective agent.


  • Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task
    • The study investigates the neural underpinnings of creative writing across different participant groups: “Brain-only” (unassisted), “llm” (Large Language Model-assisted), and “Search Engine” (search engine-assisted). A primary focus is on Delta band connectivity, which consistently shows the most significant disparities between groups. The Brain-only group demonstrates significantly higher and more widespread Delta band activity, indicating greater engagement of deep, slow integrative brain processes, multisensory integration, and internally-driven thought during unassisted writing. Conversely, llm and Search Engine groups exhibit more externally anchored or intermittently guided cognitive engagement, with notably weaker delta interactions.
    • N-gram analysis complements these neurological findings, revealing distinct linguistic patterns across groups for different topics. The “Brain-only” group tends to use n-grams reflecting internal thought processes and prosocial framing, while llm and Search Engine groups show patterns indicative of external sourcing or more direct task-oriented language. The study also tracks changes in brain connectivity patterns across sessions, noting a general trend of increasing connectivity in later sessions for both Brain-only and llm groups, suggesting adaptation to the task
  • Ultrafast coherent dynamics of microring modulators
    • Next-generation computing clusters require ultra-high-bandwidth optical interconnects to support large-scale artificial-intelligence applications. These electronic–photonic co-integrated systems necessitate densely integrated high-speed electro-optical converters. In this context, microring modulators (mrms) emerge as a promising solution, prized for their exceptional compactness and energy efficiency.
  • Potemkin Understanding in Large Language Models
    • Success on benchmarks only demonstrates potemkin understanding: the illusion of understanding driven by answers irreconcilable with how any human would interpret a concept.

  • goose
    • an open source, extensible AI agent that goes beyond code suggestions—install, execute, edit, and test with any llm
  • bitchat-android
    • bluetooth mesh chat, irc vibes





Much of my time this month has been spent on researching tools that support the development of this site and implementing visual and basic quality-of-life features to establish a solid foundation for the future.

KR

New Links & Articles: newsletter & rssRSS icon