Tuesday, March 24, 2026

WebAssembly inside a Transformer; “computers inside transformers”

 The work you’re referring to is a very recent research direction (2025–2026) associated with Christos Tzamos and collaborators, often discussed under “computers inside transformers” or “WASM-in-transformer”.

Here’s a clear breakdown of what it is and why it’s important:


๐Ÿ”ง Core Idea: WebAssembly inside a Transformer

Recent work (e.g., from Percepta AI) shows that a transformer can embed and execute a WebAssembly (WASM) interpreter directly inside its weights.

  • Instead of calling external tools or running code separately
  • The model itself executes programs step-by-step internally

๐Ÿ‘‰ In effect, the transformer becomes both:

  • a neural model (learning patterns)
  • AND a symbolic computer (executing programs)

They “implemented a WebAssembly interpreter inside the transformer weights”


⚙️ What exactly is happening?

1. Compile a computer into the model

  • A RAM-like abstract machine + WASM interpreter is encoded into the transformer
  • Represented using attention + matrix operations

2. Feed in a program

  • The input tokens include:
    • program (in WASM or compiled form)
    • initial memory/state

3. Transformer executes it

  • Each forward pass simulates steps of execution
  • Produces:
    • execution trace
    • final output

๐Ÿ‘‰ No Python, no external runtime — everything happens inside the transformer


๐Ÿš€ Key Results (from reports)

  • Deterministic execution (not probabilistic text generation)
  • Runs tens of thousands of steps per second on CPU (~33K tokens/sec)
  • Demonstrates that transformers can act as general-purpose computers

๐Ÿง  Why WebAssembly?

WebAssembly (WASM) is ideal because:

  • Low-level, structured bytecode (like a portable assembly)
  • Deterministic and safe
  • Easy to model as state transitions

So it becomes a natural bridge between:

  • symbolic computation
  • neural computation

๐Ÿงฉ Conceptual Shift

This work challenges the usual architecture:

Traditional LLMTzamos-style system
Predicts next tokenExecutes programs
Uses external tools (Python, APIs)Internal execution
ProbabilisticDeterministic
Text interfaceComputational substrate

๐Ÿ‘‰ This suggests:

Transformers are not just language models — they can be universal computers


๐Ÿ”ฌ Relation to Theory

This connects to known results:

  • Transformers are Turing complete (in theory)
  • This work shows a practical construction
  • Bridges:
    • neural networks
    • programming languages
    • compilers

๐Ÿงช Implications for Research (very relevant to your work)

Given your interests (AI + biology + hypothesis-driven models), this paradigm could enable:

1. Hybrid symbolic–neural models

  • Embed biological simulation rules inside transformers
  • Combine:
    • learned representations
    • mechanistic models

2. Deterministic reasoning inside LLMs

  • Replace:
    • fragile chain-of-thought
  • With:
    • executable programs inside the model

3. “Model = program + data”

  • Instead of prompting → reasoning
  • You compile algorithms into weights

⚠️ Limitations / Open Questions

  • Scaling to large real-world programs
  • Training vs hand-encoding interpreters
  • Efficiency vs specialized runtimes
  • Debuggability

๐Ÿงญ Bottom Line

The “Christos Tzamos + WebAssembly + transformer” work shows:

A transformer can contain and execute a full programming language runtime internally, effectively turning it into a neural computer.

This is a major step toward:

  • neural-symbolic unification
  • tool-free AI systems
  • and potentially programmable foundation models

openai life sciences

 

https://openaifoundation.org/news/update-on-the-openai-foundation

At the Foundation, we’ve identified three initial focus areas where we think this work could make a real difference:

  • AI for Alzheimer’s: Alzheimer’s is one of the hardest and most heartbreaking diseases families face – and one of the toughest problems in medicine. AI’s ability to reason across complex data could help researchers uncover new insights. We will be partnering with leading research institutions, with an initial focus on mapping disease pathways, detecting biomarkers for clinical care and clinical trials, and accelerating personalization of treatments – including, where possible, repurposing existing FDA-approved molecules.

  • Public Data for Health: Many of medicine’s biggest advances have been made possible by shared scientific data, and public access to data is essential to deliver the promise of AI for scientific breakthroughs. We will help partners create and expand open, high-quality datasets – and, where appropriate, help responsibly open previously closed ones – so researchers everywhere can leverage AI and use data to drive progress across diseases.

  • Accelerating Progress on High-Mortality and High-Burden Diseases: We believe AI can help lead to scientific breakthroughs, and lower the cost and risk of developing or repurposing therapies, particularly in high-mortality and high-burden disease areas that are underfunded. We will bring together AI researchers and disease experts, starting with a focused workshop to identify how best to empower scientists with AI tools and surface promising opportunities.

Saturday, March 7, 2026

DNABERT waterfield

 

todo: DNABERT2


DNABERT is installed.


  - Found repo: https://github.com/jerryji1993/DNABERT.git

  - Cloned to: DNABERT (commit b6da04e)

  - Created Conda env: dnabert (Python 3.6)

  - Installed:

      - DNABERT in editable mode (pip install -e .)

      - Example dependencies

      - pybedtools/pysam (via mamba + pip wheel fallback)

      - PyTorch stack (torch 1.10.2+cu102, torchvision 0.11.2)


  Verification passed in dnabert:


  - torch 1.10.2+cu102

  - transformers 2.5.0

  - pybedtools 0.8.1

  - pysam 0.19.0


  Use it with:


  source ~/miniforge3/etc/profile.d/conda.sh

  conda activate dnabert

  cd /home/xxx/github/DNABERT




Tuesday, March 3, 2026

waterfield GPUs

 

GPU Inventory Summary

GPU TypePartitionTotal NodesGPUs per NodeTotal GPUs
NVIDIA RTX P6000rtxp6000flex-130130
rtxp6000flex-212224
rtxp6000flex-48432
rtxp6000flex-88864
NVIDIA H100h100flex-130130
h100flex-28216
h100flex-48432
h100flex-88864
NVIDIA H200h200flex-88864
NVIDIA B200b200flex-82816
NVIDIA MegaGPU*reserved-clync008188

Total Counts by Model

  • RTX P6000 Series: 150 GPUs

  • H100 Series: 142 GPUs

  • H200 Series: 64 GPUs

  • B200 Series: 16 GPUs

  • Reserved/Special: 8 GPUs

Thursday, February 19, 2026

modern's mrna flu vaccine

 The story of Moderna's mRNA flu vaccine (mRNA-1010) has recently shifted from a regulatory stalemate to a potential path forward, largely due to a rare public disagreement and subsequent reversal by the FDA.

The Conflict and Reversal

Initially, the FDA had expressed significant hesitation regarding Moderna's data, leading to what reports described as a public dispute between the agency and the pharmaceutical company. Moderna's vaccine had shown strong results against some flu strains (Influenza A) but was less effective against others (Influenza B).

However, in a notable turn of events, the FDA reversed its course and agreed to officially review the vaccine. This change of heart reportedly followed significant pressure from the White House, as the administration looks to modernize flu shot technology beyond traditional egg-based methods.

Key Milestones in the Story

  • The Technology: Unlike traditional flu shots, Moderna uses mRNA technology (the same used in their COVID-19 vaccine), which allows for faster manufacturing and the ability to update the vaccine more quickly to match circulating strains.

  • Mixed Data: Clinical trials showed the vaccine was highly effective at generating antibodies for the H1N1 and H3N2 strains but struggled to meet the "non-inferiority" bar for the B/Victoria and B/Yamagata strains compared to existing vaccines.

  • The Resolution: After the public back-and-forth, the FDA will now consider the new flu shot for approval, aiming to resolve the technical and regulatory disagreements that had previously stalled its progress.

Why This Matters

If approved, Moderna's vaccine would be the first mRNA flu shot on the market. This is seen as a crucial step toward their ultimate goal: a combination vaccine that protects against the flu, COVID-19, and potentially RSV in a single annual shot.


Would you like me to look up the latest clinical trial efficacy numbers for Moderna's combination flu/COVID vaccine?

In clinical trials, Moderna's flu vaccine (mRNA-1010) faced a specific regulatory hurdle: while it successfully induced a strong immune response against the "A" strains of the flu, it initially fell short of the required performance standards for the "B" strains.

To understand why this mattered, you have to look at the four specific targets of a "quadrivalent" (four-strain) flu shot:

1. The Targets: A vs. B Strains

Seasonal flu vaccines are designed to protect against four distinct versions of the virus that circulate in humans:

  • Influenza A (H1N1 and H3N2): These are the "heavy hitters." They are more common, mutate rapidly, and are responsible for most flu-related hospitalizations and all historical pandemics. mRNA-1010 performed exceptionally well here, showing results that were equal to or better than traditional shots.

  • Influenza B (Victoria and Yamagata): These lineages only infect humans and generally mutate more slowly. While they cause fewer hospitalizations in adults than Type A, they are a major cause of illness in children.

2. The "Non-Inferiority" Bar

In vaccine science, a new product doesn't always have to be "better" than what already exists; it often just needs to prove it is "not unacceptably worse." This is the non-inferiority bar.

  • The Test: Researchers compare the antibody levels (geometric mean titers) produced by the new vaccine against those produced by an already approved "gold standard" vaccine.

  • The Result: For the B/Victoria and B/Yamagata strains, Moderna’s early data showed that the antibody levels were lower than those produced by the traditional vaccines. They didn't miss by a lot, but they failed to meet the strict statistical threshold (the "bar") required to prove they were "just as good" as the shots people already get.

3. Why did the mRNA struggle with Type B?

This is a technical challenge scientists are still solving. mRNA vaccines work by teaching your cells to make a specific protein (the hemagglutinin or HA). For reasons not fully understood, the mRNA instructions for the Type B HA proteins didn't seem to "take" as strongly as the instructions for Type A.

Note: Since 2020, the B/Yamagata strain has actually disappeared from global circulation (likely due to COVID-19 lockdowns). As a result, the FDA and WHO have recently recommended moving back to trivalent (three-strain) vaccines that exclude Yamagata entirely.

Current Status

Moderna has since updated its formulation. Their latest Phase 3 data (from June 2025) showed that the newer version of mRNA-1010 finally achieved "superiority" or "non-inferiority" across all targeted strains, including B/Victoria. This breakthrough is what finally prompted the FDA to agree to review the vaccine for approval.


Would you like me to look up the specific side effects reported in these mRNA flu trials compared to traditional egg-based shots?