As a reminder, for any presentations or publications showcasing your pilot work, please make sure to include a funding acknowledgment: “Research reported in this publication/ presentation was supported by the National Institute On Aging of the National Institutes of Health under Award Number P30AG073105. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.”
Open Notebook
This site is to serve as my note-book and to effectively communicate with my students and collaborators. Every now and then, a blog may be of interest to other researchers or teachers. Views in this blog are my own. All rights of research results and findings on this blog are reserved. See also http://youtube.com/c/hongqin @hongqin
Friday, March 27, 2026
Tuesday, March 24, 2026
WebAssembly inside a Transformer; “computers inside transformers”
The work you’re referring to is a very recent research direction (2025–2026) associated with Christos Tzamos and collaborators, often discussed under “computers inside transformers” or “WASM-in-transformer”.
Here’s a clear breakdown of what it is and why it’s important:
๐ง Core Idea: WebAssembly inside a Transformer
Recent work (e.g., from Percepta AI) shows that a transformer can embed and execute a WebAssembly (WASM) interpreter directly inside its weights.
- Instead of calling external tools or running code separately
- The model itself executes programs step-by-step internally
๐ In effect, the transformer becomes both:
- a neural model (learning patterns)
- AND a symbolic computer (executing programs)
They “implemented a WebAssembly interpreter inside the transformer weights”
⚙️ What exactly is happening?
1. Compile a computer into the model
- A RAM-like abstract machine + WASM interpreter is encoded into the transformer
- Represented using attention + matrix operations
2. Feed in a program
-
The input tokens include:
- program (in WASM or compiled form)
- initial memory/state
3. Transformer executes it
- Each forward pass simulates steps of execution
-
Produces:
- execution trace
- final output
๐ No Python, no external runtime — everything happens inside the transformer
๐ Key Results (from reports)
- Deterministic execution (not probabilistic text generation)
- Runs tens of thousands of steps per second on CPU (~33K tokens/sec)
- Demonstrates that transformers can act as general-purpose computers
๐ง Why WebAssembly?
WebAssembly (WASM) is ideal because:
- Low-level, structured bytecode (like a portable assembly)
- Deterministic and safe
- Easy to model as state transitions
So it becomes a natural bridge between:
- symbolic computation
- neural computation
๐งฉ Conceptual Shift
This work challenges the usual architecture:
| Traditional LLM | Tzamos-style system |
|---|---|
| Predicts next token | Executes programs |
| Uses external tools (Python, APIs) | Internal execution |
| Probabilistic | Deterministic |
| Text interface | Computational substrate |
๐ This suggests:
Transformers are not just language models — they can be universal computers
๐ฌ Relation to Theory
This connects to known results:
- Transformers are Turing complete (in theory)
- This work shows a practical construction
-
Bridges:
- neural networks
- programming languages
- compilers
๐งช Implications for Research (very relevant to your work)
Given your interests (AI + biology + hypothesis-driven models), this paradigm could enable:
1. Hybrid symbolic–neural models
- Embed biological simulation rules inside transformers
-
Combine:
- learned representations
- mechanistic models
2. Deterministic reasoning inside LLMs
-
Replace:
- fragile chain-of-thought
-
With:
- executable programs inside the model
3. “Model = program + data”
- Instead of prompting → reasoning
- You compile algorithms into weights
⚠️ Limitations / Open Questions
- Scaling to large real-world programs
- Training vs hand-encoding interpreters
- Efficiency vs specialized runtimes
- Debuggability
๐งญ Bottom Line
The “Christos Tzamos + WebAssembly + transformer” work shows:
A transformer can contain and execute a full programming language runtime internally, effectively turning it into a neural computer.
This is a major step toward:
- neural-symbolic unification
- tool-free AI systems
- and potentially programmable foundation models
openai life sciences
https://openaifoundation.org/news/update-on-the-openai-foundation
At the Foundation, we’ve identified three initial focus areas where we think this work could make a real difference:
AI for Alzheimer’s: Alzheimer’s is one of the hardest and most heartbreaking diseases families face – and one of the toughest problems in medicine. AI’s ability to reason across complex data could help researchers uncover new insights. We will be partnering with leading research institutions, with an initial focus on mapping disease pathways, detecting biomarkers for clinical care and clinical trials, and accelerating personalization of treatments – including, where possible, repurposing existing FDA-approved molecules.
Public Data for Health: Many of medicine’s biggest advances have been made possible by shared scientific data, and public access to data is essential to deliver the promise of AI for scientific breakthroughs. We will help partners create and expand open, high-quality datasets – and, where appropriate, help responsibly open previously closed ones – so researchers everywhere can leverage AI and use data to drive progress across diseases.
Accelerating Progress on High-Mortality and High-Burden Diseases: We believe AI can help lead to scientific breakthroughs, and lower the cost and risk of developing or repurposing therapies, particularly in high-mortality and high-burden disease areas that are underfunded. We will bring together AI researchers and disease experts, starting with a focused workshop to identify how best to empower scientists with AI tools and surface promising opportunities.
Saturday, March 7, 2026
DNABERT waterfield
todo: DNABERT2
• DNABERT is installed.
- Found repo: https://github.com/jerryji1993/DNABERT.git
- Cloned to: DNABERT (commit b6da04e)
- Created Conda env: dnabert (Python 3.6)
- Installed:
- DNABERT in editable mode (pip install -e .)
- Example dependencies
- pybedtools/pysam (via mamba + pip wheel fallback)
- PyTorch stack (torch 1.10.2+cu102, torchvision 0.11.2)
Verification passed in dnabert:
- torch 1.10.2+cu102
- transformers 2.5.0
- pybedtools 0.8.1
- pysam 0.19.0
Use it with:
source ~/miniforge3/etc/profile.d/conda.sh
conda activate dnabert
cd /home/xxx/github/DNABERT
Tuesday, March 3, 2026
waterfield GPUs
GPU Inventory Summary
| GPU Type | Partition | Total Nodes | GPUs per Node | Total GPUs |
| NVIDIA RTX P6000 | rtxp6000flex-1 | 30 | 1 | 30 |
| rtxp6000flex-2 | 12 | 2 | 24 | |
| rtxp6000flex-4 | 8 | 4 | 32 | |
| rtxp6000flex-8 | 8 | 8 | 64 | |
| NVIDIA H100 | h100flex-1 | 30 | 1 | 30 |
| h100flex-2 | 8 | 2 | 16 | |
| h100flex-4 | 8 | 4 | 32 | |
| h100flex-8 | 8 | 8 | 64 | |
| NVIDIA H200 | h200flex-8 | 8 | 8 | 64 |
| NVIDIA B200 | b200flex-8 | 2 | 8 | 16 |
| NVIDIA MegaGPU* | reserved-clync008 | 1 | 8 | 8 |
Total Counts by Model
RTX P6000 Series: 150 GPUs
H100 Series: 142 GPUs
H200 Series: 64 GPUs
B200 Series: 16 GPUs
Reserved/Special: 8 GPUs
Thursday, February 19, 2026
modern's mrna flu vaccine
The story of Moderna's mRNA flu vaccine (mRNA-1010) has recently shifted from a regulatory stalemate to a potential path forward, largely due to a rare public disagreement and subsequent reversal by the FDA.
The Conflict and Reversal
Initially, the FDA had expressed significant hesitation regarding Moderna's data, leading to what reports described as a
However, in a notable turn of events, the FDA
Key Milestones in the Story
The Technology: Unlike traditional flu shots, Moderna uses mRNA technology (the same used in their COVID-19 vaccine), which allows for faster manufacturing and the ability to update the vaccine more quickly to match circulating strains.
Mixed Data: Clinical trials showed the vaccine was highly effective at generating antibodies for the H1N1 and H3N2 strains but struggled to meet the "non-inferiority" bar for the B/Victoria and B/Yamagata strains compared to existing vaccines.
The Resolution: After the public back-and-forth, the
the new flu shot for approval, aiming to resolve the technical and regulatory disagreements that had previously stalled its progress.FDA will now consider
Why This Matters
If approved, Moderna's vaccine would be the first mRNA flu shot on the market. This is seen as a crucial step toward their ultimate goal: a combination vaccine that protects against the flu, COVID-19, and potentially RSV in a single annual shot.
Would you like me to look up the latest clinical trial efficacy numbers for Moderna's combination flu/COVID vaccine?
In clinical trials, Moderna's flu vaccine (mRNA-1010) faced a specific regulatory hurdle: while it successfully induced a strong immune response against the "A" strains of the flu, it initially fell short of the required performance standards for the "B" strains.
To understand why this mattered, you have to look at the four specific targets of a "quadrivalent" (four-strain) flu shot:
1. The Targets: A vs. B Strains
Seasonal flu vaccines are designed to protect against four distinct versions of the virus that circulate in humans:
Influenza A (H1N1 and H3N2): These are the "heavy hitters." They are more common, mutate rapidly, and are responsible for most flu-related hospitalizations and all historical pandemics. mRNA-1010 performed exceptionally well here, showing results that were equal to or better than traditional shots.
Influenza B (Victoria and Yamagata): These lineages only infect humans and generally mutate more slowly. While they cause fewer hospitalizations in adults than Type A, they are a major cause of illness in children.
2. The "Non-Inferiority" Bar
In vaccine science, a new product doesn't always have to be "better" than what already exists; it often just needs to prove it is "not unacceptably worse." This is the non-inferiority bar.
The Test: Researchers compare the antibody levels (geometric mean titers) produced by the new vaccine against those produced by an already approved "gold standard" vaccine.
The Result: For the B/Victoria and B/Yamagata strains, Moderna’s early data showed that the antibody levels were lower than those produced by the traditional vaccines. They didn't miss by a lot, but they failed to meet the strict statistical threshold (the "bar") required to prove they were "just as good" as the shots people already get.
3. Why did the mRNA struggle with Type B?
This is a technical challenge scientists are still solving. mRNA vaccines work by teaching your cells to make a specific protein (the hemagglutinin or HA). For reasons not fully understood, the mRNA instructions for the Type B HA proteins didn't seem to "take" as strongly as the instructions for Type A.
Note: Since 2020, the B/Yamagata strain has actually disappeared from global circulation (likely due to COVID-19 lockdowns). As a result, the
have recently recommended moving back to trivalent (three-strain) vaccines that exclude Yamagata entirely. FDA and WHO
Current Status
Moderna has since updated its formulation. Their latest Phase 3 data (from
Would you like me to look up the specific side effects reported in these mRNA flu trials compared to traditional egg-based shots?