https://arxiv.org/abs/2406.06608
This site is to serve as my note-book and to effectively communicate with my students and collaborators. Every now and then, a blog may be of interest to other researchers or teachers. Views in this blog are my own. All rights of research results and findings on this blog are reserved. See also http://youtube.com/c/hongqin @hongqin
Mermaid Chart, Flowcharts - Basic Syntax
https://mermaid.js.org/syntax/flowchart.html
**Summary of AI and Physics Connection:**
Yann LeCun and Derrick Hodge highlight the deep interconnections between AI and physics. LeCun notes that the derivation of backpropagation from classical mechanics and optimal control was detailed in his 1987 PhD thesis and 1988 paper. Hodge elaborates on how various AI methods correspond to physical principles:
1. **Variational Bayesian Inference** relates to thermodynamics.
2. **Bayesian Ensembling** is analogous to path integrals.
3. **Backpropagation** can be derived via the Lagrangian saddle point method.
4. **Convergence in non-convex objectives** mirrors spontaneous symmetry breaking.
5. **ConvNet pooling** parallels renormalization group theory and MERA tensor networks.
6. **Lorentz-equivariant ConvNets**.
These connections suggest that AI and physics share foundational computational mechanisms, leading to unified theories, new research frontiers, and deeper understanding of both fields.
https://hanlab.mit.edu/courses/2023-fall-65940
This course focuses on efficient machine learning and systems. This is a crucial area as deep neural networks demand extraordinary levels of computation, hindering its deployment on everyday devices and burdening the cloud infrastructure. This course introduces efficient AI computing techniques that enable powerful deep learning applications on resource-constrained devices. Topics include model compression, pruning, quantization, neural architecture search, distributed training, data/model parallelism, gradient compression, and on-device fine-tuning. It also introduces application-specific acceleration techniques for large language models and diffusion models. Students will get hands-on experience implementing model compression techniques and deploying large language models (Llama2-7B) on a laptop.
Tuesday/Thursday 3:35-5:00pm Eastern Time
Thursday 5:00-6:00 pm Eastern Time, 38-344 Meeting Room
Final project: reports, slides and demo videos
Final report and course evaluation due
Mid-term survey: https://forms.gle/xMgCohDLX73cd4af9
Lab 5 is out.
Lab 1 due (extended to Sep 30 at 11:59 p.m)
Dec 14: Project report and course evaluation due