Sunday, August 10, 2025

courses: deep generative learning and generative AI.

 deep generative learning and generative AI. Here are some notable sources:

  • Stanford University: The course "CS236: Deep Generative Models" by Prof. Stefano Ermon provides detailed lecture videos and slides focused on the foundations, challenges, and applications of generative models in image, text, video, medicine, robotics, and more. The course website with slides is available at https://deepgenerativemodels.github.io/ and videos are on YouTube.youtube

  • Cornell University: The course "CS 6785: Deep Generative Models," taught by Prof. Vadir Kuleshov, offers an introduction to deep generative models, recent advances, algorithms, and applications including NLP and biology. The lectures are recorded and available on YouTube.youtube

  • MIT (Massachusetts Institute of Technology): The "Introduction to Deep Learning 6.S191" program covers deep learning basics along with generative AI applications in media, vision, NLP, and biology. All lecture slides, labs, and code are open-sourced and free to use, accessible at https://introtodeeplearning.com/ with lecture videos like "Generative AI for Media" by Google’s Doug Eck on YouTube.introtodeeplearningyoutube

  • Harvard University: There are presentations specifically on generative AI’s role in education, including outlines for PowerPoint slides targeting its teaching and learning impact, available in PDF form (e.g., from Harvard AI Sandbox materials).hcsra.sph.harvard

  • University of Virginia (UVA): UVA SEAS offers collections of slides on the technical foundations of generative AI with practical uses in engineering design and analysis.teaching.virginia

  • Other resources:

    • NVIDIA Deep Learning Institute has a teaching kit for generative AI with lecture slides, labs, and Jupyter notebooks focused on GPU-accelerated generative AI development.developer.nvidia

    • Stony Brook University provides teaching resources on generative AI including PowerPoint slides for educators.stonybrook

If you want ready-to-use lecture slides or full course materials, the Stanford CS236, Cornell CS6785, and MIT 6.S191 courses are among the most comprehensive and authoritative sources from major universities. Their materials are typically publicly available online for educational use.

Would you like direct access links, or specific slide decks on any of these?

  1. https://hcsra.sph.harvard.edu/sites/projects.iq.harvard.edu/files/hcsra/files/presentation_on_ai1.pdf
  2. https://www.youtube.com/watch?v=XZ0PMRWXBEU
  3. https://blog.uwgb.edu/catl/files/2023/02/Introduction-to-Generative-AI-CATL-Presentation-Slides.pdf
  4. https://www.youtube.com/watch?v=IZgvgLy1wyg
  5. https://teaching.virginia.edu/collections/uva-seas-resources-teaching-genai-use-for-engineering-design-and-analysis/272
  6. https://developer.nvidia.com/blog/nvidia-deep-learning-institute-releases-new-generative-ai-teaching-kit/
  7. https://www.sdccd.edu/docs/IIE/ProfessionalDevelopment/Presentations/10252024_AI-Demystified-Intro-to-Generative-AI.pdf
  8. https://introtodeeplearning.com
  9. https://www.stonybrook.edu/celt/teaching-resources/aibot.php
  10. https://www.youtube.com/watch?v=P7Hkh2zOGQ0

Thursday, August 7, 2025

medical AI



Artificial Intelligence in Medical Education: The 2025 IACAI Vision and Integration Frameworks

 

https://www.medbiq.org/initiatives/international-advisory-committee-artificial-intelligence


Sunday, August 3, 2025

asr whisper wahab

 test run old slurm job and it worked. 

git clone to wahahb


hqin@wahab-01 fairASR25]$ pwd
/home/hqin/github/fairASR25





speech-to-text corpora, accent

 speech-to-text corpora—one of which is Meta FAIR’s fairness-oriented dataset:

  • LibriSpeech ASR Corpus
    A corpus of roughly 1,000 hours of 16 kHz read English speech, derived from LibriVox audiobooks, carefully segmented and aligned. Released under a CC BY 4.0 license. (openslr.org)

  • Multilingual LibriSpeech (MLS)
    A large-scale ASR dataset by Facebook AI Research (Meta), comprising ∼50,000 hours of public-domain audiobooks across eight languages (English, German, Dutch, French, Spanish, Italian, Portuguese, Polish). (Meta AI, voxforge.org)

  • Mozilla Common Voice
    A crowdsourced, multilingual speech corpus with millions of volunteer-recorded, validated sentences and transcriptions, released under CC0 (public domain). (Wikipedia)

  • TED-LIUM v3
    An English ASR corpus of 452 hours of TED talk recordings with aligned transcripts, freely available for research. (openslr.org)

  • VoxForge
    A community-collected GPL-licensed speech corpus in multiple languages, built to support open-source ASR engines (e.g., CMU Sphinx, Julius). (voxforge.org)

  • Fair-Speech Dataset (Meta FAIR)
    A fairness-oriented evaluation set containing 26,471 utterances from 593 U.S. speakers, designed to benchmark bias and robustness in speech recognition. (Meta AI, Meta AI)

  • GigaSpeech
    A multi-domain English ASR corpus featuring 10,000 hours of high-quality transcribed audio (plus 40,000 hours of additional audio for semi-/unsupervised research).

  • VoxPopuli
    Contains over 1 million hours of unlabeled multilingual speech and 1.8 k hours of transcribed speeches in 16 languages (with aligned interpretation pairs), for representation learning and semi-supervised ASR. (arxiv.org)


Here are several publicly available speech-to-text corpora that include regional and non-native accents—many of which you can filter or mine for Southern Chinese (e.g., Cantonese-influenced) accent patterns (such as /s/ vs /ʃ/ or –ing vs –in):

  • Speech Accent Archive

    A growing, global collection of ~2,500 English recordings of the same Harvard paragraph, each with narrow phonetic transcription and speaker metadata (including L1 and region). You can browse by “Chinese” and then drill down to Cantonese vs. other dialect regions. (ResearchGate, accent.gmu.edu)

  • L2-ARCTIC

    A corpus of non-native English speech from ten Mandarin (plus Hindi, Korean, Spanish, Arabic) speakers reading CMU ARCTIC prompts. It includes orthographic transcripts, forced-aligned phonetic annotations, and expert mispronunciation tags. (psi.engr.tamu.edu)

  • CSLU Foreign-Accented English (Release 1.2)

    ~4,925 telephone-quality utterances by speakers of various L1s (including Chinese), with transcript, speaker background, and perceptual accent ratings. (borealisdata.ca)

  • speechocean762

    5,000 English utterances from 250 non-native speakers (half children), each annotated at the sentence, word, and phoneme level. Designed for pronunciation assessment, freely downloadable via OpenSLR. (arXiv)

  • ShefCE: Cantonese-English Bilingual Corpus

    Audio & transcripts from 31 Hong Kong L2 English learners reading parallel Cantonese and English texts—ideal for studying Cantonese-influenced English phonetics. (orda.shef.ac.uk)

  • Sell-Corpus: Multi-Accented Chinese English Speech

    First open-source English speech corpus covering seven major Chinese dialect regions (including Southern dialects), with recordings & transcripts for accent variation research. (sigport.org)

  • Mozilla Common Voice

    Crowdsourced, multilingual speech data (CC0) with per-speaker accent tags—you can filter English recordings by “Chinese (Hong Kong)” or “Chinese (Mainland)” to get regional accent samples. (Wikipedia)

  • ICNALE Spoken Monologues

    4,400 60-second monologues (~73 h) by 1,100 Asian learners (incl. Mainland China, Hong Kong, Taiwan), with transcripts—useful for comparing Southern vs. Northern Chinese L1 influence on English pronunciation. (language.sakura.ne.jp, language.sakura.ne.jp)

  • International Dialects of English Archive (IDEA)

    Free archive of scripted & unscripted English dialect samples worldwide. Browse the “Asia → China” section to find Cantonese- and Mandarin-accented speakers, all with transcripts. (Wikipedia)

Each of these datasets provides aligned audio and text (and often phonetic detail) that you can mine to analyze pronunciation patterns—like the s/ʃ or –ing/–in contrasts—among Southern Chinese speakers learning or using English.

Real-time Out-of-distribution Detection in Learning-Enabled Cyber-Physical Systems

 Real-time Out-of-distribution Detection in Learning-Enabled Cyber-Physical Systems


Here are the details:

A Generative Security Application Engineering Curriculum

 

https://arxiv.org/html/2501.10900v1

A Generative Security Application Engineering Curriculum


https://youtube.com/playlist?list=PLdCTMpWhMq8KuRHaXHuf2U53hM1NkuiCe&si=TqgEulf57ZnUOXON



o All material has been made publicly available (https://codelabs.cs.pdx.eduhttps://github.com/wu4f/cs410g-srchttps://github.com/wu4f/cyberpdx-crypto)

o YouTube playlist for Generative Security Applications course is available at (https://youtube.com/playlist?list=PLdCTMpWhMq8KuRHaXHuf2U53hM1NkuiCe&si=S5dXzkslHPX5N13w)

o CyberPDX module available at https://crypto.cyberpdx.org


Monday, July 28, 2025

funding acknowledgement fall 2025

 for AI works

HQ thanks USA NSF 2525493 and  2200138, a catalyst award from the USA National Academy of Medicine,  and internal support of the Old Dominion University


Friday, July 18, 2025

ODU CS Candidacy exam

CS  Candidacy Exam.

 

The catalog description of the exam is found at

https://catalog.odu.edu/graduate/sciences/computer-science/computer-science-phd/#additionalrequirementstext

 

He’s doing Option 1, which is summary of papers relevant to his dissertation research topic.

 

The guidelines for the length of the document are just guidelines, it can be longer.  And same for the length of the presentation. It can be as long as the committee would like.

 


Sunday, July 13, 2025

goodnight moon audo data and labels?


Children’s Speech Recognition Challenge

https://kidsasr.drivendata.org/

https://github.com/hongqin/goodnight-moon


https://www.drivendata.org/competitions/298/literacy-screening/


Saturday, July 12, 2025

shap summry plot by chatGPT

 chatGPT repeatly make mistake for shap summary plot. 


for different class label, the index should be the 3rd position:  shap_vals[:,:,idx],


Wednesday, July 2, 2025

national family survey of pregancy

  national family survey of pregancy

https://www.cdc.gov/nchs/nsfg/index.htm


todo: request to restrickted access variables. 


Saturday, June 28, 2025

Fall 2025 schedule

August 23 - Dec 12, 2025, Thursday 6p - 8:40pm.  

Scheduled Meeting Times
TypeTimeDaysWhereDate RangeSchedule TypeInstructors
Scheduled In-Class Meetings6:00 pm - 8:40 pmRENGINEERING & COMP SCI BLDG 2120Aug 23, 2025 - Dec 12, 2025LECTURE

Tuesday, June 24, 2025

ZIP Code RUCA Approximation,

 

https://depts.washington.edu/uwruca/ruca-approx.php?utm_source=chatgpt.com


All of Us survey, data codebooks

 All of Us survey, data codebooks

https://docs.google.com/spreadsheets/d/1pODkE2bFN-kmVtYp89rtrJg7oXck4Fsex58237x47mA/edit?usp=sharing


Friday, June 20, 2025

A model for the assembly map of bordism-invariant functors

 The paper "A model for the assembly map of bordism-invariant functors" by Levin, Nocera, and Saunier (2025) develops advanced categorical frameworks for algebraic topology, particularly through oplax colimits of stable/hermitian/Poincaré categories and bordism-invariant functors123. While not directly addressing machine learning (ML) or large language models (LLMs), its contributions could indirectly influence these fields through three key pathways:

1. Enhanced Categorical Frameworks for ML

The paper's formalization of oplax colimits and Poincaré-Verdier localizing invariants13 provides new mathematical tools for structuring compositional systems. This could advance:

  • Model Architecture Design: By abstracting relationships between components (e.g., neural network layers) as bordism-invariant functors, enabling more rigorous analysis of model behavior under transformations5.

  • Geometric Deep Learning: Topological invariants and assembly maps could refine methods for learning on non-Euclidean data (e.g., graphs, manifolds) by encoding persistence of features under deformations5.

2. Invariance Learning and Equivalence

The bordism-invariance concept—where structures remain unchanged under continuous deformations—offers a mathematical foundation for invariance principles in ML:

  • Data Augmentation: Formalizing "bordism equivalence" could guide the design of augmentation strategies that preserve semantic content (e.g., image rotations as "topological bordisms")5.

  • Robust Feature Extraction: Kernels of Verdier projections13 might model noise subspaces to exclude during feature learning, improving adversarial robustness.

3. LLMs for Structured Reasoning

The paper’s explicit decomposition of complex functors (e.g., Shaneson splittings with twists13) parallels challenges in LLM-based reasoning:

  • Program Invariant Prediction: LLMs that infer program invariants6 could adopt categorical decompositions to handle twisted or hierarchical constraints (e.g., loop invariants in code).

  • Categorical Data Embeddings: LLM-generated numerical representations of categorical data4 might leverage bordism-invariance to ensure embeddings respect equivalence classes (e.g., "color" as a deformation-invariant attribute).

Limitations and Future Directions

The work is highly theoretical, with no direct ML/LLM applications in the paper. Bridging this gap requires:

  • Translating topological bordisms into data-augmentation pipelines.

  • Implementing Poincaré-Verdier invariants as regularization terms in loss functions.

  • Extending LLM-based invariant predictors6 to handle categorical assembly maps.

While speculative, these connections highlight how advanced category theory could enrich ML’s theoretical foundations and LLMs’ reasoning capabilities.

  1. https://arxiv.org/abs/2506.05238
  2. https://arxiv.org/pdf/2506.05238.pdf
  3. https://www.arxiv.org/pdf/2506.05238.pdf
  4. https://pubmed.ncbi.nlm.nih.gov/39348252/
  5. https://www.aimodels.fyi/papers/arxiv/category-theoretical-topos-theoretical-frameworks-machine-learning
  6. https://openreview.net/pdf?id=mXv2aVqUGG
  7. https://x.com/CTpreprintBot
  8. https://keik.org/profile/mathat-bot.bsky.social
  9. https://www.alphaxiv.org/abs/2506.05238
  10. https://publications.mfo.de/bitstream/handle/mfo/4263/OWR_2024_47.pdf?sequence=1&isAllowed=y
  11. https://x.com/CTpreprintBot/status/1930943445977518380
  12. https://www.themoonlight.io/en/review/a-model-for-the-assembly-map-of-bordism-invariant-functors
  13. https://library.slmath.org/books/Book69/files/wholebook.pdf
  14. https://www.reed.edu/math-stats/thesis.html
  15. https://math.mit.edu/events/talbot/2020/syllabus2020.pdf
  16. https://webhomes.maths.ed.ac.uk/~v1ranick/papers/quinnass.pdf
  17. https://msp.org/agt/2009/9-4/agt-v9-n4-p16-s.pdf
  18. https://webhomes.maths.ed.ac.uk/~v1ranick/papers/owsem.pdf

Friday, June 6, 2025

Mendely preprint error

 In Mendely, if an article has unspecified type, it often list it as "preprint". 

To fix this, just change the document type to 'jounral article' or 'conference proceedings' or other appropriate type. 

Thursday, May 29, 2025

Beyond Attention: Toward Machines with Intrinsic Higher Mental States

 

Beyond Attention: Toward Machines with Intrinsic Higher Mental States

https://techxplore.com/news/2025-05-architecture-emulates-higher-human-mental.html#google_vignette


Wednesday, May 28, 2025

USDA Rural-Urban Commuting Area Codes

 USDA

Rural-Urban Commuting Area Codes

https://www.ers.usda.gov/data-products/rural-urban-commuting-area-codes?utm_source=chatgpt.com


three digit 360 zipcode in Alabama

 Based on our discussion, I have looked further into the ‘360’ zipcode region, and found it contain 14 counties below:

 

counties = [    "Autauga County", "Barbour County", "Bullock County", "Butler County",

    "Chilton County", "Coosa County", "Covington County", "Crenshaw County",

    "Elmore County", "Lowndes County", "Macon County", "Montgomery County",

    "Pike County", "Tallapoosa County"]

 

Amon them, there are “Barbour”, “Bullock”, and “Macon” counties.

 

So, not sure how useful the three-digit zip code of ‘360’ might be relevant to TU MCH project. 



pastbin

 https://pastebin.com/uBcFUXCA

  1. import pandas as pd
  2. dataset = %env WORKSPACE_CDR
  3. query = """
  4. SELECT
  5. p.person_id AS person_id,
  6. c.concept_name AS state_name
  7. FROM `{dataset}.person` AS p
  8. LEFT JOIN `{dataset}.concept` AS c ON p.state_of_residence_concept_id = c.concept_id
  9. WHERE c.concept_name LIKE 'PII State: %'
  10. """
  11.  
  12. state_df = pd.read_gbq(query.format(dataset = dataset))
  13.  
  14. state_df['state_of_residence'] = state_df['state_name'].str.replace('PII State: ', '')
  15.  
  16. state_df.head()
  17. state_df.shape
  18.  
  19. homeless_state_df=state_df[state_df['person_id'].isin(homeless_respiratory_status['person_id'])]
  20. homeless_state_df['state_of_residence'].value_counts()


All of us research platform, data explore for rural health research.

 All of us research platform, data explore for rural health research.

Zip3, observational table


Workspaces > Beginner Intro to AoU Data and the Workbench > Analysis >

In ‘test-state_data_v060523’, found sample on state, county, and zip3.

 

 

Monday, May 26, 2025

ODU CS and DSC courses taught by Hong Qin

 

https://catalog.odu.edu/courses/cs/#graduatecoursestext

https://catalog.odu.edu/courses/dasc/


CS 781  AI for Health Sciences  (3 Credit Hours)  

This course explores the application of AI in health sciences, focusing on machine learning, NLP, computer vision, generative AI techniques for diagnostics, treatment planning, patient monitoring, and biomedical research. It covers precision medicine, ethical AI, and the integration of AI into practice. Students will gain a deep understanding and practical skills to develop innovative AI solutions that address real-world challenges in health sciences.

Prerequisites: Prior programming experience  
CS 782  Generative AI  (3 Credit Hours)  

This course provides a deep dive into the foundations and current advancements in generative AI. It covers key concepts such as transformer models, GANs, VAEs, LLMs, and their applications across various fields, emphasizing both theory and hands-on learning, including ethical considerations such as fairness and bias mitigation. Students will develop a comprehensive understanding of generative AI and gain practical experience.

Prerequisites: Prior programming experience  

CS 881  AI for Health Sciences  (3 Credit Hours)  

This course explores the application of AI in health sciences, focusing on machine learning, NLP, computer vision, generative AI techniques for diagnostics, treatment planning, patient monitoring, and biomedical research. It covers precision medicine, ethical AI, and the integration of AI into practice. Students will gain a deep understanding and practical skills to develop innovative AI solutions that address real-world challenges in health sciences.

Prerequisites: Prior programming experience  
CS 882  Generative AI  (3 Credit Hours)  

This course provides a deep dive into the foundations and current advancements in generative AI. It covers key concepts such as transformer models, GANs, VAEs, LLMs, and their applications across various fields, emphasizing both theory and hands-on learning, including ethical considerations such as fairness and bias mitigation. Students will develop a comprehensive understanding of generative AI and gain practical experience.

Prerequisites: Prior programming experience  

DASC 781  AI for Health Sciences  (3 Credit Hours)  

This course explores the application of AI in health sciences, focusing on machine learning, NLP, computer vision, generative AI techniques for diagnostics, treatment planning, patient monitoring, and biomedical research. It covers precision medicine, ethical AI, and the integration of AI into practice. Students will gain a deep understanding and practical skills to develop innovative AI solutions that address real-world challenges in health sciences.

Prerequisites: Prior programming experience  
DASC 782  Generative AI  (3 Credit Hours)  

This course provides a deep dive into the foundations and current advancements in generative AI. It covers key concepts such as transformer models, GANs, VAEs, LLMs, and their applications across various fields, emphasizing both theory and hands-on learning, including ethical considerations such as fairness and bias mitigation. Students will develop a comprehensive understanding of generative AI and gain practical experience.

Prerequisites: Prior programming experience  

DASC 881  AI for Health Sciences  (3 Credit Hours)  

This course explores the application of AI in health sciences, focusing on machine learning, NLP, computer vision, generative AI techniques for diagnostics, treatment planning, patient monitoring, and biomedical research. It covers precision medicine, ethical AI, and the integration of AI into practice. Students will gain a deep understanding and practical skills to develop innovative AI solutions that address real-world challenges in health sciences.

Prerequisites: Prior programming experience  
DASC 882  Generative AI  (3 Credit Hours)  

This course provides a deep dive into the foundations and current advancements in generative AI. It covers key concepts such as transformer models, GANs, VAEs, LLMs, and their applications across various fields, emphasizing both theory and hands-on learning, including ethical considerations such as fairness and bias mitigation. Students will develop a comprehensive understanding of generative AI and gain practical experience.

Prerequisites: Prior programming experience