cv

You can download a PDF version. Although the cv is more updated on this website. I intend to disclose the whole journey of mine here, however it may look winding and strange. I want to build, one honest work a time, and then one item in this page. In the end it's life worth living.

Basics

Name Zhonghao He (何忠豪)
Label AI Alignment and Human-AI Interaction Researcher
Affiliation Leverhulme Center for Future Intelligence, University of Cambridge
Summary I am creating AI-assistants for human moral progress and preventing LLM-induced lock-in. My work is the only necessary form of my existence. I build, therefore I exist.

Education

  • 2022.09 - PRESENT

    Cambridge, UK

    Master
    University of Cambridge
    AI Ethics and Society
    • Machine Learning Alignment
    • AI Ethics
    • AI Governance
    • CS230 Deep Learning
    • ML Safety
    • Discrete Mathematics
    • CS234 Reinforcement Learning
    • CS109 Probability for Computer Scientists
    • Mechanistic Interpretability
    • Algorithms and Data Structure
  • 2019.06 - 2019.09

    Palo Alto, USA

    Summer Student
    Stanford University
    Cognitive Science & Philosophy
    • Mathematics Foundation of Computing
    • Minds and Machines
    • Introduction to Neuroscience
  • 2014.08 - 2019.06

    Shantou, China

    Bachelor of Arts
    Shantou University
    English & Global Studies
    • Machine Learning and relevant maths
    • Research Methodology
    • Linguistics

Projects

  • 2025.01 - Present
    Algorithms that Assist Truth Seeking and Prevent Lock-in
    We propose an alternative to RLHF that uses " helping humans to seek truth as training objective " and human opinion change data as ground truth. By doing so we aim to remove feedback loop-incurred lock-in from its root and revolves alignment evaluation around " LLM-assited human performanceWe ".
    • Position: Project Co-founder
    • Collaborators: Prof Andreea Bobu (MIT), Tianyi (Alex) Qiu (CHAI, Berkley & PKU
  • 2024.10 - 2025.01
    Position: AI Systematically Rewires the Flow of Ideas
    We propose "AI influence " as a field of studies on AI's impact on epistemics and morality. Early algorithmic interventions and data curation practices will remove mechanisms (such as dual influence) leading to systematic negative impact on human knowledge and values.
    • Position: Project Co-founder
    • Collaborators: Prof Max Kleiman-Weiner (UW), Tianyi (Alex) Qiu (CHAI, Berkley & PKU, Prof Atoosa Kasirzadeh, Prof John P Wihbey, Dr Moshe Glickman, Tao Lin.
  • 2024.10 - 2025.01
    The Lock-in Hypothesis: Stagnation by Algorithms
    We are concerned with the problems of LLM-incured value lock-in and knowledge collapse (as probable as model collapse since increasingly our discourses are mediated by AI systems and iterative training becomes more prevalence), with the consequence being more destructive. Our team is working on toy model demonstration of the mechanisms of LLM-incurred value lock-in, as long as establishing real-world evidence of lock-in from human-AI interactions.
    • Position: Project Co-founder
    • Collaborators: Prof Max Kleiman-Weiner (UW), Tianyi (Alex) Qiu (CHAI, Berkley & PKU
  • 2023.12 - 2025.02
    Multilevel analytical framework for interpretability
    Research on cognitive science and neuroscience to address interpretability challenges in ML.
    • Position: Project Lead
    • Goal: Publication in Transactional Machine Learning Research
    • Senior authors: Grace W. Lindsay(NYU), Prof Anna Ivanova (GeorgiaTech)
  • 2023.07 - 2023.10
    Comprehensive Survey on AI Alignment
    Survey paper on alignment research for newcomers.
    • Focus: Interpretability challenges in ML
    • Collaborators: Yaodong Yang, Jiaming Ji, Tianyi Qiu
  • 2022.12 - 2023.03
    Harms from agentic algorithmic systems
    Research on safety and harms from agentic systems in AI.
    • Highlight: Published paper cited by GPT-4 and high-profile AI safety reports
  • 2021.06 - 2022.02
    Stanford Existential Risks Initiative (SERI)
    Research on China's AI governance approach.
    • Position: Research Fellow

Awards

Skills

Mathematics
Calculus
Information Theory
Linear Algebra
Formal Methods
ML Engineering
Machine Learning
Deep Learning
Data Analysis
ML Safety
Git
Experimental
Data Visualization
Mechanistic Interpretability
Simulation
Programming
Python (advanced)
Pytorch
R (intermediary)
web stuff (intermediary)
Matlab (basic)
C/C++ (basic)

Languages

English
Close to Native
Chinese
Native
French
Basic

Interests

Physical Activities
Rowing
Hiking
Other Interests
Debate
Greek Literature