🚀 Updates: As an incoming assistant professor at King’s College London (fall 2025), I am actively recruiting passionate PhDs and Master’s students interested in interpretable and robust reasoning with language models.

Hi everyone! I’m a PostDoc (01.2024-) at King’s College London, NLP Group, School of Informatics. I passed my PhD viva with no corrections after a great time at the University of Warwick (10.2020-01.2024), advised by Prof. Yulan He and Dr. Lin Gui. I finished my M.S. at Peking University(09.2017-07.2020) and my B.E. at Beihang University(09.2013-06.2017).

During Ph.D., I started my Causality Journey(10.2022-02.2023) in visiting Prof. Kun Zhang affiliated with the Causal Learning and Reasoning Group@CMU. Before Ph.D., I started my NLP journey(07.2019-10.2019) in visiting Prof. Wenjie Li affiliated NLP Group @PolyU Hong Kong.

I’ve been incredibly lucky to have a number of amazing collaborators and mentors across KCL and a range of other institutions, including Carnegie Mellon University, MIT, Peking University, MBZUAI, Hong Kong Polytechnic University, University College London,University of Warwick. None of the research so far—would be possible without their kind help and support.

🔍 Research Summary

My research interests lie in the intersection of Machine Learning and Natural Language Processing, i.e., incorporating fundamental representation learning to enhance the interpretability and reliability of different NLP models, with 10+ (co-)first-authored papers published at top-tier venues:

  • Mechanistic interpretability (neuron-level) in language models and multi-modal models [EMNLP24,NeurIPS24-RBMF]; self-explainable models with a conceptualised layer linking the input and decision layer [Computational Linguistics 22,TKDE24].
  • Empirical and principled methods to enhance model robustness over various test inputs, e.g., position bias [ACL24-findings,ACL21-oral], distribution shifts [NeurIPS23] and representation inefficiency in transformer-based models [EMNLP24,EACL23-findings,UAI22-spotlight].
  • Understanding and enhancing LM’s reasoning capabilities via injecting external knowledge[ACL21-oral][ACL25], weak supervision [EMNLP24], applying a self-refinement mechanism for factual knowledge reasoning [ACL24]. More recently focus in the two directions:
    • Science literature understanding, such as code generation for scientific paper replication on our own SciReplicate-Bench and novelty assessment.
    • Reasoning in latent space, such as a position paper at ICML25 about meta-reasoning, CODI in implicit CoT, in retrieved-based QA [ACL25], navigating search in latent space [ICML25spotlight] and sparse feature for preference optimization [ICML25].

🔥 News

07.2025: I go to Vienna, Austria🇦🇹 to present our faithful rationale generation and RAG papers accepted at ACL 2025. 🏞️ Excited to escape city life and explore Gosau & Hallstatt!
07.2025: I go to Vancouver, Canada🇨🇦 to present our LLM reasoning papers accepted at ICML 2025. 🍜 Can’t wait to revisit my favorite Chinese restaurant there!
07.2025: 🔥 1*paper, SciReplicate-Bench, accepted by COLM25, a benchmark for paper replication via code generation.
05.2025" 🔥 2*papers accepted by ACL25 main conference, including a co-first author paper in faithful rationale generation during LLM inference.
05.2025: 🔥 3*papers accepted by ICML25, including a first-author paper about meta-reasoning in the Position paper track.
11.2024: I go to Miami☀️🌊🍹🏝, US for EMNLP24 to present our accepted papers and connect with like-minded researchers👩‍💻👨‍💻.
10.2024: 1*first-author paper about monosemantic neuron in multi-modal model is accepted by Neurips-RBMF workshop.
09.2024: 3*papers (monomsemantic neurons, oral survey in ICL, weak2strong event extraction) are accepted by EMNLP24 Main Conference. 🎉
08.2024: I go to Bangkok, Thailand🇹🇭 for ACL24. ✈️
06.2024: 2* paper accepted by ACL24, including 1 1st author in main and 1 in findings.
04.2024: I pass the PhD viva with no corrections 🎓.
01.2024: I become a PostDoc👩‍🏫 at King's College London, NLP Group.
01.2024: I finish my PhD thesis (draft) on the same day of my birthday.
01.2024: 1*first-author paper is finally accepted by TKDE.
12.2023: I go to New Orleans🎷, US to present our Neurips paper.
07.2023: I go to Hawaii🌴, US to present our ICML-workshop paper.
07.2023: 1*first-author paper is accepted by Neurips (my Neurips paper).
02.2023: I go back to the UK from Abu Dhabi, UAE🇦🇪, finish my Machine Learning trip in MBZUAI.
02.2023: I attend the EMNLP23 held in Abu Dhabi, to present our Computational Linguistics paper.
01.2023: 1*paper is accepted by EACL23🇭🇷-findings (first time as a mentor for a master's student).
12.2022: Lionel Messi leads Argentina to win the ⚽️World Cup championship.
10.2022: I start to be a funded visiting student in Machine Learning, Department at MBZUAI🏫, Abu Dhabi, UAE, advised by Prof. Kun Zhang.
08.2022: I go to Eindhoven, Netherlands🇳🇱 to present our UAI paper.
05.2022: 1*first-author paper is accepted by UAI23-spotlight (🥳my first ML paper)
05.2021: The first time to get paper accepted! 1*first-author paper is accepted by ACL21 🌟Oral. A super encouragement in my early PhD career.
10.2020: I start my PhD📚 journey at University of Warwick, UK🇬🇧.

👩‍🏫 Professional Service

  • Co-Chair of the Asian Chapter of the Association for Computational Linguistics (Student Research Workshop) 2022
  • Organiser of Event - PreTrain 2025: Spotlight on ACL/ICML/ICLR at King’s College London
  • Area Chair: ACL25, EMNLP25
  • Reviewers:
    • NLP: AACL, NAACL, EACL, EMNLP, ACL, COLM25
    • AI/ML: UAI, AISTATS, NEURIPS, ICLR, ICML, AAAI26
    • Journal: NeuroComputing, TOIS, TMLR, Transactions on Big Data, Transactions on Artificial Intelligence.

💬 Invited Talks

  • University of Southampton, Computer Science , 04/2025. LLM Reasoning with internal and controllable signals
  • University of Surrey, Computer Science , 04/2025. Controllable generation framework for LLM reasoning
  • University of Bath, Computer Science , 04/2025. Disentanglement for interpretability and controllability
  • University of Birmingham, Computer Science , 12/2024. Understanding and improving LLM reasoning via representation learning
  • Huawei, LLM research team in Singapore (Remote) . Controllable generation and Causality in LLMs.
  • Amazon, Artificial Generative Intelligence team in London (Remote) . Robust and Interpretable NLP via Representation Learning and Applications in LLMs.
  • Fudan University, NLP Group , 07/2024. Representation Learning and Mechanistic Interpretability
  • UC San Diego, NLP Group , 02/2024. Robust and Interpretable NLP via representation learning and Path Ahead
  • Yale University, NLP Group 01/2024. Robust and Interpretable NLP via representation learning and Path Ahead
  • Turing AI Fellowship Event , London, 03/2023, Distinguishability Calibration to In-Context Learning
  • UKRI Fellows Workshop , University of Edinburgh, 04/2022. Interpreting Long Documents and Recommendation Systems via Latent Variable Models

💬 Mentee

  • PhD students .
    • 2*scientific literature understanding (accepted by COLM25 and in submission).
    • 2* explainable AI (language and multimodal model). (accepted by ACL25 and in submission)
    • 3* robust LLM reasoning (accepted by ACL24-findings/ACL25 and in submission).
    • 1* event extraction (accepted by EMNLP24)
  • Master students .
    • 1*rank efficiency in transformer representation (accepted by EACL23-findings).
    • 5*explainable AI (cognition perspective).

📝 Publications

(* indicates equal contribution)

Position: LLMs Need a Bayesian Meta-Reasoning Framework for More Robust and Generalizable Reasoning
H. Yan, L. Zhang, J. Li, Z. S, Y. He
ICML25, Position Track | Paper
Application

Drift: Enhancing LLM Faithfulness in Rationale Generation via Dual-Reward Probabilistic Inference
J. Li, H. Yan, Y. He
ACL25, Main | Paper
Application Interpretability


Navigating Solution Spaces in Large Language Models through Controlled Embedding Exploration
Q. Zhu, R. Zhao. H. Yan, Y. He, Y. Chen, L. Gui
ICML25, Spotlight | Paper
Representation

Direct preference optimization using sparse feature-level constraints
Q. Yin, C. Leong, H. Zhang, M. Zhu, H. Yan, Q. Zhang, Y. He, W. Li, J. Wang, Y. Zhang, L. Yang
ICML25 | Paper
Interpretability Representation

Encourage or Inhibit Monosemanticity? Revisit Monosemanticity from a Feature Decorrelation Perspective
H. Yan, Y. Xiang, G Chen, Y. Wang, L. Gui, Y. He
EMNLP24, main | Paper
Interpretability Representation

Weak Reward Model Transforms Generative Models into Robust Causal Event Extraction Systems
I. Silva, H. Yan, L. Gui, Y. He
EMNLP24, main | Paper
Causality application

The Mystery and Fascination of LLMs: A Comprehensive Survey on the Interpretation and Analysis of Emergent Abilities
Y. Zhou, J. Li, Y.Xiang, H.Yan, L. Gui, Y. He
EMNLP24, main | Paper
Interpretability

Mirror: A Multiple-perspective Self-Reflection Method for Knowledge-rich Reasoning
H. Yan, Q. Zhu, X. Wang, L. Gui, Y. He
ACL24, main | Paper
application

Addressing Order Sensitivity of In-Context Demonstration Examples in Causal Language Models.
Y. Xiang, H. Yan, L. Gui, Y. He
ACL24, findings | Paper
Representation

Counterfactual Generation with Identifiability Guarantee
H. Yan, L. Kong, L. Gui, Y. Chi, Eric. Xing, Y. He, K. Zhang
Neurips23, main | Paper
Causality Representation application

Explainable Recommender with Geometric Information Bottleneck
H. Yan, L. Gui, M. Wang, K. Zhang and Y. He
TKDE | Paper
Interpretability application

Hierarchical Interpretation of Neural Text Classification
H. Yan, L. Gui and Y. He
Computational Linguistics, Present at EMNLP23 | Paper
Interpretability application

Addressing Token Uniformity in Transformers via Singular Value Transformation
H. Yan, L. Gui, W. Li and Y. He
UAI22, spotlight | Paper
Representation

Distinguishability Calibration to In-Context Learning
H. Li, H. Yan, L. Gui, W. Li and Y. He
EACL23, findings | Paper
Representation

A Knowledge-Aware Graph Model for Emotion Cause Extraction
H. Yan, L. Gui and Y. He
ACL21, Oral | Paper
Causality application

📝 Notes

o1-technical report (notes for video)
Machine Unlearning via CausalLens and in NLP tasks
Reading List For Large Language Model
Identifiability101 in Causality (3rd PhD)
Induction Head_ contribute to In-context Learning (3rd PhD)
Recommendation with Causality (2nd PhD)
Causality101 (Feb 2022, 2nd PhD)
Explaining Neural Networks (Oct 2020 1st PhD)