WU JUNCHAO
- I am WU JUNCHAO (吴俊潮). Currently, I am a first-year Ph.D. student in Computer Science at NLP2CT Lab, University of Macau, fortunately advised by Prof. Derek F. Wong. Previously, I completed my M.S. in Data Science (Computational Lingustics) at the same lab, co-advised by Prof. Derek F. Wong and Prof. Yuan Yulin. I achineved my bachelor’s degree at Beijing Normal University, Zhuhai, supervised by Prof. JiangYing and Prof. Yangjing.
- My current research interests are Natural Language Processing, Machine Translation and Trustworthy AI. Please feel free to contact me via email!
Research Interests
- Regulatable & Trustworthy AI & Safety: LLM-Generated Text Detection; Hallucination Detection & Mitigation; Internal interpretability of LLMs; Cybersecurity Ability of LLMs
- Machine Translation: Low-Resource, LLM-based MT (Document-Level; Culture; Idioms)
- Other Interests: Knowledge Graphs and Ontology, Public Opinion Analysis
News
- [2025-06-27] One Evaluation Paper accepted by CCL 2025 Evaluation Workshop as co-author: Overview of CCL25-Eval Task 4: Factivity Inference Evaluation 2025, we will update the paper in arxiv soon~
- [2025-06-13] One Evaluation Paper accepted by NLPCC 2025 as first-author, Overview of the NLPCC 2025 Shared Task 1: LLM-Generated Text Detection. We provide a comprehensive overview of the task, including the dataset, task design, evaluation results, and an in-depth analysis of the submitted solutions.
- [2025-05-16] Two paper accepted by ACL 2025 Findings as co-author. Congratulations to Shu🌳, Xinyi, and all co-authors!
- [2025-04-03] Our new preprint Understanding Aha Moments: from External Observations to Internal Mechanisms released. We reveal that “aha moments” help LRMs solve complex problems by using anthropomorphic language, adjusting uncertainty, and preventing reasoning collapse, while internally balancing self-reflection with reasoning and adapting to task difficulty.
- [2025-03-04] 🔥 We’re excited to announce a shared task at NLPCC 2025: LLM-Generated Text Detection !!! This task focuses on Chinese and offers a great opportunity to explore unique findings compared to English detection. Join us and create your own detector now! 🚀
- [2024-11-31] One paper accepted by COLING 2025 as first-author: Who Wrote This? The Key to Zero-Shot LLM-Generated Text Detection Is GECScore. We introduce a novel black-box zero-shot method for detecting LLM-generated text, called GECScore 🔍. This method is remarkably simple yet effective, based on the observation that LLMs tend to prefer grammatically correcting human-written text over LLM-generated text.
- [2024-11-28] One paper accepted by Computational Linguistics (CL) as first-author: A Survey on LLM-generated Text Detection: Necessity, Methods, and Future Directions. This survey offers a comprehensive overview of the latest advancements in LLM-generated text detection, highlighting the urgent need for more robust methods. It reviews mainstream approaches, addresses key challenges, and outlines promising future research directions. The paper serves as both a clear introduction for newcomers and a valuable resource for experts seeking updates in the field.
- [2024-09-27] One paper accepted by NeurIPS 2024 D&B Track as first-author: DetectRL: Benchmarking LLM-Generated Text Detection in Real-World Scenarios. We introduce a new benchmark, DetectRL, which covers multiple realistic scenarios, including usage of various prompts, human revisions of LLM-generated text, adversarial spelling errors, taking measures to attack detectors, etc., provide real utility to researchers on the topic and practitioners looking for consistent evaluation methods.
Publications
2025
2024
2023
2021
“The Canton Canon” Digital Library Based on Knowledge Graph - Taking the Revolutionary Archives of Canton in the Republic of China as an Example
Services
- Conference Reviewer: ACL ARR, ICML, ICLR, NeurIPS, CCL, NLPCC
- Journal Reviewer: Proc. IEEE, IEEE TIFS, ACM TALLIP, ACM TIST
- Student Volunteer: MT Summit 2023
- Teaching Assistant
- Computational Linguistics (MSc) programme (2023 Fall)
- AHGC7315 Language and Linguistics (2023 Spring)
Experience
- Alibaba Group - Research Intern (Jul. 2025 - Present)
- PRADA Lab, KAUST - Visiting Intern (Mar. 2025 - Jun. 2025)
Professional skills
- English Level: IELTS(6.5), CET-6(507)
- Programming Languages: C, Python, Java, JavaScript, SQL, Bash
- Development Frameworks: SpringBoot, React.js, Flask
- Deep Learning Tools: Scikit-learn, PyTorch, Fairseq, Transformers
- DB Tools: Neo4j, MySQL, Oracle DB
- Other Tools: WebLogic, Apache Ant, Jena, Git
Others
- 🎓 If you are also interested in natural language processing and machine translation, you can follow my lab and lab members: NLP2CT LAB, I love them.
- ✨Zhan Runzhe is my current advisor, who focus on machine translation and LLMs. He is a brilliant scientist and one of the nicest people I have ever met.
- 🌈Chen Xin is one of my best friends, who is a incoming PhD student at Nanjing University and is also committed to NLP research., he has many interesting dreams.