Ryuto Koike
Email: ryuto.koike [at] nlp.c.titech.ac.jp
|
|
About me
Hello! I am a final-year PhD student (est. March 2026) at the Institute of Science Tokyo, advised by Prof. Naoaki Okazaki.
My research interests lie in the field of Responsible AI and NLP, including model
robustness, the
creation of interpretable
NLP systems, ensuring the safe application of LLMs, and rigorous evaluation grounded in
practical
scenarios. My work has
been published at leading conferences, including ACL, EMNLP, and AAAI. I have led multiple research
projects with Prof.
Chris
Callison-Burch at UPenn and Prof. Preslav
Nakov at MBZUAI.
Particularly, my research on
detecting AI-generated text
has gathered notable attention with coverage in the Nikkei, received conference awards, and
collectively has over 100 citations.
Outside of academia, I am involved as a research advisor for a
startup on multi-lingual text generation.
I'm actively looking for research positions in academia or industry starting
in April 2026. Please feel free to contact me if you are interested in my research.
|
Selected Works (β : undergraduate/master's mentee.)
[
Google Scholar
]
|
|
Robustness
OUTFOX: LLM-Generated Essay
Detection Through In-Context Learning with Adversarially
Generated Examples
Ryuto Koike, Masahiro
Kaneko, Naoaki Okazaki
AAAI 2024
TL;DR - We propose OUTFOX, a novel framework that improves the robustness of LLM text detectors by
allowing both the detector and the attacker to adversarially learn from each other's output through
in-context learning. In this framework, the attacker uses the detector's prediction labels as examples
for in-context learning and adversarially generates essays that are harder to detect, while the
detector uses the adversarially generated essays as examples for in-context learning to learn to
detect essays from a strong attacker.
|
|
Robustness Evaluation
How You Prompt Matters! Even Task-Oriented Constraints in Instructions Affect
LLM-Generated Text Detection
Ryuto Koike, Masahiro
Kaneko, Naoaki Okazaki
EMNLP 2024 (Findings)
TL;DR - We reveal that current detectors are brittle to instruction variation in text
generation.
Specifically, even task-oriented constraints -- constraints that would naturally be included in
an
instruction and are not related to detection-evasion -- cause existing powerful detectors to have a
large variance in detection performance.
We raise awareness of the need to ensure prompt diversity when creating a detection benchmark and
open-source our constrained dataset.
|
|
Interpretability Reliability
ExaGPT: Example-Based Machine-Generated Text Detection for Human Interpretability
Ryuto Koike, Masahiro
Kaneko, Ayana Niwa, Preslav Nakov, Naoaki Okazaki
NeurIPS (In submission)
TL;DR - We propose ExaGPT, an interpretable AI text detector that identifies a text by checking
whether it shares more similar spans with human-written vs. machine-generated texts from a datastore
and presents those spans as evidence for users to assess how reliably correct the decision is.
|
|
Reliability Evaluation
Likelihood-based Mitigation of Evaluation Bias in Large Language Models
Masanari Ohiβ , Masahiro Kaneko, Ryuto
Koike, Mengsay Loem, Naoaki Okazaki
ACL 2024 (Findings)
TL;DR - We present a self-preference bias in LLM-as-a-judge i.e., LLMs overrate texts with higher
likelihoods while underrating those
with lower likelihoods. We also propose a simple but effective mitigation method via in-context
learning.
|
Experiences
|
|
Institute of Science Tokyo, Tokyo, Japan
Doctoral Researcher (2023.04 - Present)
Advisor: Prof. Naoaki Okazaki
|
|
University of Pennsylvania, Philadelphia, PA, USA
Visiting Researcher (2024.10 - Present)
Advisor: Prof. Chris Callison-Burch
|
|
Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), Abu Dhabi, UAE
Research Collaborate (2024.04 - Present)
Advisor: Prof. Preslav Nakov
|
|
Exawizards, Inc., Tokyo, Japan
Machine Learning Engineer Intern (2022.02 - 2022.03)
|
|
CyberAgent, Inc., Tokyo, Japan
Research Intern (2021.09 - 2022.01)
Software Engineer Intern (2021.07 - 2021.08)
|
Education
|
|
Institute of Science Tokyo (formerly Tokyo Institute of Technology), Tokyo, Japan
Ph.D. in Computer Science (2023.04 - est. 2026.04)
|
|
Keio University, Tokyo, Japan
M.S. in Information and Computer Science (2021.04 - 2023.03)
B.S. in Information and Computer Science (2017.04 - 2021.03)
|
Grants
- Off-Campus Study Plus in Tokyo Tech SPRING
Scholarship
Tokyo Institute of Technology, 2024.
Research Funds: 900,000 JPY / APPROX 6,000 USD
- Tobitate! (Leap for Tomorrow)
Study Abroad Scholarship (Acceptance Rate: 16.7%)
The Ministry of Education, Culture, Sports, Science and Technology (MEXT), 2024.
Scholarship: 1,920,000 JPY / APPROX 12,100 USD per year, Preparation Funds: 350,000 JPY /
APPROX 2,200 USD
- Tokyo Tech SPRING Scholarship
Tokyo Institute of Technology, Apr. 2024 - Mar.2026.
Scholarship: 2,160,000 JPY / APPROX 14,400 USD per year, Research Funds: 300,000 JPY / APPROX
2,000 USD per year,
Full Tuition Exemption.
- Tokyo Tech Advanced Human Resource
Development Fellowship for Doctoral Students
Tokyo Institute of Technology, Apr. 2023 - Mar. 2024.
Scholarship: 1,800,000 JPY / APPROX 12,000 USD per year, Research Funds: 300,000 JPY / APPROX
2,000 USD per year,
Full Tuition Exemption.
|
Honors
- Encouragement
Award (Top
23/187=12.3%)
The 19th Symposium of Young Researcher Association for NLP Studies (YANS 2024). Easily Detectable
LLMs Without Sacrificing Its Generative Capability.
- Sponsorship
Award from CyberAgent, Inc. (Top 2/187=1.1%)
The 19th Symposium of Young Researcher Association for NLP Studies (YANS 2024). Easily Detectable
LLMs Without Sacrificing Its Generative Capability.
- Sponsorship
Award from PKSHA Technology (Top 1/140=0.7%)
The 18th Symposium of Young Researcher Association for NLP Studies (YANS 2023). OUTFOX:
LLM-Generated Essay Detection Through In-Context Learning with Adversarially Generated
Examples.
- Sponsorship
Award from HAKUHODO Technologies (Top 1/140=0.7%)
The 18th Symposium of Young Researcher Association for NLP Studies (YANS 2023). OUTFOX:
LLM-Generated Essay Detection Through In-Context Learning with Adversarially Generated
Examples.
- Silver Medal (Top 165/4373=3.8%)
Kaggle, Mechanisms of Action (MoA) Prediction, 2020.
|
|