The slides used for the guest lecture in the ML in feedback systems class (CS6784) at Cornell.
video recording: https://vod.video.cornell.edu/media/Guest+Lecture%3A+Off-Policy+Evaluation+and+Learning+%28ML+In+Feedback+Sys+F25%29+/1_eyiazrlc
class info: https://github.com/ml-feedback-sys/materials-f25/tree/main
----
The main content is (1) Intro to off-policy evaluation and learning (OPE/L) and (2) Research example of OPL for sentence personalization. I also briefly mentioned my research interests related to "ML in feedback systems" topics, and mentioned the following papers in the lecture.
OPL for sentence personalization
paper: https://arxiv.org/abs/2504.02646
slides: https://speakerdeck.com/harukakiyohara_/opl-prompt
Steering systems for long-term objectives
paper: https://arxiv.org/abs/2502.01792
slides: https://speakerdeck.com/harukakiyohara_/dynamics-two-stage-rec
Scalable and adaptable RecSys under practical constraints
(on-going, workshop) paper: https://drive.google.com/file/d/1pc7aa5dvv9cpMRnbUDaeh9-dKP6wzC5J/view?usp=drive_link
----
My research statement is available here.
https://drive.google.com/file/d/1LqONxB8Qw4Z0GSUavAwSSV_oCANcl9TI/view?usp=sharing