layout | title |
---|---|
News |
**Nov 2024**: Happy to have been part of the exciting [Tülu 3](https://allenai.org/papers/tulu-3-report.pdf) and [OLMo 2](https://allenai.org/olmo) releases! My primary contribution is scaling-up our preference data using a synthetic on-policy pipeline that led to improvements in our DPO models.
**Oct 2024**: Our paper on routing preference instances to human or LM annotators, [Hybrid Preferences](https://arxiv.org/abs/2410.19133), is now available. This is the first work I co-led (with Yizhong Wang) at Ai2!
**Oct 2024**: Our paper on evaluating reward models in multilingual settings, [M-RewardBench](https://arxiv.org/abs/2410.15522), is now available. This was a fun collab with folks from Cohere for AI!
**Sep 2024**: My cross-institutional collabs, [Consent in Crisis](https://arxiv.org/abs/2407.14933) and [SEACrowd](https://arxiv.org/abs/2406.10118), were accepted to NeurIPS D&B and EMNLP 2024, respectively.
**Aug 2024**: 🏆 Our work on evaluating reward models in multilingual settings won Silver Prize in [Cohere for AI's Aya Expedition](https://cohere.com/blog/empowering-others-to-explore-the-next-frontier-expedition-aya)!
**Jul 2024**: I gave a [guest lecture](/notebook/2024/07/02/talk-dlsu/) at DLSU about building Filipino NLP resources. Thanks to Dr. Charibeth Cheng for inviting me!
**Mar 2024**: [Universal NER](https://arxiv.org/abs/2311.09122) was accepted to NAACL 2024. I hope to still work on linguistic aspects of NLP in the future!
**Mar 2024**: We released [RewardBench](https://arxiv.org/abs/2403.13787), the first benchmark for evaluating reward models.