A photo of me!

Hi, I’m Hamish! I’m (currently) a PhD student at the University of Washington at H2Lab, advised by Hannaneh Hajishirzi. I’m generally interested in NLP research, with a focus on post-training for language models. I’m interested in making language models more usable for more people, and exploring ways to improve them that go beyond next-token training. Additionally, I’m also interested in improving and exploring language model data mixtures, and have dabbled in exploring alternatives approaches to language modelling.

I’m from Sydney and did my undergraduate at the University of Sydney, doing a Bachelor of Arts and IT and triple majoring in Linguistics, Classical Greek, and Computer Science. I also did some NLP with the UsydNLP group, examining multi-hop question answering. Throughout my undergrad (and just after), I spent some time at the Commonwealth Bank of Australia, start-up-y stuff, and Optiver. Before my PhD, I was a predoctoral researcher at AI2 on the AllenNLP team.

If you have questions about my work, general academia/software/research-related stuff, or want to chat, feel free to reach out at hamishiv [at] cs [dot] washington [dot] edu. I am generally happy to answer most questions! You can also find me on various social media at @hamishivi.


Papers

See below for papers I’ve worked on. You can also check out my Semantic Scholar and Google Scholar profiles.

In addition to these, I also help maintain Open-Instruct, a codebase for general LM post-training. Send me a note if you need some help with it!

    2 OLMo 2 FuriousTeam OLMo (inc. Hamish Ivison). 2024.
    Tülu 3: Pushing Frontiers in Open Language Model Post-TrainingNathan Lambert*, Jacob Morrison*, Valentina Pyatkin*, Shengyi Huang*, Hamish Ivison*, Faeze Brahman*, Lester James V. Miranda*, Alisa Liu, Nouha Dziri, Shane Lyu, Yuling Gu, Saumya Malik, Victoria Graf, Jena D. Hwang, Jiangjiang Yang, Ronan Le Bras, Oyvind Tafjord, Chris Wilhelm, Luca Soldaini, et al. 2024.
    Personalizing Reinforcement Learning from Human Feedback with Variational Preference LearningSriyash Poddar*, Yanming Wan*, Hamish Ivison, Abhishek Gupta, and Natasha Jaques. 2024. NeurIPS.
    Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference FeedbackHamish Ivison, Yizhong Wang, Jiacheng Liu, Zeqiu Wu, Valentina Pyatkin, Nathan Lambert, Noah A. Smith, Yejin Choi, and Hannaneh Hajishirzi. 2024. NeurIPS.
    OLMo: Accelerating the Science of Language ModelsDirk Groeneveld, Iz Beltagy, ..., Hamish Ivison, ..., Noah A. Smith, and Hannaneh Hajishirzi. 2024. ACL.
    Backtracking Mathematical Reasoning of Language Models to the Pretraining DataYasaman Razeghi*, Hamish Ivison*, Sameer Singh, and Yanai Elazar. 2024. The Second Tiny Papers Track at ICLR 2024.
    TESS: Text-to-Text Self-Conditioned Simplex DiffusionRabeeh Karimi Mahabadi*, Hamish Ivison*, Jaesung Tae, James Henderson, Iz Beltagy, Matthew E. Peters, and Arman Cohan. 2024. EACL.
    Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2Hamish Ivison*, Yizhong Wang*, Valentina Pyatkin, Nathan Lambert, Matthew Peters, Pradeep Dasigi, Joel Jang, David Wadden, Noah A. Smith, Iz Beltagy, and Hannaneh Hajishirzi. 2023. technical report.
    How Far Can Camels Go? Exploring the State of Instruction Tuning on Open ResourcesYizhong Wang*, Hamish Ivison*, Pradeep Dasigi, Jack Hessel, Tushar Khot, Khyathi Raghavi Chandu, David Wadden, Kelsey MacMillan, Noah A. Smith, Iz Beltagy, and Hannaneh Hajishirzi. 2023. NeurIPS Datasets and Benchmarks Track.
    HINT: Hypernetwork Instruction Tuning for Efficient Zero-Shot GeneralisationHamish Ivison, Akshita Bhagia, Yizhong Wang, Hannaneh Hajishirzi, and Matthew Peters. 2023. ACL.
    Data-Efficient Finetuning Using Cross-Task Nearest NeighborsHamish Ivison, Noah A. Smith, Hannaneh Hajishirzi, and Pradeep Dasigi. 2023. Findings of ACL.
    Hyperdecoders: Instance-specific decoders for multi-task NLPHamish Ivison and Matthew E. Peters. 2022. Findings of EMNLP.
    Local Interpretations for Explainable Natural Language Processing: A SurveySiwen Luo*, Hamish Ivison*, Soyeon Caren Han, and Josiah Poon. 2021. ACM Computing Surveys.
    Would you like fries with that? Modular Multi-hop ReasoningHamish Ivison. 2020. November.