Seventh Meeting
Presenter
Qirun Dai and Shixuan Liu.
References
- Akyürek, Ekin, Tolga Bolukbasi, Frederick Liu, Binbin Xiong, Ian Tenney, Jacob Andreas, and Kelvin Guu. “Towards tracing factual knowledge in language models back to the training data.” arXiv preprint arXiv:2205.11482 (2022).
- Li, Dongfang, Zetian Sun, Xinshuo Hu, Zhenyu Liu, Ziyang Chen, Baotian Hu, Aiguo Wu, and Min Zhang. “A survey of large language models attribution.” arXiv preprint arXiv:2311.03731 (2023).
- Youssef, Paul, Osman Alperen Koraş, Meijie Li, Jörg Schlötterer, and Christin Seifert. “Give me the facts! a survey on factual knowledge probing in pre-trained language models.” arXiv preprint arXiv:2310.16570 (2023).
- Worledge, Theodora, Judy Hanwen Shen, Nicole Meister, Caleb Winston, and Carlos Guestrin. “Unifying corroborative and contributive attributions in large language models.” In 2024 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), pp. 665-683. IEEE, 2024.
- Konz, Nicholas, Charles Godfrey, Madelyn Shapiro, Jonathan Tu, Henry Kvinge, and Davis Brown. “Attributing Learned Concepts in Neural Networks to Training Data.” arXiv preprint arXiv:2310.03149 (2023).