Slide 13
Slide 13 text
参考⽂献
12
● Takuya Akiba, Makoto Shing, Yujin Tang, Qi Sun, and David Ha. Evolutionary optimization of
model merging recipes, 2024.
● Anthropic. Introducing the next generation of claude. available at:
https://www.anthropic.com/news/ claude-3-family.
● Jinze Bai, Shuai Bai, Shusheng Yang, Shijie Wang, Sinan Tan, Peng Wang, Junyang Lin, Chang
Zhou, and Jingren Zhou. Qwen-vl: A versatile vision-language model for understanding,
localization, text reading, and beyond, 2023.
● Yash Goyal, Tejas Khot, Douglas Summers-Stay, Dhruv Batra, and Devi Parikh. Making the V in
VQA Matter: Elevating the role of image understanding in visual question answering. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR),
July 2017.
● Drew A. Hudson and Christopher D. Manning. Gqa: A new dataset for real-world visual
reasoning and compositional question answering. In Proceedings of the IEEE/CVF
Conference on Computer Vision and Pattern Recognition (CVPR), June 2019.
● Meng Lee, Fujiki Nakamura, Makoto Shing, Paul McCann, Takuya Akiba, and Naoki Orii.
Japanese stablelm base alpha 7b.