Multimodal Task Vectors Enable Many-Shot Multimodal In-Context Learning
Brandon Huang and Chancharik Mitra and Assaf Arbelle and Leonid Karlinsky and Trevor Darrell and Roei Herzig
EECS Department, University of California, Berkeley
Technical Report No. UCB/EECS-2025-71
May 15, 2025
http://www2.eecs.berkeley.edu/Pubs/TechRpts/2025/EECS-2025-71.pdf
BibTeX citation:
@mastersthesis{Huang:EECS-2025-71, Author= {Huang, Brandon and Mitra, Chancharik and Arbelle, Assaf and Karlinsky, Leonid and Darrell, Trevor and Herzig, Roei}, Title= {Multimodal Task Vectors Enable Many-Shot Multimodal In-Context Learning}, School= {EECS Department, University of California, Berkeley}, Year= {2025}, Month= {May}, Url= {http://www2.eecs.berkeley.edu/Pubs/TechRpts/2025/EECS-2025-71.html}, Number= {UCB/EECS-2025-71}, }
EndNote citation:
%0 Thesis %A Huang, Brandon %A Mitra, Chancharik %A Arbelle, Assaf %A Karlinsky, Leonid %A Darrell, Trevor %A Herzig, Roei %T Multimodal Task Vectors Enable Many-Shot Multimodal In-Context Learning %I EECS Department, University of California, Berkeley %D 2025 %8 May 15 %@ UCB/EECS-2025-71 %U http://www2.eecs.berkeley.edu/Pubs/TechRpts/2025/EECS-2025-71.html %F Huang:EECS-2025-71