Building Next-Gen Multimodal Foundation Models for General-Purpose Assistants

LLaVA is an open-source project, collaborating with research community to advance the state-of-the-art in AI. LLaVA represents the first end-to-end trained large multimodal model (LMM) that achieves impressive chat capabilities mimicking spirits of the multimodal GPT-4. The LLaVA family continues growing to support more modalities, capabilities, applications and beyond.

人数

Open research collaboration across universities in the research community and multiple Microsoft team, pushing the SoTA in new capabilities scale and applications etc.

Hao Chengの肖像

Hao Cheng

Principal Researcher

Michel Galleyの肖像

Michel Galley

Senior Principal Researcher

Jianfeng Gaoの肖像

Jianfeng Gao

Distinguished Scientist & Vice President

Yong Jae Leeの肖像

Yong Jae Lee

Associate Professor

University of Wisconsin-Madison

Lars Lidenの肖像

Lars Liden

Principal Research Software Engineer Manager

Haotian Liuの肖像

Haotian Liu

Ph.D. student

University of Wisconsin-Madison

Xiaodong Liuの肖像

Xiaodong Liu

Senior Principal Researcher

Yadong Luの肖像

Yadong Lu

Researcher

Microsoft Azure AI

Tristan Naumannの肖像

Tristan Naumann

Principal Researcher

Hoifung Poonの肖像

Hoifung Poon

General Manager, Health Futures

Yelong Shenの肖像

Yelong Shen

Principal Researcher

Microsoft Azure AI

Swadheen Shuklaの肖像

Swadheen Shukla

Principal Program Manager

Andrea Tupiniの肖像

Andrea Tupini

Research Software Engineer

Naoto Usuyamaの肖像

Naoto Usuyama

Principal Researcher

Cliff Wongの肖像

Cliff Wong

Principal Data Scientist

Sheng Zhangの肖像

Sheng Zhang

Principal Researcher