Building Next-Gen Multimodal Foundation Models for General-Purpose Assistants

LLaVA is an open-source project, collaborating with research community to advance the state-of-the-art in AI. LLaVA represents the first end-to-end trained large multimodal model (LMM) that achieves impressive chat capabilities mimicking spirits of the multimodal GPT-4. The LLaVA family continues growing to support more modalities, capabilities, applications and beyond.

인원

Open research collaboration across universities in the research community and multiple Microsoft team, pushing the SoTA in new capabilities scale and applications etc.

Hao Cheng의 초상화

Hao Cheng

Principal Researcher

Michel Galley의 초상화

Michel Galley

Senior Principal Researcher

Jianfeng Gao의 초상화

Jianfeng Gao

Distinguished Scientist & Vice President

Yong Jae Lee의 초상화

Yong Jae Lee

Associate Professor

University of Wisconsin-Madison

Lars Liden의 초상화

Lars Liden

Principal Research Software Engineer Manager

Haotian Liu의 초상화

Haotian Liu

Ph.D. student

University of Wisconsin-Madison

Xiaodong Liu의 초상화

Xiaodong Liu

Senior Principal Researcher

Yadong Lu의 초상화

Yadong Lu

Researcher

Microsoft Azure AI

Tristan Naumann의 초상화

Tristan Naumann

Principal Researcher

Hoifung Poon의 초상화

Hoifung Poon

General Manager, Health Futures

Yelong Shen의 초상화

Yelong Shen

Principal Researcher

Microsoft Azure AI

Swadheen Shukla의 초상화

Swadheen Shukla

Principal Program Manager

Andrea Tupini의 초상화

Andrea Tupini

Research Software Engineer

Naoto Usuyama의 초상화

Naoto Usuyama

Principal Researcher

Cliff Wong의 초상화

Cliff Wong

Principal Data Scientist

Sheng Zhang의 초상화

Sheng Zhang

Principal Researcher