haotian-liu/LLaVA
LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Builder

haotian-liu
haotian-liu • individual
Stars
24,623
Using upstream star count
Forks
2,751
Using upstream fork count
Open Issues
0
Activity Score
0/100
0 commits in 30d
Created
Apr 17, 2023
Project creation date
README Summary
LLaVA (Large Language and Vision Assistant) is a multimodal AI model that combines visual and language understanding through instruction tuning. The project aims to build GPT-4V level capabilities by training large language models to process and respond to both text and image inputs. It was presented as an oral paper at NeurIPS 2023 and represents a significant advancement in visual instruction following.
AI Dev Skills
Unmapped
Tags
Taxonomy
Deployment Context
Modalities
Skill Areas
Recent Activity
Updated 1 years ago
7 Days
0
30 Days
0
90 Days
0
Quality
research- Quality
- high
- Maturity
- research
Categories
PM Skills
Languages
Timeline
- Project created
- Apr 17, 2023
- Forked
- Mar 13, 2026
- Your last push
- 1 years ago
- Upstream last push
- 1 years ago
- Tracked since
- Aug 12, 2024
Similar Repos
pgvector cosine similarity · $0
Loading…