Library/flash-moe
Library/flash-moeForked

danveloper/flash-moe

flash-moe

Running a big model on a small laptop

Builder

danveloper

danveloper

danveloper • individual

Stars

3,181

Using upstream star count

Forks

368

Using upstream fork count

Open Issues

0

Activity Score

0/100

0 commits in 30d

Created

Mar 18, 2026

Project creation date

README Summary

Flash-moe is a project focused on enabling the execution of large machine learning models on resource-constrained devices like laptops. It appears to optimize model inference to work within the memory and computational limitations of smaller hardware configurations.

AI Dev Skills

Unmapped

Mixture of Experts ArchitectureModel OptimizationEfficient InferenceMemory ManagementMobile/Edge AI Deployment

Tags

Mixture of Experts ArchitectureModel OptimizationEfficient InferenceMemory ManagementMobile/Edge AI DeploymentEdge ComputingEdge/MobileSelf-hostedModel EfficiencyDemocratized AI AccessOn-device AITextOffline AI ApplicationsResource-Constrained Model DeploymentLocal AI Model InferenceObjective-C

Taxonomy

Recent Activity

Updated 25 days ago

7 Days

0

30 Days

0

90 Days

0

Quality

research
Quality
low
Maturity
research

Categories

Edge & Mobile AIPrimaryInference & ServingOther AI / ML

PM Skills

Scale & ReliabilityCost & Efficiency

Languages

Objective-C100.0%

Timeline

Project created
Mar 18, 2026
Forked
Mar 23, 2026
Your last push
25 days ago
Upstream last push
25 days ago
Tracked since
Mar 19, 2026

Similar Repos

pgvector cosine similarity · $0

Loading…