Microsoft

MiniLMv2-L6-H384

Efficient, smaller embeddings for resource-constrained applications.

high performance at lower dimensionsfast inference speedgood for mobile and edge devices
Today's score
89.0
Try MiniLMv2-L6-H384

Where it ranks today

Best for / Not great for

Best for
  • real-time search on devices
  • low-latency applications
  • embedding generation on a budget
Not great for
  • complex semantic understanding
  • very large scale enterprise search
  • capturing subtle nuances

Why it ranks here

Despite its smaller size, MiniLMv2 provides surprisingly strong performance, making it ideal for applications where efficiency and speed are paramount over maximum accuracy.

30-day trend

Score breakdown

Search trends90
Benchmarks88
Developer buzz91
News mentions86

Pricing

API: $0.00 in · $0.00 out per 1M tokens · Consumer: $0.00/mo

Pricing plans

Popular
Hugging Face Hub
Download and use freely.
Free
  • Open-source model
  • Community availability
  • Easy integration
Get on Hugging Face
Cloud Inference
Deploy for scalable inference.
Custom
  • Managed endpoints
  • Pay-per-call
  • Customizable infrastructure
Deploy via SageMaker
Compare with another modelHow is this score calculated? →Snapshot 2026-05-12