multi-qa-MiniLM-L6-cos-v1

Maintained By
sentence-transformers

multi-qa-MiniLM-L6-cos-v1

PropertyValue
Parameter Count22.7M
Embedding Dimensions384
Training Data215M QA pairs
Model TypeSentence Transformer

What is multi-qa-MiniLM-L6-cos-v1?

multi-qa-MiniLM-L6-cos-v1 is a compact but powerful sentence embedding model designed specifically for semantic search applications. It transforms text inputs into 384-dimensional dense vector representations, enabling efficient similarity comparisons between sentences and paragraphs. The model was trained on an extensive dataset of 215 million question-answer pairs from diverse sources including WikiAnswers, Stack Exchange, and MS MARCO.

Implementation Details

The model implements a mean pooling architecture with normalized embeddings and supports multiple frameworks including PyTorch, TensorFlow, and ONNX. It processes text up to 512 word pieces (with optimal performance for texts under 250 word pieces) and produces normalized embeddings that can be compared using dot-product or cosine similarity metrics.

  • Optimized for semantic search and question-answering tasks
  • Supports multiple deep learning frameworks
  • Efficient 384-dimensional embeddings
  • Pre-trained on MiniLM-L6-H384-uncased architecture

Core Capabilities

  • Semantic search across document collections
  • Question-answer matching
  • Text similarity comparison
  • Dense passage retrieval
  • Cross-encoder pre-screening

Frequently Asked Questions

Q: What makes this model unique?

The model's distinctive feature is its optimal balance between size and performance, trained on one of the largest and most diverse question-answer datasets (215M pairs). It produces normalized embeddings that enable fast similarity computations using simple dot products.

Q: What are the recommended use cases?

The model excels in semantic search applications, document retrieval, and question-answering systems. It's particularly effective for applications requiring fast and accurate semantic similarity matching between shorter texts (up to 250 word pieces).

The first platform built for prompt engineering