SIE supports 85 pre-configured models across encode (embeddings), score (reranking), and extract (NER, relations, classification, vision). All models are quality-verified against MTEB benchmarks.
For model selection guidance, see Choosing a Model .
Model Dims Max Length Languages Bundle BAAI/bge-m31024 8192 100+ default Alibaba-NLP/gte-Qwen2-1.5B-instruct1536 32768 Multi default Alibaba-NLP/gte-Qwen2-7B-instruct3584 32000 Multi sglang Alibaba-NLP/gte-multilingual-base768 8192 50+ default NovaSearch/stella_en_400M_v51024 512 English default NovaSearch/stella_en_1.5B_v51024 512 English default intfloat/e5-small-v2384 512 English default intfloat/e5-base-v2768 512 English default intfloat/e5-large-v21024 512 English default intfloat/multilingual-e5-large1024 512 100+ default intfloat/multilingual-e5-large-instruct1024 512 100+ default intfloat/e5-mistral-7b-instruct4096 4096 English sglang sentence-transformers/all-MiniLM-L6-v2384 256 English default nomic-ai/nomic-embed-text-v2-moe768 2048 English default nvidia/NV-Embed-v24096 32768 English default nvidia/llama-embed-nemotron-8b4096 8192 English sglang Salesforce/SFR-Embedding-Mistral4096 4096 English sglang Salesforce/SFR-Embedding-2_R4096 8192 English sglang GritLM/GritLM-7B4096 8192 English default Linq-AI-Research/Linq-Embed-Mistral4096 32768 English sglang google/embeddinggemma-300m768 2048 English default Qwen/Qwen3-Embedding-0.6B1024 32768 Multi default Qwen/Qwen3-Embedding-4B2560 32768 Multi sglang
Model Vocab Size Max Length Bundle BAAI/bge-m3250002 8192 default naver/splade-v330522 512 default naver/splade-cocondenser-selfdistil30522 512 default prithivida/Splade_PP_en_v230522 256 default rasyosef/splade-mini30522 128 default ibm-granite/granite-embedding-30m-sparse30522 512 default opensearch-project/opensearch-neural-sparse-encoding-v1— — default opensearch-project/opensearch-neural-sparse-encoding-v2-distill— — default opensearch-project/opensearch-neural-sparse-encoding-doc-v2-distill— — default opensearch-project/opensearch-neural-sparse-encoding-doc-v2-mini— — default opensearch-project/opensearch-neural-sparse-encoding-doc-v3-distill— — default opensearch-project/opensearch-neural-sparse-encoding-doc-v3-gte— — default
Model Token Dim Max Length Bundle jinaai/jina-colbert-v2128 8192 default colbert-ir/colbertv2.0128 512 default answerdotai/answerai-colbert-small-v196 512 default mixedbread-ai/mxbai-colbert-large-v1128 512 default mixedbread-ai/mxbai-edge-colbert-v0-32m128 512 default lightonai/GTE-ModernColBERT-v1128 8192 default lightonai/Reason-ModernColBERT128 8192 default nvidia/llama-nemoretriever-colembed-3b-v11024 512 default
Model Dims Resolution Task Bundle openai/clip-vit-base-patch32512 224 Image+text embedding default openai/clip-vit-large-patch14768 224 Image+text embedding default laion/CLIP-ViT-B-32-laion2B-s34B-b79K512 224 Image+text embedding default laion/CLIP-ViT-H-14-laion2B-s32B-b79K1024 224 Image+text embedding default google/siglip-so400m-patch14-2241152 224 Image+text embedding default google/siglip-so400m-patch14-3841152 384 Image+text embedding default vidore/colpali-v1.3-hf128 1024 Document vision (ColBERT) default vidore/colqwen2.5-v0.2128 1024 Document vision (ColBERT) default
Model Max Length Languages Bundle BAAI/bge-reranker-base512 English default BAAI/bge-reranker-large512 English default BAAI/bge-reranker-v2-m38192 100+ default jinaai/jina-reranker-v2-base-multilingual8192 100+ default mixedbread-ai/mxbai-rerank-base-v28192 English default mixedbread-ai/mxbai-rerank-large-v28192 English default Alibaba-NLP/gte-reranker-modernbert-base8192 English default cross-encoder/ms-marco-MiniLM-L-6-v2512 English default cross-encoder/ms-marco-MiniLM-L-12-v2512 English default
ColBERT models can also be used for reranking via MaxSim scoring. See the Multi-Vector section above .
Model Languages Notes Bundle urchade/gliner_small-v2.1English Small gliner urchade/gliner_medium-v2.1English Medium gliner urchade/gliner_large-v2.1English Large gliner urchade/gliner_multi-v2.1Multilingual Recommended gliner urchade/gliner_multi_pii-v1Multilingual PII detection gliner EmergentMethods/gliner_large_news-v2.1English News domain gliner Ihor/gliner-biomed-large-v1.0English Biomedical gliner NeuML/gliner-bert-tinyEnglish Tiny, fastest gliner numind/NuNER_ZeroEnglish Zero-shot gliner numind/NuNER_Zero-spanEnglish Span extraction gliner
Model Notes Bundle jackboyla/glirel-large-v0Zero-shot relation extraction gliner
Model Approach Max Length Bundle knowledgator/gliclass-small-v1.0GLiClass 512 gliner knowledgator/gliclass-base-v1.0GLiClass 512 gliner MoritzLaurer/deberta-v3-base-zeroshot-v2.0NLI 512 default MoritzLaurer/deberta-v3-large-zeroshot-v2.0NLI 512 default
Model Tasks Bundle microsoft/Florence-2-baseOCR, caption, detection florence2 microsoft/Florence-2-largeOCR, caption, detection florence2 microsoft/Florence-2-base-ftOCR, caption, detection florence2 mynkchaudhry/Florence-2-FT-DocVQADocument QA florence2 naver-clova-ix/donut-base-finetuned-docvqaDocument QA florence2 naver-clova-ix/donut-base-finetuned-cord-v2Receipt parsing florence2 naver-clova-ix/donut-base-finetuned-rvlcdipDocument classification florence2
Model Notes Bundle IDEA-Research/grounding-dino-tinySmaller, faster default IDEA-Research/grounding-dino-baseHigher quality default google/owlv2-base-patch16-ensembleOWL-ViT based default
Models require specific bundles due to dependency conflicts:
Bundle Image Tag Models defaultcuda12-defaultMost models (embeddings, rerankers, ColBERT, NLI classification) glinercuda12-glinerGLiNER, GLiREL, GLiClass, NuNER models sglangcuda12-sglangLLM-based models (e5-mistral-7b, Nemotron, SFR, etc.) florence2cuda12-florence2Florence-2, Donut vision models
See Bundles for details.
You can programmatically query which models are available on a running SIE instance:
from sie_sdk import SIEClient
from sie_sdk.types import Item
client = SIEClient( "http://localhost:8080" )
models = client.list_models()
print ( f " { model.name } : { model.dims } dims, loaded= { model.loaded } " )
import { SIEClient } from "@sie/sdk" ;
const client = new SIEClient ( "http://localhost:8080" );
const models = await client. listModels ();
for ( const model of models) {
console. log ( `${ model . name }: ${ model . dims ?. dense } dims, loaded=${ model . loaded }` );
SIE can serve any HuggingFace model that fits an existing adapter. See Adding Models .