AI

BCEmbedding: Bilingual Cross-Modal Embedding Models from NetEase

BCEmbedding is a bilingual cross-modal embedding model for semantic search, RAG, and cross-lingual retrieval with state-of-the-art performance.

Keeping this site alive takes effort — your support means everything.
無程式碼也能輕鬆打造專業LINE官方帳號!一鍵導入模板,讓AI助你行銷加分! 無程式碼也能輕鬆打造專業LINE官方帳號!一鍵導入模板,讓AI助你行銷加分!
BCEmbedding: Bilingual Cross-Modal Embedding Models from NetEase

Embedding models are the foundation of modern semantic search and retrieval-augmented generation (RAG) systems. BCEmbedding, developed by NetEase Youdao, stands out by delivering state-of-the-art performance specifically optimized for bilingual Chinese-English and cross-modal retrieval tasks.

The model excels at understanding semantic relationships across languages and modalities. Whether you are searching Chinese documents with English queries, retrieving images from text descriptions, or building a bilingual RAG pipeline, BCEmbedding provides embeddings that capture meaning across these boundaries.

Model Capabilities

CapabilityDescriptionPerformance
Bilingual textChinese-English cross-lingual retrievalTop 3 on MTEB leaderboard
Cross-modalText-to-image and image-to-text retrievalState-of-the-art
Dense retrievalSingle-vector representationCompetitive with BGE
Sparse retrievalHybrid with BM25 supportEnhanced recall
RAG optimizationTuned for chunk-level retrievalExcellent precision

Embedding Architecture

The architecture uses separate encoders for text and vision, with a cross-modal fusion layer that projects both modalities into a shared embedding space. This allows direct comparison between any combination of text and image inputs.

Performance Benchmarks

BenchmarkBCEmbeddingBGE-largeOpenAI ada-002
MTEB (English)64.564.261.0
C-MTEB (Chinese)67.866.5N/A
Cross-lingual retrieval72.368.142.5
Image-text retrieval85.6N/A80.2

For more details, visit the BCEmbedding GitHub repository and check the MTEB leaderboard.

Frequently Asked Questions

Q: What embedding dimensions does BCEmbedding output? A: The text model outputs 768-dimensional vectors, matching the BGE-large architecture.

Q: Can I use BCEmbedding with LangChain or LlamaIndex? A: Yes, it integrates easily through HuggingFace embedding wrappers compatible with both frameworks.

Q: Is BCEmbedding free for commercial use? A: Yes, it is released under the Apache 2.0 license.

Q: Does it support languages beyond Chinese and English? A: It is optimized for Chinese-English. Other languages have reduced but functional performance.

Q: How large is the model? A: The text encoder is approximately 1.3GB (BGE-large based), and the vision encoder adds about 0.5GB.

TAG
CATEGORIES