--- license: bigscience-bloom-rail-1.0 datasets: - cmarkea/mmarco-contrastive language: - fr - en pipeline_tag: feature-extraction --- # Bloomz-560m-retriever-v2 ## Presentation We introduce the Bloomz-560m-retriever-v2 model, based on the [Bloomz-560m-dpo-chat](https://huggingface.co/ArkeaIAF/bloomz-560m-dpo-chat) model. This bi-encoder projects articles and queries into the same vector space, ensuring the proximity of queries to related articles. The model is language-agnostic for French and English, meaning a query in either language will be close to an article regardless of whether it is in French or English. This model is ideal for Open Domain Question Answering (ODQA) and can be complemented by the rerankers [Bloomz-560m-reranking](https://huggingface.co/cmarkea/bloomz-560m-reranking) or [Bloomz-3b-reranking](https://huggingface.co/cmarkea/bloomz-3b-reranking). ## Training The training dataset used is a variant of [mMARCO](https://huggingface.co/datasets/cmarkea/mmarco-contrastive) enabling contrastive learning and filtering out false negatives. The filtering threshold was set at 0.8, and a positive observation is confronted with 10 hard negatives, ordered by decreasing score (the 10 hardest). The model was trained on a uniform distribution of languages (1/4 French-French, 1/4 French-English, 1/4 English-French, 1/4 English-English). The learning objective is of the InfoNCE type with a trainable temperature parameter as presented for the [CLIP](https://arxiv.org/abs/2103.00020) model. ## Note Unlike the [Bloomz-560m-retriever](https://huggingface.co/cmarkea/bloomz-560m-retriever), this much more efficient model uses cosine distance as its metric (instead of L2 distance as previously). ## Benchmark The performance evaluation is based on the evaluation portion of SQuAD (5921 queries over 1204 articles across 35 different topics). One interesting aspect of this dataset is having multiple articles associated with a single theme, representing challenging contexts where a query may be close to several relevant articles. On average, there are about thirty articles per theme (see [Bloomz-560m-reranking](https://huggingface.co/cmarkea/bloomz-560m-reranking) for the exact distribution). We compare the performances using the average top rank of the articles targeted by a query (Top-mean), the standard deviation of the top ranks (Top-std), the percentage of correct articles within Top-1, Top-5, and Top-10, and finally, the mean reciprocal rank (MRR) across the 1204 articles. | Model (FR/FR) | Top-mean | Top-std | Top-1 (%) | Top-5 (%) | Top-10 (%) | MRR (%) | |----------------------------------------------------------------------------------------------------:|:--------:|:-------:|:---------:|:---------:|:----------:|:--------:| | BM25 | 16.8 | 100.8 | 71.7 | 88.3 | 91.8 | 79.2 | | [CamemBERT](https://huggingface.co/camembert/camembert-base) | 269.6 | 303.0 | 5.6 | 12.5 | 16.5 | 9.7 | | [STS-CamemBERT](h4c5/sts-camembert-base) | 23.1 | 85.5 | 36.0 | 63.0 | 74.0 | 48.5 | | [Sentence-BERT](https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2) | 10.2 | 40.1 | 43.9 | 73.9 | 84.0 | 57.3 | | [E5-base](https://huggingface.co/intfloat/multilingual-e5-base) | 6.1 | 29.7 | 59.9 | 84.9 | 91.0 | 71.1 | | [E5-large](https://huggingface.co/intfloat/multilingual-e5-large) | 5.2 | 29.2 | 67.0 | 89.2 | 93.7 | 76.7 | | [Bloomz-560m-retriever](https://huggingface.co/cmarkea/bloomz-560m-retriever) | 10.2 | 46.6 | 51.5 | 78.1 | 86.2 | 63.5 | | [Bloomz-3b-retriever](https://huggingface.co/cmarkea/bloomz-3b-retriever) | 8.8 | 36.4 | 49.2 | 77.5 | 86.1 | 62.0 | | [Bloomz-560m-retriever-v2](https://huggingface.co/cmarkea/bloomz-560m-retriever-v2) | 4.0 | 17.1 | 68.0 | 89.9 | 94.4 | 77.7 | | [Bloomz-3b-retriever-v2](https://huggingface.co/cmarkea/bloomz-3b-retriever-v2) | 2.8 | 14.8 | 76.5 | 94.4 | 97.2 | 84.4 | | Model (EN/FR) | Top-mean | Top-std | Top-1 (%) | Top-5 (%) | Top-10 (%) | MRR (%) | |----------------------------------------------------------------------------------------------------:|:--------:|:-------:|:---------:|:---------:|:----------:|:--------:| | BM25 | 280.7 | 371.8 | 23.9 | 37.4 | 43.3 | 30.4 | | [CamemBERT](https://huggingface.co/camembert/camembert-base) | 355.0 | 328.3 | 0.9 | 3.7 | 6.4 | 3.13 | | [STS-CamemBERT](h4c5/sts-camembert-base) | 102.2 | 196.9 | 13.1 | 30.5 | 40.7 | 22.1 | | [Sentence-BERT](https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2) | 10.6 | 41.2 | 43.3 | 72.4 | 82.7 | 56.5 | | [E5-base](https://huggingface.co/intfloat/multilingual-e5-base) | 9.9 | 38.1 | 49.8 | 77.2 | 85.4 | 62.6 | | [E5-large](https://huggingface.co/intfloat/multilingual-e5-large) | 5.6 | 26.9 | 62.9 | 86.9 | 92.5 | 73.8 | | [Bloomz-560m-retriever](https://huggingface.co/cmarkea/bloomz-560m-retriever) | 11.0 | 47.8 | 48.3 | 75.7 | 84.7 | 60.4 | | [Bloomz-3b-retriever](https://huggingface.co/cmarkea/bloomz-3b-retriever) | 8.9 | 37.6 | 48.8 | 77.4 | 86.1 | 61.6 | | [Bloomz-560m-retriever-v2](https://huggingface.co/cmarkea/bloomz-560m-retriever-v2) | 4.4 | 18.9 | 66.6 | 89.3 | 94.1 | 76.6 | | [Bloomz-3b-retriever-v2](https://huggingface.co/cmarkea/bloomz-3b-retriever-v2) | 2.7 | 14.2 | 75.7 | 94.5 | 97.1 | 83.9 | ## How to Use Bloomz-560m-retriever-v2 **With Transformers API:** ```python from typing import Union, List import numpy as np import torch from transformers import AutoTokenizer, AutoModel from scipy.spatial.distance import cdist tokenizer = AutoTokenizer.from_pretrained('cmarkea/bloomz-560m-retriever-v2') model = AutoModel.from_pretrained('cmarkea/bloomz-560m-retriever-v2') def infer(txt: Union[str, List[str]]): tok = tokenizer(txt, padding=True, return_tensors='pt') with torch.no_grad(): embedding = model(**tok) # Inportant: take only last token! return embedding.get('last_hidden_state')[:,-1,:].numpy() list_of_contexts: List[str] = [...] emb_contexts = infer(list_of_contexts) list_of_queries: List[str] = [...] emb_queries = infer(list_of_queries) # Important: take cosine distance! dist = cdist(emb_queries, emb_contexts, 'cosine') top_k = lambda x: [ [list_of_contexts[qq] for qq in ii] for ii in dist.argsort(axis=-1)[:,:x] ] # top 5 nearest contexts for each queries top_contexts = top_k(5) ``` **With Pipeline API:** ```python import numpy as np from transformers import pipeline from scipy.spatial.distance import cdist retriever = pipeline('feature-extraction', 'cmarkea/bloomz-560m-retriever-v2') # Inportant: take only last token! infer = lambda x: [ii[0][-1] for ii in retriever(x)] list_of_contexts: List[str] = [...] emb_contexts = np.concatenate(infer(list_of_contexts), axis=0) list_of_queries: List[str] = [...] emb_queries = np.concatenate(infer(list_of_queries), axis=0) # Important: take cosine distance! dist = cdist(emb_queries, emb_contexts, 'cosine') top_k = lambda x: [ [list_of_contexts[qq] for qq in ii] for ii in dist.argsort(axis=-1)[:,:x] ] # top 5 nearest contexts for each queries top_contexts = top_k(5) ``` Citation -------- ```bibtex @online{DeBloomzRetv2, AUTHOR = {Cyrile Delestre}, ORGANIZATION = {Cr{\'e}dit Mutuel Ark{\'e}a}, URL = {https://huggingface.co/cmarkea/bloomz-560m-retriever-v2}, YEAR = {2024}, KEYWORDS = {NLP ; Transformers ; LLM ; Bloomz}, } ```