https://github.com/guenthermi
jina.ai/news/what-we...
jina.ai/news/what-we...
jina.ai/news/how-ima...
jina.ai/news/how-ima...
👨💻 github.com/jina-ai/jina...
📑 jina.ai/news/jinavdr...
👨💻 github.com/jina-ai/jina...
📑 jina.ai/news/jinavdr...
📅 I'll present it on July 17th
📝 Pre-print: arxiv.org/abs/2409.04701
🔗 Workshop: sigir-2025-workshop-on-robust-ir.github.io
📅 I'll present it on July 17th
📝 Pre-print: arxiv.org/abs/2409.04701
🔗 Workshop: sigir-2025-workshop-on-robust-ir.github.io
Details in 🧵
Details in 🧵
jina.ai/news/quantiz...
jina.ai/news/quantiz...
- multilingual text-to-text and text-to-image search w/o modality gap
- also visual docs (e.g. pdfs, maps) - trained on a wider scope than DSE, ColPali, etc.
+ MRL, late interaction, etc.
🤗 huggingface.co/jinaai/jina-...
📄 arxiv.org/abs/2506.18902
- multilingual text-to-text and text-to-image search w/o modality gap
- also visual docs (e.g. pdfs, maps) - trained on a wider scope than DSE, ColPali, etc.
+ MRL, late interaction, etc.
🤗 huggingface.co/jinaai/jina-...
📄 arxiv.org/abs/2506.18902
jina.ai/news/on-the-...
jina.ai/news/on-the-...
Supports Multiple Languages and Dynamic Resolution (up to 4K)
🤗 huggingface.co/jinaai/jina-...
Supports Multiple Languages and Dynamic Resolution (up to 4K)
🤗 huggingface.co/jinaai/jina-...
Online Event: lu.ma/j8g0wnit
Paper: arxiv.org/abs/2503.09516
Online Event: lu.ma/j8g0wnit
Paper: arxiv.org/abs/2503.09516
It sees to work out-of-the-box and generally boost the performance of embedding models. However, it requires more latency. Would be interesting to see more about this.
📃: jina.ai/news/query-e...
🛠️: github.com/jina-ai/llm-...
It sees to work out-of-the-box and generally boost the performance of embedding models. However, it requires more latency. Would be interesting to see more about this.
📃: jina.ai/news/query-e...
🛠️: github.com/jina-ai/llm-...
Baichuan releases Baichuan-Omni-1.5
Open-source Omni-modal Foundation Model Supporting Text, Image, Video, and Audio Inputs as Well as Text and Audio Outputs.
Both model ( huggingface.co/baichuan-inc... ) and base ( huggingface.co/baichuan-inc... ).
Baichuan releases Baichuan-Omni-1.5
Open-source Omni-modal Foundation Model Supporting Text, Image, Video, and Audio Inputs as Well as Text and Audio Outputs.
Both model ( huggingface.co/baichuan-inc... ) and base ( huggingface.co/baichuan-inc... ).
github.com/guenthermi/t...
github.com/guenthermi/t...
jina.ai/news/what-sh...
jina.ai/news/what-sh...
jina.ai/news/readerl...
jina.ai/news/readerl...
At the ECIR Industry Day my colleague @str-saba.bsky.social presents how we train the latest version of our text embedding model.
More details on ECIR: ecir2025.eu
More details about the model: arxiv.org/abs/2409.10173
At the ECIR Industry Day my colleague @str-saba.bsky.social presents how we train the latest version of our text embedding model.
More details on ECIR: ecir2025.eu
More details about the model: arxiv.org/abs/2409.10173
jina.ai/news/scaling...
jina.ai/news/scaling...
t.co/BLC3WTU3LP
t.co/BLC3WTU3LP
We are releasing SmolVLM: a new 2B small vision language made for on-device use, fine-tunable on consumer GPU, immensely memory efficient 🤠
We release three checkpoints under Apache 2.0: SmolVLM-Instruct, SmolVLM-Synthetic and SmolVLM-Base huggingface.co/collections/...
We are releasing SmolVLM: a new 2B small vision language made for on-device use, fine-tunable on consumer GPU, immensely memory efficient 🤠
We release three checkpoints under Apache 2.0: SmolVLM-Instruct, SmolVLM-Synthetic and SmolVLM-Base huggingface.co/collections/...