folder_special
Избранное
school
Академические публикации

февраль 17, 2026
jina-embeddings-v5-text: Task-Targeted Embedding Distillation

февраль 11, 2026
Embedding Inversion via Conditional Masked Diffusion Language Models

ICLR 2026
январь 22, 2026
Embedding Compression via Spherical Coordinates

декабрь 29, 2025
Vision Encoders in Vision-Language Models: A Survey

ICLR 2026
декабрь 04, 2025
Jina-VLM: Small Multilingual Vision Language Model

AAAI 2026
октябрь 01, 2025
jina-reranker-v3: Last but Not Late Interaction for Document Reranking

NeurIPS 2025
август 31, 2025
Efficient Code Embeddings from Code Generation Models

EMNLP 2025
июнь 24, 2025
jina-embeddings-v4: Universal Embeddings for Multimodal Multilingual Retrieval

ICLR 2025
март 04, 2025
ReaderLM-v2: Small Language Model for HTML to Markdown and JSON

ACL 2025
декабрь 17, 2024
AIR-Bench: Automated Heterogeneous Information Retrieval Benchmark

ICLR 2025
декабрь 12, 2024
jina-clip-v2: Multilingual Multimodal Embeddings for Text and Images

ECIR 2025
сентябрь 18, 2024
jina-embeddings-v3: Multilingual Embeddings With Task LoRA

SIGIR 2025
сентябрь 07, 2024
Late Chunking: Contextual Chunk Embeddings Using Long-Context Embedding Models

EMNLP 2024
август 30, 2024
Jina-ColBERT-v2: A General-Purpose Multilingual Late Interaction Retriever

WWW 2025
июнь 21, 2024
Leveraging Passage Embeddings for Efficient Listwise Reranking with Large Language Models

ICML 2024
май 30, 2024
Jina CLIP: Your CLIP Model Is Also Your Text Retriever

февраль 26, 2024
Multi-Task Contrastive Learning for 8192-Token Bilingual Text Embeddings

октябрь 30, 2023
Jina Embeddings 2: 8192-Token General-Purpose Text Embeddings for Long Documents

EMNLP 2023
июль 20, 2023
Jina Embeddings: A Novel Set of High-Performance Sentence Embedding Models
Всего публикаций 19.
folder_special
Избранное
school
Академический
Все
пресс-релиз
Технический блог
Событие
Мнение
март 11, 2026 • 7 минуты чтения
Бутстрэппинг аудиоэмбеддингов на базе мультимодальных LLM
Превратите любую мультимодальную LLM в малую векторную модель аудио, которая превосходит CLAP, используя в 25 раз меньше данных.


март 06, 2026 • 6 минуты чтения
Идентификация векторных моделей по сырым числовым значениям
Крошечный трансформер, который создает цифровые отпечатки векторных моделей, считывая необработанные цифры. Без проектирования признаков.


февраль 19, 2026 • 7 минуты чтения
jina-embeddings-v5-text: новые SOTA компактные мультиязычные векторные модели
Две многоязычные модели Embeddings объемом менее 1 млрд параметров с лучшей в своем классе производительностью, доступные в Elastic Inference Service, Llama.cpp и MLX.


февраль 17, 2026
jina-embeddings-v5-text: Task-Targeted Embedding Distillation
Text embedding models are widely used for semantic similarity tasks, including information retrieval, clustering, and classification. General-purpose models are typically trained with single- or multi-stage processes using contrastive loss functions. We introduce a novel training regimen that combines model distillation techniques with task-specific contrastive loss to produce compact, high-performance embedding models. Our findings suggest that this approach is more effective for training small models than purely contrastive or distillation-based training paradigms alone. Benchmark scores for the resulting models, jina-embeddings-v5-text-small and jina-embeddings-v5-text-nano, exceed or match the state-of-the-art for models of similar size. jina-embeddings-v5-text models additionally support long texts (up to 32k tokens) in many languages, and generate embeddings that remain robust under truncation and binary quantization. Model weights are publicly available, hopefully inspiring further advances in embedding model development.

февраль 11, 2026
Embedding Inversion via Conditional Masked Diffusion Language Models
We frame embedding inversion as conditional masked diffusion, recovering all tokens in parallel through iterative denoising rather than sequential autoregressive generation. A masked diffusion language model is conditioned on the target embedding via adaptive layer normalization, requiring only 8 forward passes through a 78M parameter model with no access to the target encoder. On 32-token sequences across three embedding models, the method achieves 81.3% token accuracy and 0.87 cosine similarity.

январь 22, 2026
Embedding Compression via Spherical Coordinates
We present a compression method for unit-norm embeddings that achieves 1.5x compression, 25% better than the best prior lossless method. The method exploits that spherical coordinates of high-dimensional unit vectors concentrate around pi/2, causing IEEE 754 exponents to collapse to a single value and high-order mantissa bits to become predictable, enabling entropy coding of both. Reconstruction error is below 1e-7, under float32 machine epsilon. Evaluation across 26 configurations spanning text, image, and multi-vector embeddings confirms consistent improvement. The method requires no training.

ICLR 2026
декабрь 29, 2025
Vision Encoders in Vision-Language Models: A Survey
Vision encoders have remained comparatively small while language models scaled from billions to hundreds of billions of parameters. This survey analyzes vision encoders across 70+ vision-language models from 2023–2025 and finds that training methodology matters more than encoder size: improvements in loss functions, data curation, and feature objectives yield larger gains than scaling by an order of magnitude. Native resolution handling improves document understanding, and multi-encoder fusion captures complementary features no single encoder provides. We organize encoders into contrastive, self-supervised, and LLM-aligned families, providing a taxonomy and practical selection guidance for encoder design and deployment.

декабрь 04, 2025 • 7 минуты чтения
Jina-VLM: Маленькая многоязычная модель Vision Language Model
Новая модель vision language на 2B достигла SOTA в многоязычном VQA, без катастрофического забывания в задачах, связанных только с текстом.


декабрь 04, 2025
Jina-VLM: Small Multilingual Vision Language Model
We present jina-vlm, a 2.4B parameter vision-language model that achieves state-of-the-art multilingual visual question answering among open 2B-scale VLMs. The model couples a SigLIP2 vision encoder with a Qwen3 language backbone through an attention-pooling connector that enables token-efficient processing of arbitrary-resolution images. Across standard VQA benchmarks and multilingual evaluations, jina-vlm achieves leading results while preserving competitive text-only performance. Model weights and code are publicly released.

ICLR 2026
октябрь 03, 2025 • 7 минуты чтения
Jina Reranker v3: 0.6B Listwise Reranker для SOTA Мультиязычного Поиска
Новый списочный реранкер с 0.6B параметрами, который рассматривает запрос и все документы-кандидаты в едином контекстном окне.

