Late Chunking in Long-Context Embedding Models
* https://jina.ai/news/late-chunking-in-long-context-embedding-models/ (2024/8/22)
* https://x.com/JinaAI_/status/1826649439324254291 (2024/8/22)
* https://x.com/JinaAI_/status/1841903098174046494 (2024/10/4)
* https://jina.ai/news/what-late-chunking-really-is-and-what-its-not-part-ii/ (2024/10/4)
* 利用長上下文嵌入模型的力量來嵌入短分塊
* https://weaviate.io/blog/late-chunking (2024/9/5)
* https://x.com/JinaAI_/status/1854538252922802412 (2024/11/7)
* https://colab.research.google.com/drive/1iz3ACFs5aLV2O_uZEjiR1aHGlXqj0HY7?usp=sharing
* 和 [[Contextual Retrieval]] 比較: https://medium.com/kx-systems/late-chunking-vs-contextual-retrieval-the-math-behind-rags-context-problem-d5a26b9bbd38
> **封閉 API(如 OpenAI 的 embedding API)** 無法支援 Late Chunking,因為它通常只能直接返回 chunk-level embedding,而不能讓你獲取 transformer 中間層的 token embeddings