SOTA Code Retrieval with Efficient Code Embedding Models — https://www.qodo.ai/blog/qodo-embed-1-code-embedding-code-retreival/
#HackerNews #SOTA #Code #Retrieval #Code #Embedding #AI #Technology #Machine #Learning

SOTA Code Retrieval with Efficient Code Embedding Models — https://www.qodo.ai/blog/qodo-embed-1-code-embedding-code-retreival/
#HackerNews #SOTA #Code #Retrieval #Code #Embedding #AI #Technology #Machine #Learning
#AI and #RAG - Learning the basics: What exactly is an #embedding and how to use them in #MariaDB?
https://www.youtube.com/watch?v=XkB2DLK60JU
GitHub - lancedb/lancedb: Developer-friendly, serverless vector database for AI applications. Easily add long-term memory to your LLM apps! https://github.com/lancedb/lancedb #persistence #OpenSource #embedding #database #GitHub #search #vector #ai
初めてのAI開発!ワクワクしながら作った問い合わせ対応チャットボット
https://qiita.com/SatoRyota_zvc/items/c5d647f5174ca8136bcb?utm_campaign=popular_items&utm_medium=feed&utm_source=popular_items
So basically face recognision is: compare current #embedding distance to database of embeddings and closest is considered a match?
#facerecognition #opencv
BedrockにCohereのマルチモーダル埋め込みモデルが来た!で、何に使うの?
https://qiita.com/moritalous/items/b881becb2c67d74446dd?utm_campaign=popular_items&utm_medium=feed&utm_source=popular_items
I’m excited to share my newest blog post, "Don't sure cosine similarity carelessly"
https://p.migdal.pl/blog/2025/01/dont-use-cosine-similarity
We often rely on cosine similarity to compare embeddings—it's like “duct tape” for vector comparisons. But just like duct tape, it can quietly mask deeper problems. Sometimes, embeddings pick up a “wrong kind” of similarity, matching questions to questions instead of questions to answers or getting thrown off by formatting quirks and typos rather than the text's real meaning.
In my post, I discuss what can go wrong with off-the-shelf cosine similarity and share practical alternatives. If you’ve ever wondered why your retrieval system returns oddly matched items or how to refine your embeddings for more meaningful results, this is for you!
`
I want to thank Max Salamonowicz and Grzegorz Kossakowski for their feedback after my flash talk at the Warsaw AI Breakfast, Rafał Małanij for inviting me to give a talk at the Python Summit, and for all the curious questions at the conference, and LinkedIn.
Damn, this is really cool, but I wish it had a big “pre-requisites” in the readme with “NVIDIA” in it #AI #RAG #Embedding #Documents #Ollama https://github.com/TilmanGriesel/chipper
Encoder only model that's a direct drop-in replacement for existing BERT models
- First major upgrade to BERT-style models in six years
- Significantly reduced processing costs for large-scale applications
- Enables longer document processing without chunking
- Better performance in retrieval tasks
- Suitable for consumer-grade GPU deployment
#llm #ai #embedding
https://huggingface.co/blog/modernbert
SQLite's Use Of Tcl (2017): I had no idea the database was originally written to be used as a TCL extension. Explains a lot of good things.
https://www.tcl.tk/community/tcl2017/assets/talk93/Paper.html
#via:lobsters #programming #embedding #sqlite #tcl #+
Ollama サーバーを立ててベクトル生成とベクトル検索を試してみた
https://qiita.com/yuji-arakawa/items/1f6ab3f68b98ff56de12?utm_campaign=popular_items&utm_medium=feed&utm_source=popular_items
Fine-tuning #embedding models clarifies enterprise semantics, business metrics, and ranking relevance prior to users issuing prompts.
https://thenewstack.io/the-secret-sauce-for-vector-search-training-embedding-models/
The current relevation that LLMs can’t reason is causing a lot of shade&fraud, but it’s not purely true
An LLM could reason, if you gave it a corpus of sentences (in whichever languages) which explicitly and unambiguously described a whole big bag of causal relationships and outcomes and things that happen because other things happen, and general structures such as that described clearly and formally and without any possibility of confusion
The embeddings which result from such a corpus could well work as a reference source of logic or cause or common sense or reason, about lots of things, and the next step would be to make it so that these embeddings are generalisable so that the common sense of the way life is, can be applied widely (again using vector comparison) so that yes it is possible to apply reason to a LLM, the main thing is that there probably isn’t an emphasis on that kind of descriptive and even prescriptive literature in and among the source learning in the first place – there’ll be a lot, there’ll be some, but I don’t think it was emphasised
By introducing it at the RAG level, and then the embeddings migrating back into the future models, I believe it could be possible to emulate a lot of common sense about the world and the way things are, purely through description of such – after all, the embeddings produced from such a block (a very massive block) of description, as vectors, are only numbers, which is what LLMs are really operating on, just vectors, not words, not tokens, just numbers
Consequently my dreams of applying real-world sensor/actuator ways of learning about the real world and building common sense are probably able to be supplanted just by a rigorous and hefty major project of just describing it instead of actually doing it – but the thing to watch would be in the description itself, it’d have to be as detailed and accurate and wide-ranging as the experiential model would be, and this might be where the difficulty lies, people describing common sense in the world would tend to abbreviate, generalise prematurely, miss things out, misunderstand, and above all, they’ll assume a lot #AI #LLM #reasoning #CommonSense #vector #embedding
Распределённый инференс llama.cpp через RPC
Приветствую, хабровчане! Идея создания данной публикации крутилась с моей голове уже давно, дело в том, что одно из моих хобби связанно с распределёнными вычислениями, а другое хобби связанно с нейросетями и мне давно не давала покоя идея запустить инференс LLM на нескольких компьютерах, но так чтобы все они выполняли работу над одно и той же моделью параллельно. Погуглив некоторое время узнал, что проект LocalAI уже относительно давно поддерживает такую возможность, недолго думая я раскатал на нескольких компьютерах данный проект, после чего выполнил все необходимые настройки связав все инстансы в единую систему и, мягко говоря, был разочарован, уж слишком "фатально-недостаточным" оказалось данное решение, Docker-образ собран неоптимально, он был огромный по весу и только под amd64 , неотключаемый веб-интерфейс шел в комплекте с проектом, скупой выбор моделей, некоторые из доступных LLM не работали в режиме RPC, все эмбеддинговые модели тоже отказывались запускаться в таком режиме, и так далее и тому подобное. Повозившись ещё немного, полез в исходники и обнаружил упоминание проекта llama.cpp , затем нашёл вызов бинарника rpc-server . И вот я оказался на странице llama.cpp/examples/rpc и всё заверте...
Jina Al just released Jina ColBERT v2, a Multilingual Late Interaction Retriever for #Embedding and #Reranking. The new model supports 89 languages with superior retrieval performance, user-controlled output dimensions, and 8192 token-length.
The server test has enabled users who embed content to skirt copyright infringement. However, the 2007 ruling faces another major challenge.
https://www.plagiarismtoday.com/2024/08/08/the-server-test-suffers-a-major-blow/
#Development #Techniques
External, styleable, and scalable SVGs · SVG embeddings that leave little to be desired https://ilo.im/15zn1a
_____
#VectorGraphic #SVG #Embedding #WebPage #WebDev #Frontend #HTML #CSS #CustomProperty
#Development #Pitfalls
YouTube embeds are bananas heavy · Lighter ways to add YouTube videos on your website https://ilo.im/15zdd6
_____
#Video #Youtube #Embedding #WebComponent #ProgressiveEnhancement #WebPerf #WebDev #Frontend #HTML #JavaScript
Gave https://ollama.com/avr/sfr-embedding-mistral a spin but took way to long (+3hours) to generate 5K embeddings on my m3 pro (32gb).. #llm #embedding #ollama
Can #AI understand emotions?
I dove deep into #LLM #embedding models to extract emotional signals in the text, because why not? And yeah, some models capture more than others, but all of them have it to some extent.