<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>DEV Community: Vector Podcast</title>
    <description>The latest articles on DEV Community by Vector Podcast (@vectorpodcast).</description>
    <link>https://dev.to/vectorpodcast</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Forganization%2Fprofile_image%2F7515%2F98fb0688-cd71-4a11-8f48-24e85b35f75f.jpeg</url>
      <title>DEV Community: Vector Podcast</title>
      <link>https://dev.to/vectorpodcast</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://dev.to/feed/vectorpodcast"/>
    <language>en</language>
    <item>
      <title>Novel idea in vector search: Wormhole vectors</title>
      <dc:creator>Dmitry Kan</dc:creator>
      <pubDate>Fri, 07 Nov 2025 06:16:39 +0000</pubDate>
      <link>https://dev.to/vectorpodcast/novel-idea-in-vector-search-wormhole-vectors-3dg8</link>
      <guid>https://dev.to/vectorpodcast/novel-idea-in-vector-search-wormhole-vectors-3dg8</guid>
      <description>&lt;p&gt;Vector Podcast episode: educational bit&lt;/p&gt;

&lt;p&gt;Two weeks ago I had a pleasure to co-host a lightning session together with Trey Grainger on a novel idea in vector search called "Wormhole vectors".&lt;/p&gt;

&lt;p&gt;The approach contrasts itself to a hybrid search approach.&lt;br&gt;
In hybrid search you would convert the same input query into different representations (keywords -&amp;gt; embeddings), run independent queries, and then combine the results.&lt;/p&gt;

&lt;p&gt;Wormhole vectors opens a new way to transcend vector spaces of different nature:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;Query in the current vector space&lt;/li&gt;
&lt;li&gt;Find a relevant document set&lt;/li&gt;
&lt;li&gt;Derive a "wormhole" vector to a corresponding region of another vector space&lt;/li&gt;
&lt;li&gt;Traverse to the other vector space with that query&lt;/li&gt;
&lt;li&gt;Repeat as desired across multiple traversals&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;More specifically, if you come from a sparse space, taking a set of returned documents you can pool their embeddings into one embedding and use it as a "wormhole" into the dense vector space.&lt;br&gt;
If as input you are dealing with a set of embeddings from a vector search, you can traverse the Semantic Knowledge Graph (SKG) to derive a sparse lexical query best representing these documents.&lt;/p&gt;

&lt;p&gt;Recording on YouTube:&lt;br&gt;


  &lt;iframe src="https://www.youtube.com/embed/fvDC7nK-_C0"&gt;
  &lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;What you'll learn:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;What are "Wormhole Vectors"?&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Learn how wormhole vectors work &amp;amp; how to use them to traverse between disparate vector spaces for better hybrid search.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Building a behavioral vector space from click stream data&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Learn to generate behavioral embeddings to be integrated with dense/semantic and sparse/lexical vector queries.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Traverse lexical, semantic, &amp;amp; behavioral vectors spaces&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Jump back and forth between multiple dense and sparse vector spaces in the same query&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Advanced hybrid search techniques (beyond fusion algorithms)&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Hybrid search is more than mixing lexical + semantic search. See advanced techniques and where wormhole vectors fit in.&lt;/p&gt;

&lt;p&gt;Find this episode on many popular platforms:&lt;br&gt;
Spotify: &lt;a href="https://open.spotify.com/episode/3fvfbAGQREqCJeUciyRb3r" rel="noopener noreferrer"&gt;https://open.spotify.com/episode/3fvfbAGQREqCJeUciyRb3r&lt;/a&gt;&lt;br&gt;


&lt;iframe src="https://open.spotify.com/embed/episode/3fvfbAGQREqCJeUciyRb3r" width="100%" height="232px"&gt;
&lt;/iframe&gt;


&lt;/p&gt;

&lt;p&gt;Apple Podcasts: &lt;a href="https://podcasts.apple.com/fi/podcast/trey-grainger-wormhole-vectors/id1587568733?i=1000735675696" rel="noopener noreferrer"&gt;https://podcasts.apple.com/fi/podcast/trey-grainger-wormhole-vectors/id1587568733?i=1000735675696&lt;/a&gt;&lt;br&gt;


&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://podcasts.apple.com/fi/podcast/trey-grainger-wormhole-vectors/id1587568733?i=1000735675696" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fis1-ssl.mzstatic.com%2Fimage%2Fthumb%2FPodcasts221%2Fv4%2F64%2Fd8%2F0d%2F64d80db7-46d0-acc7-3903-54ead5f99b82%2Fmza_14565107337929233133.png%2F1200x1200ECA.PESS01-60.jpg%3FimgShow%3DPodcasts122%2Fv4%2F26%2F00%2F20%2F260020ba-c79e-3a8d-6083-f55432f6cab1%2Fmza_15058363529978835570.jpg" height="auto" class="m-0"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://podcasts.apple.com/fi/podcast/trey-grainger-wormhole-vectors/id1587568733?i=1000735675696" rel="noopener noreferrer" class="c-link"&gt;
            Trey Grainger - Wormhole Vecto…–Vector Podcast – Apple Podcasts
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            Podcast Episode · Vector Podcast · 7 November 2025 · 1hr 19min
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fpodcasts.apple.com%2Fassets%2Ffavicon%2Ffavicon-32.png"&gt;
          podcasts.apple.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;




&lt;p&gt;RSS: &lt;a href="https://rss.com/podcasts/vector-podcast/2314900/" rel="noopener noreferrer"&gt;https://rss.com/podcasts/vector-podcast/2314900/&lt;/a&gt;&lt;br&gt;


&lt;/p&gt;
&lt;div class="crayons-card c-embed text-styles text-styles--secondary"&gt;
    &lt;div class="c-embed__content"&gt;
        &lt;div class="c-embed__cover"&gt;
          &lt;a href="https://rss.com/podcasts/vector-podcast/2314900/" class="c-link align-middle" rel="noopener noreferrer"&gt;
            &lt;img alt="" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fimg.rss.com%2Fvector-podcast%2F900%2Fep_cover_20251107_051156_724d3e0d493d36eed167f0604822b7e3.png" height="auto" class="m-0"&gt;
          &lt;/a&gt;
        &lt;/div&gt;
      &lt;div class="c-embed__body"&gt;
        &lt;h2 class="fs-xl lh-tight"&gt;
          &lt;a href="https://rss.com/podcasts/vector-podcast/2314900/" rel="noopener noreferrer" class="c-link"&gt;
            Trey Grainger - Wormhole Vectors | Podcast Episode on RSS.com
          &lt;/a&gt;
        &lt;/h2&gt;
          &lt;p class="truncate-at-3"&gt;
            This lightning session introduces a new idea in vector search - Wormhole vectors!It has deep roots in physics and allows for transcending spaces of any nature: sparse, vector and behaviour (but could theoretically be any N-dimensional space).Craft decaf &amp;amp;amp; half caf coffee, 25% discount: https://savorista.com/discount/VECTORBlog post on Medium: https://dmitry-kan.medium.com/novel-idea-in-vector-search-wormhole-vectors-6093910593b8Session page on maven: https://maven.com/p/8c7de9/beyond-hybrid-search-with-wormhole-vectors?utm_campaign=NzI2NzIx&amp;amp;amp;utm_medium=ll_share_link&amp;amp;amp;utm_source=instructorTo try the managed OpenSearch (multi-cloud, automatic backups, disaster recovery, vector search and more), go here: https://console.aiven.io/signup?utm_source=youtube&amp;amp;amp;utm_medium&amp;amp;amp;&amp;amp;amp;utm_content=vectorpodcastGet credits to use Aiven's products (PG, Kafka, Valkey, OpenSearch, ClickHouse): https://aiven.io/startupsTimecodes:00:00 Intro by Dmitry01:48 Trey's presentation03:05 Walk to the AI-Powered Search course by Trey and Doug07:07 Intro to vector spaces and embeddings19:03 Disjoint vector spaces and the need of hybrid search23:11 Different modes of search24:49 Wormhole vectors47:49 Q&amp;amp;amp;AWhat you'll learn:- What are "Wormhole Vectors"?Learn how wormhole vectors work &amp;amp;amp; how to use them to traverse between disparate vector spaces for better hybrid search.- Building a behavioral vector space from click stream dataLearn to generate behavioral embeddings to be integrated with dense/semantic and sparse/lexical vector queries.- Traverse lexical, semantic, &amp;amp;amp; behavioral vectors spacesJump back and forth between multiple dense and sparse vector spaces in the same query- Advanced hybrid search techniques (beyond fusion algorithms)Hybrid search is more than mixing lexical + semantic search. See advanced techniques and where wormhole vectors fit in.YouTube: https://www.youtube.com/watch?v=fvDC7nK-_C0
          &lt;/p&gt;
        &lt;div class="color-secondary fs-s flex items-center"&gt;
            &lt;img alt="favicon" class="c-embed__favicon m-0 mr-2 radius-0" src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fimg.rss.com%2Fvector-podcast%2F64%2Fep_cover_20251107_051156_724d3e0d493d36eed167f0604822b7e3.png"&gt;
          rss.com
        &lt;/div&gt;
      &lt;/div&gt;
    &lt;/div&gt;
&lt;/div&gt;




</description>
      <category>embeddings</category>
      <category>vectorsearch</category>
      <category>keywordsearch</category>
      <category>hybridsearch</category>
    </item>
    <item>
      <title>Vector Podcast: Simon Eskildsen, Turbopuffer</title>
      <dc:creator>Dmitry Kan</dc:creator>
      <pubDate>Fri, 19 Sep 2025 11:03:58 +0000</pubDate>
      <link>https://dev.to/vectorpodcast/vector-podcast-simon-eskildsen-turbopuffer-cfa</link>
      <guid>https://dev.to/vectorpodcast/vector-podcast-simon-eskildsen-turbopuffer-cfa</guid>
      <description>&lt;p&gt;&lt;em&gt;Vector database and search engine behind well known tools: Cursor, Notion, Linear, Superhuman, Readwise&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Vector Podcast is back — Season 4 starts now!&lt;/p&gt;

&lt;p&gt;I’ve had a pleasure to interview &lt;a class="mentioned-user" href="https://dev.to/sirupsen"&gt;@sirupsen&lt;/a&gt;, the creator of &lt;a href="https://turbopuffer.com/" rel="noopener noreferrer"&gt;Turbopuffer&lt;/a&gt; — vector database and a search engine. Prior to co-founding his startup, Simon has spent a decade at Shopify scaling the stateful engines, like Elasticsearch, Redis, MySQL at an immense scale (up to 1M reads / sec).&lt;/p&gt;

&lt;blockquote&gt;
&lt;p&gt;If you want to create a &lt;strong&gt;generational DB company&lt;/strong&gt;, I think you need 2 things: a) you need a &lt;strong&gt;new workload&lt;/strong&gt; — the new workload here is that we have almost every company on Earth, sitting on their treasure trove of data and they want to connect that to LLMs (large scale analytics on unstructured data) b) &lt;strong&gt;new storage architecture&lt;/strong&gt; — if you don’t have a new storage architecture, that is fundamentally a better trade-off for the particular workload, then there is no reason why tacking on a secondary index to your relational database, to your OLAP, to your existing search engine [would not work] — they would eat it.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F08v7b0bh3l35zibezu0s.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F08v7b0bh3l35zibezu0s.png" alt="Simon Eskildsen on Vector Podcast with Dmitry Kan"&gt;&lt;/a&gt; &lt;em&gt;Simon Eskildsen on Vector Podcast with Dmitry Kan&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Turbopuffer uses Object Storage (S3, GSC, Blob Storage) to store client data and vectors (encrypted using customer keys), and RAM and NVMe SSDs for cache of data your application actually uses. This is where the analogy to &lt;a href="https://en.wikipedia.org/wiki/Tetraodontidae" rel="noopener noreferrer"&gt;puffer fish&lt;/a&gt; finds its place: the databases inflates and deflates the cache, depending on the usage. To lower the latency towards Object Storage, Turbopuffer minimizes the number of roundtrips to 3–4 (400ms), and uses techniques like Range Fetch (available in &lt;a href="https://www.artofcode.org/blog/aws-s3-byte-range-fetch/" rel="noopener noreferrer"&gt;AWS&lt;/a&gt;, &lt;a href="https://cloud.google.com/storage/docs/samples/storage-download-byte-range" rel="noopener noreferrer"&gt;GCP&lt;/a&gt;, &lt;a href="https://learn.microsoft.com/en-us/rest/api/storageservices/specifying-the-range-header-for-blob-service-operations" rel="noopener noreferrer"&gt;Azure&lt;/a&gt;) to accelerate the cold read.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4duh6b0kktpe68ftk4ke.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4duh6b0kktpe68ftk4ke.png" alt="Turbopuffer architecture: https://turbopuffer.com/docs/architecture"&gt;&lt;/a&gt; &lt;em&gt;Turbopuffer architecture:&lt;/em&gt; &lt;a href="https://turbopuffer.com/docs/architecture" rel="noopener noreferrer"&gt;https://turbopuffer.com/docs/architecture&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;For vector search, &lt;a href="https://dl.acm.org/doi/10.1145/3600006.3613166" rel="noopener noreferrer"&gt;SPFresh&lt;/a&gt; algorithm is used, which is based on centroid index. This reminded me of the work I did as part of team Sisu implementing &lt;a href="https://github.com/DmitryKey/big-ann/tree/main/src/algorithms/sharding/kanndi" rel="noopener noreferrer"&gt;KANNDI algorithm&lt;/a&gt; for billion-scale ANN benchmarking competition. It is interesting, that centroid-based algorithms are natural for Object Storage as they minimize round-trips and allow for collocated writes. Looks like my intuition was right, and I’m glad someone has productized this!&lt;/p&gt;

&lt;p&gt;  &lt;iframe src="https://www.youtube.com/embed/I8Ztqajighg"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;All this to say, that the episode is full of insights, I’ve really enjoyed recording with Simon!&lt;/p&gt;

&lt;p&gt;Spotify: &lt;a href="https://open.spotify.com/episode/2DlO2W5QoLFVPvxwxyJaoN" rel="noopener noreferrer"&gt;https://open.spotify.com/episode/2DlO2W5QoLFVPvxwxyJaoN&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Apple Podcasts: &lt;a href="https://podcasts.apple.com/us/podcast/economical-way-of-serving-vector-search-workloads/id1587568733?i=1000727464303" rel="noopener noreferrer"&gt;https://podcasts.apple.com/us/podcast/economical-way-of-serving-vector-search-workloads/id1587568733?i=1000727464303&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;RSS: &lt;a href="https://rss.com/podcasts/vector-podcast/2222846/" rel="noopener noreferrer"&gt;https://rss.com/podcasts/vector-podcast/2222846/&lt;/a&gt;&lt;/p&gt;

</description>
      <category>vectordatabase</category>
      <category>rust</category>
      <category>objectstorage</category>
      <category>searchengine</category>
    </item>
    <item>
      <title>Adding ML layer to Search: Hybrid Search Optimizer</title>
      <dc:creator>Dmitry Kan</dc:creator>
      <pubDate>Sat, 12 Apr 2025 22:24:22 +0000</pubDate>
      <link>https://dev.to/vectorpodcast/adding-ml-layer-to-search-hybrid-search-optimizer-m0o</link>
      <guid>https://dev.to/vectorpodcast/adding-ml-layer-to-search-hybrid-search-optimizer-m0o</guid>
      <description>&lt;blockquote&gt;
&lt;p&gt;If we were to look 10 years out, I think an ideal solution is that we are not doing hybrid search anymore: just have a better approach. Something beyond vector + keyword, something better, that still supports the 0 results is the right answer (sometimes). We would have better approach, and not this slighly band-aidy approach, but for now hybrid search is exciting!&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;It is fascinating and funny how things develop, but also turn around. In 2022–23 everyone was buzzing about hybrid search. In 2024 the conversation shifted to RAG, RAG, RAG. And now we are in 2025 and back to hybrid search — on a different level. Finally, there are strides and contributions towards making hybrid search parameters learnt with ML. How cool is that?&lt;/p&gt;

&lt;p&gt;When I looked at hybrid search, I instantly knew that fiddling with a and b in a*keyword + b*neural will be the crux of succeeding with this approach to search. I also knew, that a better way than manual tweaking will be applying ML.&lt;/p&gt;

&lt;p&gt;I’m really happy someone clever did this. Daniel Wrigley and Eric Pugh, both from OpenSource Connections, decided to do exactly that: apply machine learning to the problem of computing these coefficients. In other words: what weight to give to keyword match vs neural search match. And what’s fascinating, is that they experimented with a multitude of methods from global to dynamic (per query), with different permutations, feature groups, combination methods, query sampling — it sounds an overwhelming study.&lt;/p&gt;

&lt;p&gt;What’s even cooler, is that all of this is open source.&lt;/p&gt;

&lt;p&gt;Check out this episode, let me (us) know what you think. And remember to subscribe to stay tuned for new episodes.&lt;br&gt;
  &lt;iframe src="https://www.youtube.com/embed/quY769om1EY"&gt;
  &lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;Design: Saurabh Rai, &lt;a href="https://www.linkedin.com/in/srbhr/" rel="noopener noreferrer"&gt;https://www.linkedin.com/in/srbhr/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The design of this episode is inspired by a scene in Blade Runner 2049. There’s a clear path leading towards where people want to go to, yet they’re searching for something.&lt;/p&gt;

&lt;p&gt;As usual, you can find the episode in audio form on your favorite platform.&lt;/p&gt;

</description>
      <category>federatedsearch</category>
      <category>hybridsearch</category>
      <category>neuralsearch</category>
      <category>keywordsearch</category>
    </item>
    <item>
      <title>7 AI Open Source Libraries To Build RAG, Agents &amp; AI Search</title>
      <dc:creator>𝚂𝚊𝚞𝚛𝚊𝚋𝚑 𝚁𝚊𝚒</dc:creator>
      <pubDate>Thu, 14 Nov 2024 21:08:14 +0000</pubDate>
      <link>https://dev.to/vectorpodcast/7-ai-open-source-libraries-to-build-rag-agents-ai-search-27bm</link>
      <guid>https://dev.to/vectorpodcast/7-ai-open-source-libraries-to-build-rag-agents-ai-search-27bm</guid>
      <description>&lt;h2&gt;
  
  
  What is Retrieval Augmented Generation (RAG)?
&lt;/h2&gt;

&lt;p&gt;Retrieval Augmented Generation (RAG) is an AI technique that combines searching for relevant information with generating responses. It works by first retrieving data from external sources (like documents or databases) and then using this information to create more accurate and context-aware answers. This helps the AI provide better, fact-based responses rather than relying solely on what it was trained on.&lt;/p&gt;

&lt;h2&gt;
  
  
  How does Retrieval Augmented Generation (RAG) Works?
&lt;/h2&gt;

&lt;p&gt;RAG (Retrieval-Augmented Generation) works by enhancing AI responses with relevant information from external sources. Here’s a concise explanation:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;When a user asks a question, RAG searches through various data sources (like databases, websites, and documents) to find relevant information.&lt;/li&gt;
&lt;li&gt;It then combines this retrieved information with the original question to create a more informed prompt.&lt;/li&gt;
&lt;li&gt;This enhanced prompt is fed into a language model, which generates a response that’s both relevant to the question and enriched with the retrieved information. This process allows the AI to provide more accurate, up-to-date, and context-aware answers by leveraging external knowledge sources alongside its pre-trained capabilities.&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqp5tiow7mrwykdggnce1.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fqp5tiow7mrwykdggnce1.png" alt="How does RAG Works" width="800" height="504"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  How does Retrieval Augmented Generation (RAG) helps the AI Model?
&lt;/h2&gt;

&lt;p&gt;RAG makes the AI more reliable and up-to-date by augmenting its internal knowledge with real-world, external data. RAG also improves an AI model in a few key ways:&lt;/p&gt;

&lt;ol&gt;
&lt;li&gt;
&lt;strong&gt;Access to Up-to-Date Information&lt;/strong&gt;: RAG retrieves relevant, real-time information from external sources (like documents, databases, or the web). This means the AI can provide accurate responses even when its training data is outdated.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Enhanced Accuracy&lt;/strong&gt;: Instead of relying solely on the AI’s trained knowledge, RAG ensures the model generates responses based on the most relevant data. This makes the answers more accurate and grounded in facts.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Better Contextual Understanding&lt;/strong&gt;: By combining retrieved data with a user’s query, RAG can offer answers that are more context-aware, making the AI’s responses feel more tailored and specific to the situation.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Reduced Hallucination&lt;/strong&gt;: Pure AI models sometimes "hallucinate" or make up information. RAG mitigates this by grounding responses in factual, retrieved data, reducing the likelihood of inaccurate or fabricated information.&lt;/li&gt;
&lt;/ol&gt;

&lt;h2&gt;
  
  
  7 Open Source Libraries to do Retrieval Augmented Generation
&lt;/h2&gt;

&lt;p&gt;Let's explore some open-source libraries helping you do RAG. These libraries provide the tools and frameworks necessary to implement RAG systems efficiently, from document indexing to retrieval and integration with language models.&lt;/p&gt;

&lt;h3&gt;
  
  
  1. SWIRL
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/swirlai/swirl-search" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjbx46wxs2fa74wk2pdum.png" alt="SWIRL" width="800" height="444"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;SWIRL is an open-source AI infrastructure software that powers Retrieval-Augmented Generation (RAG) applications. It enhances AI pipelines by enabling fast and secure searches across data sources without moving or copying data. SWIRL works inside your firewall, ensuring data security while being easy to implement.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;No ETL or data movement required.&lt;/li&gt;
&lt;li&gt;Fast and secure AI deployment inside private clouds.&lt;/li&gt;
&lt;li&gt;Seamless integration with over 20+ large language models (LLMs).&lt;/li&gt;
&lt;li&gt;Built for secure data access and compliance.&lt;/li&gt;
&lt;li&gt;Supports data fetching from 100+ applications.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/swirlai/swirl-search" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;⭐️ SWIRL on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h3&gt;
  
  
  2. Cognita
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/truefoundry/cognita" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fm9nwqlo42vnowcjogoic.png" alt="Cognita" width="800" height="335"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Cognita is an open-source framework for building modular, production-ready Retrieval Augmented Generation (RAG) systems. It organizes RAG components, making it easier to test locally and deploy at scale. It supports various document retrievers, embeddings, and is fully API-driven, allowing seamless integration into other systems.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Modular design for scalable RAG systems.&lt;/li&gt;
&lt;li&gt;UI for non-technical users to interact with documents and Q&amp;amp;A.&lt;/li&gt;
&lt;li&gt;Incremental indexing reduces compute load by tracking changes.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/truefoundry/cognita" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;⭐️ Cognita on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h3&gt;
  
  
  3. LLM-Ware
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/llmware-ai/llmware" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F1h4supvpf46rve3yp8mt.png" alt="LLM-Ware" width="800" height="334"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;LLM Ware is an open-source framework for building enterprise-ready Retrieval Augmented Generation (RAG) pipelines. It is designed to integrate small, specialized models that can be deployed privately and securely, making it suitable for complex enterprise workflows.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Offers 50+ fine-tuned, small models optimized for enterprise tasks.&lt;/li&gt;
&lt;li&gt;Supports a modular and scalable RAG architecture.&lt;/li&gt;
&lt;li&gt;Can run without a GPU, enabling lightweight deployments.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/llmware-ai/llmware" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;⭐️ LLMWare on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h3&gt;
  
  
  4. RAG Flow
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/infiniflow/ragflow" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F98cjqgzyjnr5vzmgj0zy.png" alt="RAG Flow" width="800" height="280"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;RagFlow is an open-source engine focused on Retrieval Augmented Generation (RAG) using deep document understanding. It allows users to integrate structured and unstructured data for effective, citation-grounded question-answering. The system offers scalable and modular architecture with easy deployment options.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Built-in deep document understanding to handle complex data formats.&lt;/li&gt;
&lt;li&gt;Grounded citations with reduced hallucination risks.&lt;/li&gt;
&lt;li&gt;Support for various document types like PDFs, images, and structured data.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/infiniflow/ragflow" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;⭐️ RAG Flow on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h3&gt;
  
  
  5. Graph RAG
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/microsoft/graphrag" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fyob0awzxn6pw3pgulbq6.png" alt="Graph RAG" width="800" height="363"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;GraphRAG is a modular, graph-based Retrieval-Augmented Generation (RAG) system designed to enhance LLM outputs by incorporating structured knowledge graphs. It supports advanced reasoning with private data, making it ideal for enterprises and research applications.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Uses knowledge graphs to structure and enhance data retrieval.&lt;/li&gt;
&lt;li&gt;Tailored for complex enterprise use cases requiring private data handling.&lt;/li&gt;
&lt;li&gt;Supports integration with Microsoft Azure for large-scale deployments.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/microsoft/graphrag" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;🌟 Graph RAG on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h3&gt;
  
  
  6. Haystack
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/deepset-ai/haystack" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F623od1u9e9fjstvfa0zm.png" alt="Haystack" width="800" height="356"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Haystack is an open-source AI orchestration framework for building production-ready LLM applications. It allows users to connect models, vector databases, and file converters to create advanced systems like RAG, question answering, and semantic search.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Flexible pipelines for retrieval, embedding, and inference tasks.&lt;/li&gt;
&lt;li&gt;Supports integration with a variety of vector databases and LLMs.&lt;/li&gt;
&lt;li&gt;Customizable with both off-the-shelf and fine-tuned models.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/deepset-ai/haystack" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;🌟 Haystack on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h3&gt;
  
  
  7. Storm
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://github.com/microsoft/graphrag" rel="noopener noreferrer"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fef6s5tek9l6tmdb7c7yt.png" alt="Storm" width="800" height="295"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;STORM is an LLM-powered knowledge curation system that researches a topic and generates full-length reports with citations. It integrates advanced retrieval methods and supports multi-perspective question-asking, enhancing the depth and accuracy of the generated content.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;What makes it unique:&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Generates Wikipedia-like articles with grounded citations.&lt;/li&gt;
&lt;li&gt;Supports collaborative human-AI knowledge curation.&lt;/li&gt;
&lt;li&gt;Modular design with support for external retrieval sources.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://github.com/stanford-oval/storm" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;🌟 Storm on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;h2&gt;
  
  
  Challenges in Retrieval Augmented Generation
&lt;/h2&gt;

&lt;p&gt;Retrieval Augmented Generation (RAG) faces challenges like ensuring data relevance, managing latency, and maintaining data quality. Some challenges are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Data relevance&lt;/strong&gt;: Ensuring the retrieved documents are highly relevant to the query can be difficult, especially with large or noisy datasets.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Latency&lt;/strong&gt;: Searching external sources adds overhead, potentially slowing down response times, especially in real-time applications.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Data quality&lt;/strong&gt;: Low-quality or outdated data can lead to inaccurate or misleading AI-generated responses.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Scalability&lt;/strong&gt;: Handling large-scale datasets and high user traffic while maintaining performance can be complex.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Security&lt;/strong&gt;: Ensuring data privacy and handling sensitive information securely is crucial, especially in enterprise settings.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Platforms like SWIRL tackle these issues by not requiring ETL (Extract, Transform, Load) or data movement, ensuring faster and more secure access to data. &lt;br&gt;
With &lt;a href="https://swirlaiconnect.com/" rel="noopener noreferrer"&gt;SWIRL&lt;/a&gt;, the retrieval and processing happen inside the user’s firewall, which helps maintain data privacy while ensuring relevant, high-quality responses. Its integration with existing large language models (LLMs) and enterprise data sources makes it an efficient solution for overcoming the latency and security challenges of RAG.&lt;/p&gt;
&lt;h3&gt;
  
  
  Thank you for reading 💜
&lt;/h3&gt;

&lt;p&gt;Thank you for reading my post and do take a look at these wonderful libraries. Share the post if you want to. I write about AI, open source tools, Resume Matcher and more.&lt;/p&gt;

&lt;p&gt;These are my handles where you can reach out to me:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://dev.to/srbhr" class="ltag_cta ltag_cta--branded"&gt;Follow me on DEV&lt;/a&gt;
&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.linkedin.com/in/srbhr/" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;Connect with me on LinkedIn&lt;/a&gt;
&lt;/p&gt;

&lt;p&gt;&lt;a href="https://github.com/srbhr" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;Follow me on GitHub&lt;/a&gt;
&lt;/p&gt;

&lt;p&gt;For collaborations send me an email at: &lt;a href="mailto:srbh077@gmail.com"&gt;srbh077@gmail.com&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwxkjiaqfwknr0pueaj9p.gif" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fwxkjiaqfwknr0pueaj9p.gif" alt="Thank you" width="480" height="480"&gt;&lt;/a&gt;&lt;/p&gt;

</description>
      <category>beginners</category>
      <category>ai</category>
      <category>python</category>
      <category>opensource</category>
    </item>
    <item>
      <title>Vector Podcast from Berlin Buzzwords’24: Sonam Pankaj, EmbedAnything</title>
      <dc:creator>Dmitry Kan</dc:creator>
      <pubDate>Fri, 20 Sep 2024 11:37:37 +0000</pubDate>
      <link>https://dev.to/vectorpodcast/vector-podcast-from-berlin-buzzwords24-sonam-pankaj-embedanything-46dh</link>
      <guid>https://dev.to/vectorpodcast/vector-podcast-from-berlin-buzzwords24-sonam-pankaj-embedanything-46dh</guid>
      <description>&lt;p&gt;I’ve just released an episode with Sonam Pankaj. She works on &lt;a href="https://github.com/StarlightSearch/EmbedAnything" rel="noopener noreferrer"&gt;EmbedAnything&lt;/a&gt;. We have recorded this episode at Berlin Buzzwords back in June, where I also got the chance to test my new audio recording gear (RØDE Wireless GO II).&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faf84ml1kxdqbql0z38tq.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/cdn-cgi/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Faf84ml1kxdqbql0z38tq.png" alt="Image description" width="" height=""&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;EmbedAnything is an infrastructure layer, that allows you to embed anything (different text formats, but also other modalities, like audio), written in Rust for performance reasons. It can embed a pdf text 40x faster than in Python.&lt;/p&gt;

&lt;p&gt;We spoke about this project, but also about metric learning, quality assurance and multimodality.&lt;/p&gt;

&lt;p&gt;There are a bunch of show notes with different papers and projects — do check them out.&lt;/p&gt;

&lt;p&gt;Find the episode on these platforms in addition to YouTube:&lt;/p&gt;

&lt;p&gt;RSS: &lt;a href="https://rss.com/podcasts/vector-podcast/1663042/" rel="noopener noreferrer"&gt;https://rss.com/podcasts/vector-podcast/1663042/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Spotify: &lt;a href="https://open.spotify.com/episode/5pUWz19iWKHqUzNT0JQ9KL" rel="noopener noreferrer"&gt;https://open.spotify.com/episode/5pUWz19iWKHqUzNT0JQ9KL&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Apple Podcasts: &lt;a href="https://podcasts.apple.com/fi/podcast/berlin-buzzwords-2024-sonam-pankaj-embedanything/id1587568733?i=1000670040161" rel="noopener noreferrer"&gt;https://podcasts.apple.com/fi/podcast/berlin-buzzwords-2024-sonam-pankaj-embedanything/id1587568733?i=1000670040161&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Patreon: &lt;a href="https://www.patreon.com/posts/vector-podcast-112350470?utm_medium=clipboard_copy&amp;amp;utm_source=copyLink&amp;amp;utm_campaign=postshare_creator&amp;amp;utm_content=join_link" rel="noopener noreferrer"&gt;https://www.patreon.com/posts/vector-podcast-112350470?utm_medium=clipboard_copy&amp;amp;utm_source=copyLink&amp;amp;utm_campaign=postshare_creator&amp;amp;utm_content=join_link&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Big thanks to &lt;a class="mentioned-user" href="https://dev.to/srbhr"&gt;@srbhr&lt;/a&gt; for designing the thumbnail of this episode.&lt;/p&gt;

</description>
    </item>
    <item>
      <title>These 8 Podcasts will help increase your knowledge and expand your mindset.</title>
      <dc:creator>𝚂𝚊𝚞𝚛𝚊𝚋𝚑 𝚁𝚊𝚒</dc:creator>
      <pubDate>Thu, 16 Nov 2023 12:10:29 +0000</pubDate>
      <link>https://dev.to/vectorpodcast/these-8-podcasts-will-help-increase-your-knowledge-and-expand-your-mindset-5lb</link>
      <guid>https://dev.to/vectorpodcast/these-8-podcasts-will-help-increase-your-knowledge-and-expand-your-mindset-5lb</guid>
      <description>&lt;p&gt;I’m a big fan of podcasts, the most underrated pure knowledge resources. Podcasts have emerged as a revolutionary medium in learning and entertainment. In recent years, they have transcended from being a niche listener’s hobby to a mainstream media format.&lt;/p&gt;

&lt;p&gt;Learning from two or more experienced people talking about a specific topic, discussing challenges they faced while doing something, and sharing their journey is fascinating. &lt;/p&gt;

&lt;p&gt;It gives incredible insights into tackling specific problems and building unique solutions. You get to know about:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Different mindsets and how other people think about things. &lt;/li&gt;
&lt;li&gt;The diverse array of tools that people use to solve problems. &lt;/li&gt;
&lt;li&gt;Prominent people and their stories and how they solve a particular problem.&lt;/li&gt;
&lt;li&gt;Come out of the mental bubble and start to see things differently.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;The other good thing is that the knowledge you acquire by listening to these podcasts becomes your latent knowledge or tacit knowledge reservoir.&lt;/p&gt;

&lt;p&gt;So, when you face challenges in a similar domain, the knowledge you gain from podcasts can give you a moment or a roadmap for building a solution to your challenge. &lt;/p&gt;

&lt;p&gt;And as an AI enthusiast and working in the knowledge and data domain. I picked up &lt;a href="https://www.youtube.com/@VectorPodcast" rel="noopener noreferrer"&gt;Vector Podcast&lt;/a&gt; because I was curious about vector databases and vector search, how they are helping companies establish artificial intelligence-based products, etc.&lt;/p&gt;

&lt;p&gt;Here are my top 8 podcast recommendations you can listen to. &lt;em&gt;I’m giving their YouTube channels, assuming everyone knows and uses YouTube. Most of them are also on Apple Podcasts, Spotify, etc.&lt;/em&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@VectorPodcast" rel="noopener noreferrer"&gt;Vector Podcast&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjii8stvlc2zhecit0b7p.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjii8stvlc2zhecit0b7p.jpg" alt="Vector Podcast" width="800" height="521"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Vector Podcast is the brainchild of Dmitry Kan—a Ph.D. Graduate Research Scientist venturing into Product Management &amp;amp; entrepreneurship. He also used to lecture at Helsinki University.&lt;/p&gt;

&lt;p&gt;With over 15 years of experience in the Search Engine Domain, Artificial Intelligence, and Software Development domains, he interviews professionals and CEOs in the field of AI.&lt;/p&gt;

&lt;p&gt;Some of the notable episodes are:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Searching with Swirl along with Sid Probstein, CEO of Swirl.&lt;/li&gt;
&lt;li&gt;Vector Databases along with Bob van Luijt, CEO of Weviate.&lt;/li&gt;
&lt;li&gt;Future of Search with Connor Shorten.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/@VectorPodcast" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;Subscribe to Vector Podcast&lt;/a&gt;
&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/c/lexfridman" rel="noopener noreferrer"&gt;Lex Fridman&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzb0jvrfnvqwa6khxhpil.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fzb0jvrfnvqwa6khxhpil.jpg" alt="Lex Fridman" width="800" height="501"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Lex Fridman is a popular figure, and because of this, he can host many famous figures, CEOs, leaders, and top-notch researchers. Bringing them to the mic-table and asking them questions (he’s pretty good at that.)&lt;/p&gt;

&lt;p&gt;Some notable people he interviewed were Sam Altman, Elon Musk, Guido van Rossum (Creator of Python Language), Stephen Wolfram, and a &lt;a href="https://www.youtube.com/watch?v=QqRV5FD8ob4" rel="noopener noreferrer"&gt;Goose&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;He’s also a Machine Learning Instructor, and you can view them in this &lt;a href="https://www.youtube.com/playlist?list=PLe8HThjUpqadLD-AewSKkhAyW5Nr4Yq4Z" rel="noopener noreferrer"&gt;playlist&lt;/a&gt;.&lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@tkppodcast" rel="noopener noreferrer"&gt;The Knowledge Project&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi7ix13d6kberdnex6suz.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fi7ix13d6kberdnex6suz.jpg" alt=" " width="800" height="432"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Hosted by Shane Parrish of Farnam Street, The Knowledge Project Podcast uncovers the best of what others have already figured out so you can apply their insights to your life.&lt;/p&gt;

&lt;p&gt;The podcast and the insights given by Shane Parrish are undoubtedly breathtaking. You have to experience the knowledge project to understand it well. &lt;/p&gt;

&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@syntaxfm" rel="noopener noreferrer"&gt;Syntax&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcul191m0c2jbdmf6knvg.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fcul191m0c2jbdmf6knvg.jpg" alt=" " width="800" height="525"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Comes from Wes Bos of 30 Days of JavaScript. This podcast talks about JavaScript and Web Development.&lt;br&gt;
It covers some fascinating topics ranging from newer features in web development to JavaScript testing and other topics. &lt;/p&gt;
&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@twimlai" rel="noopener noreferrer"&gt;This Week in Machine Learning (TWIML)&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsg5z96p0s26pw3i34n8n.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fsg5z96p0s26pw3i34n8n.jpg" alt="This week in machine learning" width="800" height="543"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Join Sam as he explores the latest trends and breakthroughs in AI, discusses the practical challenges of bringing AI-powered products to market, and examines the intersection of AI technology with business and consumer applications. &lt;/p&gt;

&lt;p&gt;This podcast is more than just a conversation; it's a gateway to understanding and leveraging the full potential of machine learning and AI to enhance our lives and communities. And I like this podcast a lot!&lt;/p&gt;
&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@Changelog" rel="noopener noreferrer"&gt;The Changelog&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffu44gmvwd13veekuukhn.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ffu44gmvwd13veekuukhn.jpg" alt="The Changelog" width="800" height="540"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Changelog is more than just a collection of podcasts; it's a dynamic community and a rich resource for developers at all stages of their journey.&lt;/p&gt;

&lt;p&gt;Whether you're a seasoned pro, a curious beginner, or somewhere in between, our shows and resources offer valuable insights, discussions, and stories that resonate with the developer experience.&lt;/p&gt;
&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@talkpython" rel="noopener noreferrer"&gt;Talk Python&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjzdyh4s3ow11c73p50f2.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fjzdyh4s3ow11c73p50f2.jpg" alt="Talk Python" width="800" height="548"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Talk Python to Me is a weekly podcast hosted by Michael Kennedy. The show covers various Python and related topics (e.g., MongoDB, AngularJS, DevOps).&lt;/p&gt;
&lt;h2&gt;
  
  
  &lt;a href="https://www.youtube.com/@BeyondCoding" rel="noopener noreferrer"&gt;Beyond Coding&lt;/a&gt;
&lt;/h2&gt;

&lt;p&gt;&lt;a href="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguwikoy09l2mv167l9wp.jpg" class="article-body-image-wrapper"&gt;&lt;img src="https://media2.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fguwikoy09l2mv167l9wp.jpg" alt=" " width="800" height="547"&gt;&lt;/a&gt;&lt;br&gt;
Patrick Akil and his guests share their journeys and perspectives for you to take with you and form your own.&lt;/p&gt;

&lt;p&gt;Beyond Coding is a weekly podcast with conversations that go "beyond coding" in a fireside chat format. Typical topics are software engineering, leadership, communication, self-improvement, and happiness.&lt;/p&gt;



&lt;p&gt;💡 &lt;strong&gt;&lt;em&gt;Fun fact:&lt;/em&gt;&lt;/strong&gt;&lt;br&gt;
Do you know that “podcast” comes from the blend of iPod and broadcast?&lt;/p&gt;
&lt;h2&gt;
  
  
  Conclusion
&lt;/h2&gt;

&lt;p&gt;I hope you liked the lists of podcasts featured in this article. If you have any suggestions, feel free to share them in the comments. &lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Checkout Vector Podcast on YouTube&lt;/strong&gt;&lt;br&gt;
We’re trying to provide you with the best-in-class content on AI, Large Language Models, and more. &lt;/p&gt;

&lt;p&gt;&lt;iframe width="710" height="399" src="https://www.youtube.com/embed/vhQ5LM5pK_Y"&gt;
&lt;/iframe&gt;
&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.youtube.com/@VectorPodcast" class="ltag_cta ltag_cta--branded" rel="noopener noreferrer"&gt;Subscribe to Vector Podcast&lt;/a&gt;
&lt;/p&gt;

&lt;p&gt;Thanks for reading.&lt;/p&gt;

</description>
      <category>webdev</category>
      <category>javascript</category>
      <category>programming</category>
      <category>python</category>
    </item>
  </channel>
</rss>
