<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Day 4: Optimization and Scale on Qdrant - Vector Database</title><link>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/</link><description>Recent content in Day 4: Optimization and Scale on Qdrant - Vector Database</description><generator>Hugo</generator><language>en-us</language><managingEditor>info@qdrant.tech (Andrey Vasnetsov)</managingEditor><webMaster>info@qdrant.tech (Andrey Vasnetsov)</webMaster><atom:link href="https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/index.xml" rel="self" type="application/rss+xml"/><item><title>Vector Quantization Methods</title><link>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/what-is-quantization/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><author>info@qdrant.tech (Andrey Vasnetsov)</author><guid>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/what-is-quantization/</guid><description>&lt;div class="date">
 &lt;img class="date-icon" src="https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/icons/outline/date-blue.svg" alt="Calendar" /> Day 4 
&lt;/div>

&lt;h1 id="vector-quantization-methods">Vector Quantization Methods&lt;/h1>
&lt;div class="video">
&lt;iframe 
 src="https://www.youtube.com/embed/oExGyAEOpP4"
 frameborder="0"
 allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share"
 referrerpolicy="strict-origin-when-cross-origin"
 allowfullscreen>
&lt;/iframe>
&lt;/div>
&lt;p>Production vector search engines face an inevitable scaling challenge: memory requirements grow with dataset size, while search latency demands vectors remain in fast storage. &lt;a href="https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/documentation/guides/quantization/">Quantization&lt;/a> provides the solution by compressing vector representations while maintaining retrieval quality - but the method you choose fundamentally determines your system&amp;rsquo;s performance characteristics.&lt;/p>
&lt;h2 id="the-memory-economics">The Memory Economics&lt;/h2>
&lt;p>Consider the mathematics of scale. OpenAI&amp;rsquo;s &lt;code>text-embedding-3-small&lt;/code> produces 1536-dimensional vectors requiring 6 KB each (1536 × 4 bytes per float32). This scales predictably: 1 million vectors consume 6 GB, 10 million require 60 GB, and 100 million demand 600 GB of memory.&lt;/p></description></item><item><title>Accuracy Recovery with Rescoring</title><link>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/rescoring-oversampling-indexing/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><author>info@qdrant.tech (Andrey Vasnetsov)</author><guid>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/rescoring-oversampling-indexing/</guid><description>&lt;div class="date">
 &lt;img class="date-icon" src="https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/icons/outline/date-blue.svg" alt="Calendar" /> Day 4 
&lt;/div>

&lt;h1 id="accuracy-recovery-with-rescoring">Accuracy Recovery with Rescoring&lt;/h1>
&lt;div class="video">
&lt;iframe 
 src="https://www.youtube.com/embed/ksw3Ok-XXqo"
 frameborder="0"
 allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share"
 referrerpolicy="strict-origin-when-cross-origin"
 allowfullscreen>
&lt;/iframe>
&lt;/div>
&lt;br/>
When we use quantization methods like Scalar, Binary, or Product Quantization, we're compressing our vectors to save memory and improve performance. However, this compression can slightly reduce the accuracy of our similarity searches because the quantized vectors are approximations of the original data. To mitigate this loss of accuracy, you can use oversampling and rescoring, which help improve the accuracy of the final search results.
&lt;p>So let&amp;rsquo;s say we are performing a search in a collection with Binary Quantization. Qdrant retrieves the top candidates using the quantized vectors based on their similarity to the query vector, as determined by the quantized data. This step is fast because we&amp;rsquo;re using the quantized vectors.&lt;/p></description></item><item><title>Large-Scale Data Ingestion</title><link>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/large-scale-ingestion/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><author>info@qdrant.tech (Andrey Vasnetsov)</author><guid>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/large-scale-ingestion/</guid><description>&lt;div class="date">
 &lt;img class="date-icon" src="https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/icons/outline/date-blue.svg" alt="Calendar" /> Day 4 
&lt;/div>

&lt;h1 id="large-scale-data-ingestion">Large-Scale Data Ingestion&lt;/h1>
&lt;div class="video">
&lt;iframe 
 src="https://www.youtube.com/embed/Rawvm7TP1XI"
 frameborder="0"
 allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share"
 referrerpolicy="strict-origin-when-cross-origin"
 allowfullscreen>
&lt;/iframe>
&lt;/div>
&lt;br/>
&lt;p>In vector search applications inserting a few thousand data points is straightforward but the dynamics change completely when dealing with millions or billions of records. Tiny inefficiencies in the ingestion process compound into significant time losses, increased memory pressure, and degraded search performance.&lt;/p>
&lt;p>Every individual upsert call initiates a transaction that consumes memory and disk I/O to build parts of the index. At scale, this naive approach can overwhelm your system, causing upload times to spike and search quality to decrease. Efficiently preparing and loading your data into Qdrant is paramount for building a robust and scalable AI application.&lt;/p></description></item><item><title>Project: Quantization Performance Optimization</title><link>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/pitstop-project/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><author>info@qdrant.tech (Andrey Vasnetsov)</author><guid>https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/course/essentials/day-4/pitstop-project/</guid><description>&lt;div class="date">
 &lt;img class="date-icon" src="https://deploy-preview-2138--condescending-goldwasser-91acf0.netlify.app/icons/outline/date-blue.svg" alt="Calendar" /> Day 4 
&lt;/div>

&lt;h1 id="project-quantization-performance-optimization">Project: Quantization Performance Optimization&lt;/h1>
&lt;p>Apply quantization techniques to your domain search engine and measure the real-world impact on speed, memory, and accuracy. You&amp;rsquo;ll discover how different quantization methods affect your specific use case and learn to optimize the accuracy recovery pipeline.&lt;/p>
&lt;h2 id="your-mission">Your Mission&lt;/h2>
&lt;p>Transform your search engine from previous days into a production-ready system by implementing quantization optimization. You&amp;rsquo;ll test different quantization methods, measure performance impacts, and tune the oversampling + rescoring pipeline for optimal results.&lt;/p></description></item></channel></rss>