Release Notes
Find out what’s new in Milvus! This page summarizes new features, improvements, known issues, and bug fixes in each release. You can find the release notes for each released version after v2.3.0 in this section. We suggest that you regularly visit this page to learn about updates.
v2.3.21
Release data: August 14, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.21 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
Miluvs v2.3.21 addresses a critical bug in data deletion (#35390) and introduces some improvements in metrics and performance.
Improvements
- Enhanced data deletion process by modifying the delegator’s delete buffer handling (#35074)
- Added metrics to track disk quota usage (#35320)
- Reduced delegator memory usage by adjusting the overloaded factor to 0.1 (#35165)
- Improved health check by skipping manually stopped components (#35122)
- Added metric to record maxinsertrate and querynodememoryhighwaterlevel (#35193)
- Upgraded the Conan version (#35217)
- Changed the default metric type for autoindex (#34328)
- Sped up the datacoord gc quit process (#35057)
- Avoided panic due to nil schema (#35065)
- Reduced duplicate primary keys in the segcore component (#35291)
- Tracked the number of times Milvus enters a force-deny-writing state (#34990)
Bug fixes
- Resolved an issue with the BloomFilter returning false negatives when the statslog contained multiple K values, which could occur when a delete operation failed to apply (#35390)
- Enabled a limiter for RESTful server (#35354)
- Added a retry mechanism to avoid getting incomplete query results (#35061)
- Fixed issues with delete-by-expression operations failing during retry progress (#35421)
- Addressed a problem where the datanode binlog size metrics included dropped segments (#35420)
- Resolved comparison operations between incompatible operands (#35308)
- Fixed an issue where the datanode could not watch the channel (#35136)
- Addressed a deadlock issue on compaction when stopping the datanode (#35199)
v2.3.20
Release data: July 25, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.20 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
Milvus v2.3.20 brings several important bug fixes focused on enhancing segment management. This release also includes an upgrade to Go version 1.21, which allows the pprof tool to capture C++ code profiles, significantly aiding in diagnosing Milvus. Additionally, the Knowhere version has been updated to v2.2.7, enabling Milvus to be compiled with glibc versions ≤ 2.30. A new metric has been introduced to record the duration that requests wait in the proxy queue, improving Milvus’s observability.
Features
- Added a metric to record the duration of requests waiting in the proxy queue (#34791).
Improvements
Bug fixes
- Initialized the hook when creating the proxy object (#34936).
- Fixed a parsing issue for the plan proto for search types (#34945).
- Resolved an issue where segments could bounce between delegator and worker (#34889).
- Prevented the generation of reduce channel tasks before channel balancing is complete (#34721).
- Avoided segment shortages caused by deduplicated segment tasks (#34882).
- Discarded compaction plans when closing DataSyncService (#34737).
- Fixed a typo in the compatible ascending index (#34712).
v2.3.19
Release date: July 15, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.19 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
Milvus v2.3.19 introduces significant improvements and bug fixes to enhance memory management, error handling, and query performance. This release continues to bolster the robustness and functionality of Milvus, ensuring optimal performance and user experience.
Improvements
- Preserved fixed-size memory in delegator node for growing segments (#34602).
- Refined Prometheus Buckets at different scales to improve monitoring precision (#34627).
- Avoided assigning too many segments/channels to new QueryNode, ensuring balanced load distribution (#34461).
- Provided clearer error messages when the string/varchar exceeded its max length (#34324, #34034).
- Prevented memory slice increase during queries, optimizing memory usage (#34256).
- Sped up the loading process for small collections, reducing latency (#33863).
Bug fixes
- Fixed error message when field name is invalid, improving user feedback (#33901).
- Ensured Segment Manager filters flushed segments correctly during import (#34650).
- Made knowhere-build-pool-size configurable on QueryNode for better resource management (#34647).
- Eliminated unnecessary vchannels’ merge when curTS is 0, optimizing performance (#34626).
- Fixed issue where IndexNode couldn’t be stopped due to a missing lifetime end (#34560).
- Enhanced RESTful API query to be more user-friendly (#34447).
- Checked load state of all partitions instead of just the first one, ensuring completeness (#34321).
- Implemented singleflight for segcore chunkcache to avoid redundant operations (#34284).
- Broadcasted collection’s new properties to DataCoord, ensuring consistency (#34147).
- DataCoord now returns checkpoint after successfully flushing segments (#34115).
- Checked nodeid wildcard when removing pkoracle, preventing potential errors (#34022).
- Ensured queries by primary key return the latest row rather than historical data, maintaining data accuracy (#34026).
v2.3.18
Release date: June 19, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.18 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
Milvus v2.3.18 includes several bug fixes and improvements to enhance query result quality and maintainability. This release also backports improvements and fixes from the 2.4 branch, which enhance control over delete data amounts (according to rate limit configuration) and increase Bloom filter processing speed. As with previous releases, this update reinforces Milvus’s commitment to functionality and robustness.
Improvements
- Added RESTful APIs to trigger component stop (#33798).
- Improved delete by expression to respect delete rate limit rules according to the matched data size instead of delete request size (#33794).
- Accelerated Bloom filter processing by submitting in batch and working in parallel (#33870).
- Enabled parallel processing for
applydelete
at the segment level (#33841). - Printed
UseDefaultConsistency
parameter in read requests (#33645).
Bug fixes
- Fixed a bug where closure capture iteration variable caused delete data to fail to apply to the correct data (#33912).
- Fixed a bug where filtered results could be less than topk by upgrading Knowhere to 2.2.10 (#33930).
- Fixed a bug where insert rate was not limited when
collection.insertrate.max.mb
config was set to 0 (#33725). - Fixed a bug where old collections created before v2.2.8 could reappear and become zombies after being dropped and restarted (#33695).
- Fixed a bug where upsert latency used the wrong unit, causing abnormal metrics (#33580).
- Fixed a bug where Milvus exception info could not be passed outside of segcore (#33395).
v2.3.17
Release date: May 31, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.17 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
Milvus version 2.3.17 primarily introduces several bug fixes to improve query accuracy, cluster reliability, and privilege maintenance. The most notable fix ensures that query iterators in Milvus return the correct number of results, rather than missing some data. This release reinforces Milvus’s commitment to functionality and robustness.
Bug fixes
- Fixed issue where query iterators missed some results when upserts and deletes were performed very frequently (#33468)
- Fixed issue where the cluster could block during startup due to improper handling of standby status (#33371)
- Fixed issue where specifying the database name when granting privileges failed (#33292)
v2.3.16
Release date: May 24, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.16 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
Milvus v2.3.16 focuses on stability, performance, and security. It includes bug fixes for auto-balancing, checkpoint handling, memory leaks, and transaction limits. Additionally, authentication for the REST v2 API has been strengthened, and query result retrieval has been optimized. Overall, this update aims to provide a more reliable, efficient, and secure platform for your needs.
Improvements
- Enhanced authentication for the REST v2 API.(#33254)
- Improved checkpoint handling for message failures.(#33159)(#33128)
- Improved the triggering mechanism for segment sync tasks so that it will only be triggered once by flushts (#33034)
- Improved the query result retrieval mechanism to get enough query results (#33177)
Bug fixes
- Fixed a bug that when suspending auto balance querynode will unexpectedly exit stopping balance (#32941)
- Resolved an issue leading to endless binlog path downloads during compaction (#33246)
- Addressed memory leaks in rendezvousflushmanager (#33112)
- Prevented system panics related to nil array field data (#33119)
- Fixed “SegmentNotLoaded” errors triggered by concurrent segment tasks in search (#33087)
- Fixed a bug that Etcd txn exceeds limit due to too many fields (#33049)
- Fixed a bug related to the incorrect calculation of the milvus_rootcoord_indexed_entity_num metric (#33002)
v2.3.15
Release date: May 11, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.15 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
This release comes with several enhancements and bug fixes to improve security and stability. One of the improvements is the inclusion of the channel checkpoint in the flush response. Another enhancement is the ability to check whether autoID is enabled during data insertion, which optimizes the data migration using Milvus-CDC.
The bug fixes address issues such as empty segments in meta, and incorrect data types during data insertion according to the schema. A critical patch has also been implemented to solve the consistent failure and retries of the syncDistribution request to the delegator. We highly recommend updating to the latest version to benefit from these fixes.
Improvements
- Supported return channel checkpoint info in flush response (#32683)
- Add config to check id during inserting data when autoID enabled (#32840)
Critical Bug Fixes
- Fixing a bug where the syncDistribution request to the delegator consistently fails and retries persistently. (#32923)
Other Bug Fixes
- Fixed a bug where the system could generate empty segments. Added support for the compactor to clean up empty segments. (#32690)
- Fixed a bug that involved validating whether the input data types match the schema. (#32790, #32845)
v2.3.14
Release date: Apr 29, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.14 | 2.3.7 | 2.3.6 | 2.3.6 | 2.3.5 |
This release focuses on improving system performance, reliability, and observability with key features. It supports RESTful APIs to execute rolling upgrades, implements task-driven collection observation, allows for configurable intervals for garbage collection (GC), and enhances the check logic.
Furthermore, it fixes bugs in RESTful v2, resolving issues such as parameter parsing errors and ineffective parameter handling. Critical bug fixes related to deadlocks and data accuracy have also been included. Additionally, performance improvements have been made when handling large numbers of segments.
Overall, these updates aim to provide a more secure, efficient, and reliable user experience.
Improvements
- Added configuration items to skip Auto ID and Partition Key checks to improve check speed.(#32671)
- Implemented the collection observer of QueryCoordV2 task-driven.(#32615)
- Supported the get sdk type by user agent in access log.(#32554)
- Supported the use of different intervals for GC scan.(#32551)
- Removed the support for the always-true expression in delete expr.(#32495)
- Allowed users to disable search optimization.(#32143)
- Added restful api for devops to execute rolling upgrade.(#31846)
Bug fixes
- Fixed the bug that the calculation precision may be different between growing index and sealed index.(#32584)
- Fixed the bug that grow task stuck when node become stopping.(#32556)
- Fixed the issue that leader view may not update when segment version did not match.(#32517)
- Fixed the bug that value in elementTypeParams not be integer in restful api.(#32362)
- Fixed dbName parameter of describeAlias request not take effect and fixed the issue which will report lack of permission when not pass the whole privileges to describe collection in restful api. (#32162)
- Fixed cp lag metrics leakage by marking channel checkpoint dropped.(#32454)
- Fixed the bug where not update segment’s version in syncDistribution.(#32320)
- Fixed a bug that coordinator register blocked on ProcessActiveStandby.(#32133)
- Fixed a bug that the reduce didn’t handle offset without limit and reduceStopForBest correctly.(#32087)
- Fixed a bug that not validata PlaceholderGroups before combine them.(#32045)
- Fixed a bug where acquire index meta’s lock for each segment.(#31798)
- Fixed a bug that Rootcoord’s stop may block in quota_center’s stop.(#31824)
v2.3.13
Release date: Apr 7, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.13 | 2.3.7 | 2.3.5 | 2.3.6 | 2.3.5 |
This release aims to enhance system performance, reliability and observability with key features like improved index information retrieval, secure Minio connections via TLS, and new metrics for better monitoring.
This release has enriched the RESTful interfaces, allowing users to perform operations related to Partition and Index. Additionally, this release includes some critical bug fixes related to deadlocks and data integrity. Check the RESTful API reference for more details.
These updates aim to provide a more secure, efficient, and reliable user experience.
Improvements
- Optimized DescribeIndex performance by implementing bulk index information retrieval(#31239)(#31429)
- Improved access efficiency to formatted keys by caching the formatted key of param item(#31402)
- Add metrics for querycoord current target cp lag(#31463)
- Implemented TLS support for Minio connections(#31292)(#31619)
- Speed up target recovery after QueryCoord restart(#31449)
- Added new metrics for entities statistics(#31511)
- Implementing the API of the resultful(#30430)
- Save collection targets by batches(#31655)
- Added validation checks for the legality of field data types(#31699)
Bug Fixes
- Fixed the issue in CurrentTargetFirst/NextTargetFirst where it would generate unexpected task behavior(#31419)
- Fixed the bug that does not cause the shard leader to fail when search fails(#31450)
- Fixed a bug where nodes were marked as unreachable when get client failure(#31451)
- Fixed client error handling to enable retries for certain types of unrecoverable errors instead of directly returning them(#31452)
- Fixed an issue where the balance channel would freeze due to deadlocks(#31455)
- Fixed a bug where nodeID was not checked when updating channel checkpoints(#31508)
- Fixed the incorrect use of double buffering for entry size exceeding the max size(#31549)
- Fixed a bug where delegator’s filtering removed all delete messages, making deleted messages temporarily accessible(#31587)
- Fixed a bug where the proxy’s context was canceled prematurely during session revocation(#31595)
- Fixed the target observer updates manual and automatic tasks were not mutually exclusive(#31603)
- Fixed a bug where channel metadata was updated even when nodeID did not match(#31665)
- Fixed a bug where compaction tasks were not cleaned up when releasing channels(#31694)
- Fixed a bug where the injection was invalid when the queue was empty(#31819)
- Fixed a bug where channel meta mergeFlushSegment was not idempotent cause data loss(#31837)
- Fixed the bug that Datacoord session disconnects before revoking it(#31747)
v2.3.12
Release date: Mar 15, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.12 | 2.3.7 | 2.3.4 | 2.3.6 | 2.3.5 |
Milvus v2.3.12 is a bug-fix release that resolves data loss in some edge cases and includes several other minor fixes. The enhancements address the high memory usage of coordinators in instances with a large number of segments and the ARM image page size issue.
It is highly recommended to upgrade to this latest version to prevent data loss.
Bug Fixes
- [Critical] Fixed a bug that some insert data would be lost when deleteBuf memory policy is triggered (#31159)
- Fixed a bug that some duplicated segments couldn’t be released after channel balancing (#31126)
- Fixed a bug that the balance checker of the Querycoord used the wrong param item for check interval (#31141)
- Fixed a bug that database info could be incorrect when describing collection with ID (#31177)
Improvements
- Change pagesize to 64k for AARCH64 platform (#31114)
- Introduce an internal API to list indexes meta and reduce memory usage during load collection (#31150)(#31163)
v2.3.11
Release date: Mar 8, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.11 | 2.3.6 | 2.3.4 | 2.3.5 | 2.3.5 |
Milvus v2.3.11 brings improvements and bug fixes aimed at enhancing performance, security, and stability. The improvements include optimized data loading, enhanced security with TLS support for Kafka connections, memory optimization, and more. Additionally, bug fixes address issues such as search/query failures, incorrect data types in outputs, and disk estimation errors. We encourage you to update to this latest version to take advantage of these enhancements and fixes.
Improvements
- Optimized JSON loading by reducing 1x memory copy (#30864)
- Implemented TLS support for Kafka connections (#30466 #30925)
- Optimized monitoring: Remove time tick delay metrics when nodes go offline (#30879)
- Optimized memory usage and loading speed for variable length data (#30900)
- Implemented support for varchar autoID in bulk insert operations (#30913)
- Added support for rate limiting for flush operations at the collection level (#29568)
- Improved compatibility for the watch DM channel request (#30954)
- Enhanced stability when reading S3 objects (#30976)
- Reduced contention for locks in the DescribeIndex operation (#30975)
- Optimized automatic balancing in QueryCoord (#30725)
- Enabled Milvus containers to run as a non-root user (#30937)
- Enhanced lock granularity in DataCoord meta (#30986)
- Enhanced memory estimation during loading by adding a load memory factor (#30999)
- Reduced memory consumption for DataNode when dealing with multiple collections (#30991)
- Optimized the connection manager in the proxy to prevent out-of-memory errors (#31009)
Bug Fixes
- Fixed search/query failed caused by passing an incorrect context and missing error handling (#30818)
- Fixed incorrect data types in the output when using the RESTful query interface (#30738)
- Fixed incorrect disk estimation when loading disk index (or index with mmap enabled) (#30948)
- Fixed the issue where redundant segments with older versions may never be released (#30953)
- Fixed the issue where incorrect disk limits may be read (#30966)
- Fixed a series of issues with the RESTful interface, such as incorrect int64 precision and failed insertion (#30873)
- Fixed panic during rolling upgrade caused by compatibility issues (#30656)
- Fixed an issue where flush and compaction were stalled due to uncleared compaction tasks in the Datanode (#30972)
- Fixed frequent lock contention in the GetCompactionTo operation, which caused channel checkpoints lag (#30965)
- Fixed the issue where Datanode failed to update channel checkpoints with multiple collections (#31024 #31082)
- Fixed insertion failures caused by frequent contention for locks (#31026)
- Fixed proxy panic caused by missing error handling (#31086)
v2.3.10
Release date: Feb 23, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.10 | 2.3.6 | 2.3.4 | 2.3.5 | 2.3.5 |
Milvus v2.3.10 is a critical patch release that follows Milvus v2.3.9, featuring several essential bug fixes. A significant fix in this release addresses the issue of missing data in hybrid search results when utilizing partition keys.
Users who have encountered hybrid search issues in release v2.3.5 through v2.3.9, especially those using partition key features, are highly recommended to upgrade to version v2.3.10 promptly.
Critical Bug Fixes
- Fixed the upsert error where the primary key field was incorrectly treated as the PartitionKey field (Issue #30607).
- Prevented flush blockages by skipping the filling of segmentID in indexBuildCh (#30749).
Bug Fixes
- Prevented accidental deletion of original data by avoiding the use of absolute paths in ChunkCache (#30679).
- Enforced the use of virtual host for Tencent Cloud object storage (#30685).
- Updated disk usage metrics following segment release (#30707).
- Released loaded growing segments if WatchDmlChannel fails (#30745).
- Resolved a panic caused by using a nil interface in the gRPC client (#30755).
v2.3.9
Release date: Feb 19, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.9 | 2.3.6 | 2.3.4 | 2.3.5 | 2.3.5 |
Milvus 2.3.9 is a critical patch release succeeding Milvus 2.3.8, incorporating a number of vital bug fixes. Key among these is the resolution of the DiskANN construction failure issue. We strongly advise users experiencing disk index construction problems in version 2.3.8 to upgrade to version 2.3.9 without delay.
Critial Bug Fixes
- Fixed DiskANN construction failure in Milvus version 2.3.8 (#30640).
- Addressed loading failures due to inability to replace the primary key index (#30578).
- Added
collectionName
to the response of the ListAliases API (#30533).
v2.3.8
Release date: Feb 7, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.8 | 2.3.6 | 2.3.4 | 2.3.5 | 2.3.5 |
Milvus v2.3.8 is a minor patch release following Milvus v2.3.7. This release includes several enhancements and bug fixes. These improvements are designed to enhance system stability and observability. One of the key bug fixes is to prevent BulkInsert from getting stuck after a node restart.
Improvements
- Improved error messaging for dimension mismatch in search vectors (#30316)
- Integrated Milvus build process details, including commit information and dependency identifiers, into monitoring metrics (#29666)
- Streamlined loading strategy for segment Binlog files to optimize performance (#30348)
- Expanded BulkInsert feature to accommodate auto-incrementing primary keys for VarChar types (#30448)
- Enhanced memory estimation algorithm during data loading to prevent out-of-memory (OOM) errors (#30475)
- Eliminated extraneous log messages for cleaner logging (#30478)
- Updated to Knowhere version 2.2.4 for improved functionality (#30513)
Critical Bug Fixes
- Fixed panic error caused by watching multiple channels in the Datanodes (#30136)
- Corrected reading of index parameters from the configuration file (#30353)
- Ensured effectiveness of the db_name parameter for DescribeAlias and ListAliases operations (#30453)
- Resolved proxy startup hang-up due to improper port occupation handling (#30416)
- Refactored BulkInsert Flush process to prevent hang-ups after restart (#30439)
v2.3.7
Release date: Jan 29, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.7 | 2.3.6 | 2.3.4 | 2.3.5 | 2.3.5 |
Milvus v2.3.7 marks a minor yet impactful update, concentrating on boosting overall functionality and stability. In this release, we have refactored the business logic for a graceful stop to prevent any data loss, introduced support for array and JSON data types through RESTFul APIs, and amped up the speed of index loading. Alongside these enhancements, we’ve made several tweaks to optimize system performance and resource management. Additionally, this release addresses critical bug fixes as well as issues like memory leaks, load timeouts, and service unavailability, ensuring a more reliable and stable user experience.
Features
Limit Collection Counts
A Milvus instance allows up to 65,536 collections. However, too many collections may result in performance issues. Therefore, it is recommended to limit the number of collections created in a Milvus instance. Read more on Limit Collection Counts.
Chunk Cache
The chunk cache mechanism enables Milvus to pre-load data into cache memory on the local hard disk of the query nodes before it is needed. This mechanism significantly improves vector retrieval performance by reducing the time it takes to load data from disk to memory. Read more on Configure Chunk Cache
Improvements
- Transform specific magic numbers into configurable options (#30070).
- Remove heartbeat delay logic for ShardLeader to prevent misjudging its availability (#30085).
- When allocating channels, shuffle ShardLeader candidates to avoid load imbalance (#30089).
- Enhance RESTful support by adding functionality for arrays and JSON (#30077).
- Add a counter monitoring for rate-limited requests (#30132).
- Accelerate index loading through concurrent methods (#30018).
- Remove the step of DataNode subscribing to the message stream during the Import phase to avoid Import timeouts (#30133).
- Introduce association logic between privileges to simplify the authorization process (#30154).
- Implement unified restrictions on the number of Collections, Partitions, and Shards (#30017).
- Incorporate proactive pre-warming logic for ChunkCache to mitigate the issue of high latency when retrieving raw vectors during cold start queries (#30289)
- Optimize the load balancing algorithm by assigning weight to growing segments (#30293)
- Remove unnecessary business logic for conversions between partition names and IDs to reduce latency in the data retrieve stage when processing search requests (#30255)
Critical Bug Fixes
- Fixed a memory leak caused by incorrect usage of OpenTelemetry in the Segcore (#30068).
- Addressed the issue of slow disk index loading by dynamically patching the index parameters (#30116).
- Resolved the problem of changes made through the “alter collection” command not being persisted (#30156).
- Fixed the issue where read request rate limiting ultimately leads to the unavailability of the read service (#30196).
- Resolve the deadlock issue when getting the configuration (#30319)
- Fix incorrect usage of the Goroutine pool on CGO calls (#30275)
- Add a timeout mechanism to the graceful stop process to prevent potential cases of getting stuck(#30320)
v2.3.5
Release date: Jan 17, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.5 | 2.3.5 | 2.3.4 | 2.3.5 | 2.3.5 |
Milvus v2.3.5 marks a significant update focused on improving overall functionality and stability. In this release, we’ve made enhancements to Role-Based Access Control (RBAC) and Alias, prioritizing user-friendliness. Additionally, various improvements have been implemented to optimize system performance and resource management. A notable feature in v2.3.5 is the reintroduction of the MVCC (Multi-Version Concurrency Control) mechanism, crucial for efficiently managing concurrent and phased queries. This release also addresses critical bug fixes, preventing issues related to data deletion loss, abnormal system panics, and deadlocks, ensuring a more reliable and stable user experience.
Features
- Role-Based Access Control (RBAC)
Improvements:
- Restore MVCC functionality (#29749)
- Add concurrency for DataCoord segment GC to increase garbage collection speed (#29557)
- Read Azure files without ReadAll to control memory usage (#29604)
- Support reading hardware metrics for cgroupv2 (#29847)
- Save lite WatchInfo into etcd in DataNode (#29751)
- Support access log printing cluster prefix (#29646)
- Rewrite generation segment plan based on assigning segments to make it more understandable (#29574)
- Performance:
Critical Bug Fixes:
- Fix the delete message disorder issue causing data loss (#29917)
- Throw an exception when the upload file fails for DiskIndex to avoid core dump (#29628)
- Fix dynamic update of rate limit config with incorrect value (#29902)
- Ensure compact operation on DataCoord meta performs as a transaction (#29776)
- Fix panic caused by type assert LocalSegment on Segment (#29018)
- Drop segment meta info with a prefix to avoid etcd txn out of limit (#29857)
- Remove unnecessary lock-in config manager (#29855)
- Rectify memory leaks when reading data from Azure. (#30006)
- Resolve the issue of mistakenly duplicating dynamic fields when handling RESTful insert requests. (#30043)
- Rectify the deadlock issue in the BlockAll operation within the flowgraph. (#29972)
- Resolve the issue of failing to parse lengthy and complex expressions. (#30021)
v2.3.4
Release date: Jan 2, 2024
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.4 | 2.3.4 | 2.3.3 | 2.3.4 | 2.3.5 |
Milvus 2.3.4 brings significant enhancements, focusing on availability and usability. The update introduces access logs for better monitoring and integrates Parquet for efficient bulk imports. A key feature is the binlog index on growing segments for faster searches. Major improvements include support for up to 10,000 collections/partitions, reduced memory usage, clearer error messages, quicker loading, and better query shard balance. It addresses critical issues like resource leakage, load/release failures, and concurrency challenges. However, it discontinues regular expression searches in partitions to save resources, with an option to re-enable this feature in the configuration.
Features
Access Logs:
- Milvus now supports access logs for monitoring external interfaces. These logs record method names, user requests, response times, and error codes.
- Note: Currently, this feature supports only gRPC; RESTful requests are not included.
- Doc page: Configure Access Logs
Parquet File Import:
- This update introduces support for Parquet file imports, enhancing performance and memory efficiency. It also broadens data type support, including arrays and JSON.
- This feature supersedes the previous limitation of JSON and NumPy formats.
Binlog Index on Growing Segments:
- Milvus now employs a binlog index on growing segments to enhance search efficiency, allowing for advanced indices like IVF or Fast Scann.
- This improvement can increase search speeds in growing segments by up to tenfold.
Improvements
Expanded Collection/Partition Support:
- Milvus now supports up to 10,000 collections/partitions in a cluster, benefiting multi-tenant environments.
- The improvement comes from timetick mechanism refinement, goroutine management, and memory usage improvement.
- Note: Exceeding the recommended limit may affect failure recovery and resource usage. Recommended limit is 10,000 (Collection * Shard * Partition).
Reduced Memory Usage:
- Enhancements have been made to improve memory efficiency during various operations, including data retrieval and variable length data handling.
Refined Error Messaging:
- Error messages have been split into summaries and details for clearer understanding.
Accelerated Loading Speed:
- Various optimizations have been implemented to increase loading speeds, particularly in scenarios with frequent flushes and deletions.
Improved Query Shard Balance:
- Implemented balance channel in
querycoord
and other improvements for efficient shard management.
- Implemented balance channel in
Other Enhancements:
- Includes security improvements, MMap support for index loading, partition-level privileges, and more.
Critical Bug Fixes
Resource Leakage Fixes:
- Addressed critical memory leaks in Pulsar producer/consumer and improved garbage collection of meta snapshots.
Load/Release Failure Fixes:
- Resolved issues causing load/release operations to stall, especially in clusters with many segments.
Concurrency Issues:
- Fixed problems related to concurrent insertions, deletions, and queries.
Other Critical Fixes:
- Fixed an issue where upgrades from version 2.2 failed due to missing
CollectionLoadInfo
. - Fixed an issue where deletions might be lost because of errors in parsing compacted file logpaths (#29276).
- Fixed an issue where flush and compaction processes could become stuck under heavy insert/delete traffic.
- Fixed the inability to perform compact operations on the array type (#29505) (#29504).
- Fixed an issue where collections with more than 128 partitions failed to be released (#28567).
- Fixed an issue related to parsing expressions that include quotation marks (#28418).
- Addressed a failure in Azure Blob Storage’s
ListObjects
operation causing garbage collection failures (#27931) (#28894). - Fixed an issue with missing target database names in
RenameCollection
operations (#28911). - Fixed an issue where iterators lost data in cases of duplicated results (#29406) (#29446).
- Corrected the bulk insert binlog process to consider timestamp order when processing delta data (#29176).
- Fixed an issue to exclude insert data before a growing checkpoint (#29559).
- Addressed a problem where frequent flushing caused rate limits in Minio (#28625).
- Fixed an issue where creating growing segments could introduce an excessive number of threads (#29314).
- Fixed an issue in retrieving binary vectors from chunk cache (#28866) (#28884).
- Fixed an issue where checkpoints were incorrectly updated after dropping a collection (#29221).
- Fixed an issue where upgrades from version 2.2 failed due to missing
Breaking Change
- Discontinued Regular Expression Search in Partitions:
- To reduce resource consumption, regular expression searches in partitions have been discontinued. However, this feature can be re-enabled through configuration (see #29154 for details).
v2.3.3
Release date: Nov 10, 2023
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.3 | 2.3.3 | 2.3.3 | 2.3.3 | 2.3.3 |
Milvus 2.3.3 was a bugfix version that focused on minimizing business interruption during rolling upgrades to less than 30 seconds. Additionally, it aimed to enhance the performance of vector retrieval during searches. Critical bugs related to filtering on array types and the possible appearance of deleted data were addressed. Note that we also bumped Knowhere up to 2.2.2 in this release.
Features
Supported pure list JSON in bulk insert (#28126)
Improvements
- Constructed a plan directly when searching with vector output (#27963)
- Removed binlog/delta log from getRecoveryInfoV2 (#27895) (#28090)
- Refined code for fixed-length types array (#28109)
- Improved rolling upgrade unserviceable time
- Refined stop order (#28016) (#28089)
- Set qcv2 index task priority to Low (#28117) (#28134)
- Removed retry in getShards (#28011) (#28091)
- Fixed load index for stopping node (#28047) (#28137)
- Fixed retry on offline node (#28079) (#28139)
- Fixed QueryNode panic while upgrading (#28034) (#28114)
- Fixed coordinator fast restart by deleting old session (#28205)
- Fixed check grpc error logic (#28182) (#28218)
- Delayed the cancellation of ctx when stopping the node (#28249)
- Disabled auto balance when an old node exists (#28191) (#28224)
- Fixed auto balance block channel reassign after datanode restart (#28276)
- Fixed retry when proxy stopped (#28263)
- Reduced useless ObjectExists in AzureBlobManager (#28157)
- Got vector concurrently (#28119)
- Forced set Aliyun use_virtual_host to true for all (#28237)
- Fixed delete session key with prefix causing multiple QueryNode crashes (#28267)
Bug Fixes
- Fixed script stop unable to find Milvus process (#27958)
- Fixed timestamp reordering issue with delete records (#27941) (#28113)
- Fixed prefix query with longer subarray potentially causing a crash (#28112)
- Limited max thread num for pool (#28018) (#28115)
- Fixed sync distribution with the wrong version (#28130) (#28170)
- Added a custom HTTP header: Accept-Type-Allow-Int64 for JS client (#28125)
- Fixed bug for constructing ArrayView with fixed-length type (#28186)
- Fixed bug for setting index state when IndexNode connecting failed (#28221)
- Fixed bulk insert bug that segments are compacted after import (#28227)
- Fixed the target updated before version updated to cause data missing (#28257)
- Handled exceptions while loading (#28306)
v2.3.2
Release date: Oct 26, 2023
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.2 | 2.3.2 | 2.3.2 | 2.3.2 | 2.3.3 |
We’re thrilled to unveil Milvus 2.3.2, enriched with an array of novel features. Experience support for array data types, delve into intricate delete expressions, and celebrate the return of binary metric types such as SUBSTRUCTURE/SUPERSTRUCTURE.
This release promises enhanced performance through minimized data copying during loading and better bulk insertions. Coupled with heightened error messaging and handling, you’re in for a smoother experience. Notably, our commitment to rolling upgrade stability ensures minimized service disruptions during updates.
Breaking Changes
New Features
- Array datatype now supported (#26369)
- Doc page: Use Array Fields
- Introduced complex delete expressions (#25752)
- Doc page: Delete Entities
- Reintroduced binary metric types SUBSTRUCTURE/SUPERSTRUCTURE (#26766)
- Vector index mmap now available (#26750)
- CDC: Added capability to replicate mq messages (#27240)
- Facilitated renaming of database names within collections (#26543)
- Activated bulk insert of binlog data with partition keys (#27241)
- Enhanced support for multiple index engines (#27178)
- Introduced chunk cache to fetch raw vectors:
- Newly added ChunkCache facilitates vector retrieval from storage (#26142)
- Implemented Tikv as a distributed meta solution:
- Integrated Tikv (#26246)
- Rolled out float16 vector support (#25852)
- Note: Index for float16 vector coming in the next version
- Restful updates:
Performance Enhancements
- Optimized data loading by minimizing data copy operations (#26746)
- Streamlined bulk inserts with batched varchar reading (#26199)
- Improved handling of large structs using pointer receivers (#26668)
- Removed unnecessary offset checks during data fills (#26666)
- Addressed high CPU consumption linked to proto.size (#27054)
- Optimized scalar column data with MADV_WILLNEED (#27170)
Additional Enhancements
- Robust rolling upgrade capabilities:
- Significant improvement in system availability during rolling upgrades, ensuring minimal service interruptions.
- Upgraded error messaging and handling for a seamless experience.
- Optimized flushing processes:
- Addressed issues where delete commands weren’t being saved during flush operations.
- Resolved slow flush-related issues.
- Segregated task queues for Flush and DDL to prevent mutual blockages.
- Improved RocksMQ seek speeds (#27646) and standalone recovery times.
- Streamlined compact tasks (#27899)
- Added a channel manager in DataNode (#27308)
- Refined chunk management:
- Integrated grpc compression (#27894)
- Decoupled client-server API interfaces (#27186)
- Transitioned etcd watch-related code to event manager (#27192)
- Displayed index details during GetSegmentInfo (#26981)
Bug Fixes
- Resolved concurrent string parsing expression issues (#26721)
- Fixed connection issues with Kafka under SASL_SSL (#26617)
- Implemented error responses for yet-to-be-implemented APIs, replacing panic reactions (#26589)
- Addressed data race concerns:
- Mended partition garbage collection issues (#27816).
- Rectified SIGSEGV errors encountered when operating within gdb (#27736).
- Addressed thread safety issues in glog for standalone mode (#27703).
- Fixed instances where segments were inadvertently retained post-task cancellations (#26685).
- Resolved loading failures for collections exceeding 128 partitions (#26763).
- Ensured compatibility with scalar index types such as marisa-trie and Ascending (#27638).
- Corrected issues causing retrieval to sometimes exceed specified result limits (#26670).
- Solved startup failures in rootcoord due to role number limits (#27361).
- Patched Kafka consumer connection leaks (#27224).
- Disabled the enlarging of indices for flat structures (#27309).
- Updated garbage collector to fetch metadata post-storage listing (#27203).
- Fixed instances of datanode crashes stemming from simultaneous compaction and delete processes (#27167).
- Ironed out issues related to concurrent load logic in querynodev2 (#26959).
v2.3.1
Release date: Sep 22, 2023
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.1 | 2.3.1 | 2.3.1 | 2.3.1 | 2.3.2 |
We are excited to introduce Milvus 2.3.1, a patch release that includes several enhancements and bug fixes. These improvements are designed to enhance system stability and performance.
Features
- Restored support for SUBSTRUCTURE/SUPERSTRUCTURE binary metric types (#26766).
- Displayed index information during GetSegmentInfo (#26981).
Performance Improvement
- Improved loading mechanism (#26746): Unnecessary data copies have been reduced, resulting in enhanced overall load performance.
- Optimized MMap performance (#26750): The efficiency and capacity of MMap have been enhanced.
- Refactored storage merge insert data (#26839): The merging process has been optimized, leading to improved data node performance.
- Enhanced VARCHAR bulk insert speed (#26199): Batch processing reads have greatly improved the speed of VARCHAR bulk inserts.
- Utilized a pointer receiver for large structures (#26668): Memory copy has been enhanced by utilizing a pointer receiver.
Enhancements
- Enhanced error handling in QueryNode (#26910, #26940, #26918, #27013, #26904, #26521, #26773, #26676): Error messages have been made more descriptive and informative, improving the user experience.
- Enhanced Flush All API operations (#26802, #26769, #26859): The Flush, FlushAll, and GetFlushAllState API operations have undergone several improvements for better data syncing with object storage.
- Improved resilience of the RPC client with retry mechanism (#26795): The RPC client now has an enhanced retry mechanism, improving its resilience.
- Removed invalid offset check during data filling (#26666).
- Delayed connection reset for
Canceled
orDeadlineExceeded
gRPC code (#27014). - Achieved cleaner and more efficient error code management through miscellaneous code management and control enhancements (#26881, #26725, #26713, #26732).
Bug Fixes
- Fixed the index task retry issue (#26878): Canceled tasks are no longer marked as failed without retrying.
- Addressed load stability issues (#26763, #26959, #26931, #26813, #26685, #26630, #27027): Several stability issues related to load have been resolved.
- Resolved the segment retrieval issue (#26670): Retrieving now returns the correct number of results based on the specified limit.
- Fixed memory leak when putting duplicated segments (#26693).
- Fixed the bug for concurrent parsing expressions with strings (#26721).
- Fixed the panic caused by empty traceID (#26754) (#26808).
- Fixed the issue where timeout tasks never release the queue, leading to stuck compactions (#26593).
v2.3.0
Release date: Aug 23, 2023
Milvus version | Python SDK version | Java SDK version | Go SDK version | Node.js SDK version |
---|---|---|---|---|
2.3.0 | 2.3.0 | 2.3.0 | 2.3.0 | 2.3.0 |
After months of meticulous refinement, we are pleased to announce the official release of Milvus 2.3.0. This highly anticipated release contains a wealth of exciting new features and enhancements, including GPU support, improved query architecture, enhanced load balancing capabilities, integrated message queues, Arm-compatible images, improved observability, and improved O&M tools. This represents a major leap forward in the maturity, reliability and usability of the Milvus 2.x series. We cordially invite community users to be among the first to explore them and request that any feedback or issues be submitted on GitHub. Let’s work together to further refine and stabilize this exceptional 2.3.0 release.
Breaking changes
Deprecated Time Travel Feature
Due to its inactivity and the challenges it poses to the architecture design of Milvus, the time-travel feature has been deprecated in this release.
Discontinued CentOS Support
As CentOS 7 is about to reach its end of service (EOS) and official images based on CentOS 8 and CentOS 9 are not available, Milvus no longer supports CentOS. Instead, starting from this release, Milvus will provide images using the Amazonlinux distribution. It’s worth noting that Ubuntu-based images remain the well-tested and recommended option.
Removed Index and Metrics Algorithms
The following algorithms have been removed in this release:
- ANNOY and RHNSW for index-building of float vectors
- TANIMOTO for index-building of binary vectors
- Superstructure and Substructure metrics
These changes have been made to streamline and optimize the functionality of Milvus.
Upgraded Architecture
GPU Support
Milvus had GPU support in its earlier versions (v1.x), but it was temporarily unavailable when Milvus transitioned to a distributed architecture in v2.x. Thanks to the contributions of NVIDIA engineers and their implementation of the RAFT algorithm for Milvus Knowhere, GPU support is once again available in Milvus. This latest update not only brings back GPU capabilities but also incorporates cutting-edge industry algorithms. In benchmark tests, Milvus with GPU support has demonstrated impressive performance improvements, achieving a three-fold increase in query per second (QPS) and even up to a ten-fold increase for certain datasets.
Arm64 Support
With the growing popularity of Arm CPUs among cloud providers and developers, Milvus has recognized the importance of catering to the needs of both x86 and Arm architectures. To accommodate this demand, Milvus now offers images for both platforms. Additionally, the release of Arm images aims to provide MacOS users with a seamless experience when working with Milvus on their projects.
Refactored QueryNode
QueryNode plays a vital role in data retrieval within Milvus, making its availability, performance, and extensibility essential. However, the legacy QueryNode had several reported issues, including complex status management, duplicate message queues, unclear code structure, and unintuitive error messages. To address these concerns, we have undertaken a significant refactoring effort. This involved transforming QueryNode into a stateless component and eliminating data-deletion-related message queues. These changes aim to enhance the overall functionality and usability of QueryNode within the Milvus system.
Merged IndexCoord and DataCoord
We have merged IndexCoord and DataCoord into a single component, simplifying the deployment of Milvus. This consolidation reduces complexity and streamlines operations. Moving forward, subsequent releases will also witness the integration of certain functions of IndexNode and DataNode to align with this unified approach. These updates ensure a more efficient and seamless experience when utilizing Milvus.
NATS-based Message Queue (Experimental)
The stability, extensibility, and performance of the message queue are of utmost importance to Milvus, given its log-based architecture. To expedite the development of Milvus 2.x, we have introduced support for Pulsar and Kafka as the core log brokers. However, these external log brokers have their limitations. They can exhibit instability when handling multiple topics simultaneously, complexity in managing duplicate messages, and resource management challenges when there are no messages to process. Additionally, their GO SDKs may have inactive communities.
To address these issues, we have made the decision to develop our own log broker based on NATS and Bookeeper. This custom message queue is currently undergoing experimentation, and we welcome feedback and comments from the community. Our aim is to create a robust and efficient solution that addresses the unique requirements of Milvus.
New features
Upsert
Users now can use the upsert API in Milvus for updating or inserting data. It is important to note that the upsert API combines search, delete, and insert operations, which may result in a degradation of performance. Therefore, it is recommended to use the insert APIs for specific and definitive insertions, while reserving the upsert APIs for more ambiguous scenarios. Click here to read more.
Range Search
Users now have the option to set a distance range using arguments to retrieve specific results within that range in Milvus.
// add radius and range_filter to params in search_params
search_params = {"params": {"nprobe": 10, "radius": 20, "range_filter" : 10}, "metric_type": "L2"}
res = collection.search(
vectors, "float_vector", search_params, topK,
"int64 > 100", output_fields=["int64", "float"]
)
In the above example, the returned vectors will have distances ranging from 10 to 20 regarding the query vector. It is important to note that the method of distance measurement varies depending on the chosen metric type. Therefore, it is recommended to familiarize yourself with each metric type before applying a range search. Additionally, please be aware that the maximum number of vectors returned is limited to 16384. Click here to read more.
Count
In previous releases, users would often use the num_entities API to retrieve the total number of entities in a collection. However, it is important to note that the num_entities API only applies to entities within sealed segments. Making frequent calls to the flush API can result in the creation of numerous small segments, which can negatively impact the stability of the system and the performance of data retrieval in Milvus.
In this release, Milvus introduces the count statement as an alternative solution for users to obtain the number of entities in a collection without relying on the flush API.
Please be aware that the count statement consumes system resources, and it is advisable to avoid calling it frequently to prevent unnecessary resource consumption. Click here to read more.
Cosine Metrics
The Cosine Metrics is widely regarded as the standard method for measuring the distance between vectors, particularly in Large Language Models (LLMs). With the release of Milvus 2.3.0, cosine metrics are now natively supported. As a result, users no longer need to quantize vectors for IP (Inner Product) metrics. Click here to read more.
Raw Vectors in Search Returns
Starting from Milvus 2.3.0, the capability to include raw vectors in search results is introduced for certain metrics. However, please note that including raw vectors in search results necessitates secondary searches, which can potentially impact performance. In scenarios where performance is critical, it is recommended to use indexes such as HNSW and IVF_FLAT, which inherently support the inclusion of vectors in their search results. It’s important to mention that this feature currently does not apply to quantization-related indexes like IVF_PQ and IVF_SQ8. For more detailed information, please refer to https://github.com/zilliztech/knowhere/releases.
ScaNN
Milvus now includes support for FAISS’ FastScan, which has demonstrated a 20% performance improvement compared to HNSW and a 7-fold increase compared to IVF-FLAT in multiple benchmark tests. ScaNN, an index-building algorithm similar to IVF-PQ, offers a faster index-building process. However, it’s important to note that using ScaNN may result in a potential loss of precision and therefore requires refinement using the raw vectors.
The table below presents performance comparison results obtained using VectorDBBench. It evaluates the performance of ScaNN, HNSW, and IVF-FLAT in handling data retrieval from a 768-dimensional vector dataset sourced from Cohere. Click here to read more.
Index | Case | QPS | Latency (P99) s | Recall |
---|---|---|---|---|
ScaNN | 99% filtered | 626 | 6.9 | 0.9532 |
1% filtered | 750 | 6.3 | 0.9493 | |
0% filtered | 883 | 5.1 | 0.9491 | |
IVF-FLAT | 99% filtered | 722 | 6.1 | 0.9532 |
1% filtered | 122 | 16.1 | 0.9493 | |
0% filtered | 123 | 15.4 | 0.9494 | |
HNSW | 99% filtered | 773 | 6.6 | 1.0 |
1% filtered | 355 | 8.1 | 0.9839 | |
0% filtered | 696 | 5.4 | 0.9528 |
Iterator
PyMilvus now includes support for iterators, enabling users to retrieve more than 16,384 entities in a search or range search operation. The iterator functionality operates similarly to ElasticSearch’s scroll API and the cursor concept in relational databases. Click here to read more.
JSON_CONTAINS
Starting from Milvus 2.3.0, users can utilize the JSON_CONTAINS expressions to retrieve entities whose JSON field values contain one or a specified set of elements. This feature enhances the flexibility and capability of filtering and querying data within Milvus. Click here to read more.
CDC support
Change Data Capture (CDC) is a widely used functionality in databases, typically employed in scenarios such as active/standby data synchronization, incremental data backup, and data migration. For more detailed information on CDC, please refer to https://github.com/zilliztech/milvus-cdc.
Enhancements
MMap for capacity increase
MMap is a Linux kernel feature that allows the mapping of disk space to memory. This feature enables loading data into the hard drive and mmap-ing it to memory, thereby increasing the single-machine data capacity with a marginal 20% performance degradation. Users who prioritize cost-efficiency are encouraged to test this enhancement and evaluate its benefits. Click here to read more.
Performance improvement for data filtering
In hybrid searches, Milvus performs scalar queries and vector searches sequentially. This approach often leads to a high number of entities being filtered out after the scalar queries, which can significantly degrade the performance of the built vector index. In Milvus 2.3.0, by optimizing the data filtering policies in HNSW, we have improved the scalar query performance and resolved this issue, ensuring better overall performance during hybrid searches.
Growing index
Milvus distinguishes between indexed data and stream data, each treated differently. Indexed data benefits from the built index, which accelerates the search process. On the other hand, stream data relies on brute-force search, which can potentially lead to performance degradation. To address this issue, Milvus introduces the concept of a growing index. This growing index ensures consistent search performance for both indexed and stream data, mitigating any potential performance degradation caused by brute-force search.
Increased resource usage in the multi-core environment
Approximate nearest search (ANN) is a computationally intensive task that heavily relies on CPU usage. In previous releases, the CPU usage in Milvus could remain below 70% even when multiple cores were available. However, in Milvus 2.3.0, this limitation has been overcome, and the system can now fully utilize all available CPU resources for improved performance during ANN computations.
Stability
New load balancer
Milvus 2.1.0 introduced support for multiple replicas in memory to improve QPS. However, feedback from the community highlighted issues with immediate QPS improvement, longer system recovery time after node shutdown, workload imbalance among nodes, and low CPU utilization. To tackle these issues, we redesigned the load balancing algorithm with dynamic load adjustment based on real-time load information. This enhancement enables timely detection of node status changes and workload imbalances, resulting in efficient load management.
Operability
Dynamic Configuration
Modifying configurations is a common operation for database maintenance and optimization. Starting from this version, Milvus supports dynamically modifying configuration parameters without the need to restart the cluster. There are two supported methods: modifying key-value pairs in etcd or directly modifying the milvus.yaml
configuration file. It is important to note that not all configuration parameters can be modified dynamically. Click here to read more.
Tracing support
Tracing is an important means of identifying bottleneck points in a system and is crucial for optimization. Starting from version 2.3.0, Milvus supports the Opentelemetry tracing protocol. Tracing collectors that support this protocol, such as Jaeger, can be used to observe Milvus’s invocation path and analyze system performance.
Error codes
Milvus has undergone a reorganization of its error codes according to the new design of the error code reporting system. As a result of this upgrade, error messages in Milvus will be more clear and concise, providing improved clarity in error reporting.
Tools
Birdwatcher upgrade
Through months of development, Birdwatcher incorporates the following features:
- RESTful API for seamless integration with other diagnostic systems.
- PProf command support to facilitate integration with the go pprof tool.
- Storage analysis capabilities.
- Efficient log analysis functionality.
- Ability to view and edit Milvus configuration in etcd.
Attu upgrade
A newly designed GUI makes Attu more user-friendly.
Attu