Hi Team , I have a schema with the configuration
class_schema = {
“class”:“class_”, # class name
“vectorizer”: “none”,
“properties”: [{
“name”: “record_id”,
“dataType”: [“int”]
}],
“vectorIndexType”: “hnsw”,
“vectorIndexConfig”: {
“skip”: False,
“ef”: 256,
“efConstruction”: 256,
“maxConnections”: 64,
“vectorCacheMaxObjects”:300000,
“distance”: “l2-squared”
},
“shardingConfig”: {
“virtualPerPhysical”: 128,
“desiredCount”: 2,
“actualCount”: 2,
“key”: “_id”,
“strategy”: “hash”,
“function”: “murmur3”
},
“replicationConfig”:{
“factor”:1
}
}
I currently have 50K objects with 128 dimension vector. I recognize that memory consumption is more than double the disk space occupied.
disk = 53.2M,memory =110M.
queries:
- is that memory consumption always greater than disk occupied for the above vectorIndexConfig?
- does any thing have to done to reduce memory usage without comprising search accuracy?
- What if replica to 2, does memory consumption also doubled ?