Skip to content
Merged
Show file tree
Hide file tree
Changes from 18 commits
Commits
Show all changes
28 commits
Select commit Hold shift + click to select a range
fe7a085
Update version
g-despot Oct 16, 2025
9dbe3be
Merge branch 'main' into v1-34/main
g-despot Oct 23, 2025
07d0089
Update docs
g-despot Oct 23, 2025
9637bc2
Update docs
g-despot Oct 23, 2025
73b7b34
Update docs
g-despot Oct 23, 2025
c7bef2b
Merge pull request #215 from weaviate/v1-34/acorn
g-despot Oct 24, 2025
9a0e8fa
Update docs
g-despot Oct 27, 2025
d20711e
Merge pull request #219 from weaviate/v1-34/server-side-batching
g-despot Oct 27, 2025
e75284b
Update metrics
g-despot Oct 29, 2025
cd44c95
Update docs
g-despot Oct 30, 2025
db43f58
Merge pull request #223 from weaviate/v1-34/metrics
g-despot Oct 30, 2025
bd0c244
Merge branch 'main' into v1-34/main
g-despot Oct 30, 2025
a2bdb96
Merge branch 'v1-34/main' into v1-34/flat-index-rq
g-despot Oct 30, 2025
a80e713
Merge pull request #224 from weaviate/v1-34/flat-index-rq
g-despot Oct 30, 2025
a909610
Update ssb docs
g-despot Oct 30, 2025
fdf37ba
Update code
g-despot Oct 30, 2025
b41efed
Update docs
g-despot Oct 30, 2025
8f536ee
Update docs
g-despot Oct 30, 2025
f0b552b
comments; wip
databyjp Oct 30, 2025
e32ec25
Add Contextual AI doc draft
databyjp Oct 31, 2025
7661d6c
Merge pull request #226 from weaviate/v1-34/jph-comments
g-despot Nov 4, 2025
4e1d905
Merge branch 'main' into v1-34/main
g-despot Nov 4, 2025
dd535d0
Implement feedback
g-despot Nov 4, 2025
cd6ac2a
Revision
databyjp Nov 4, 2025
16f0d6a
Merge pull request #227 from weaviate/v1-34/contextual-ai-models
g-despot Nov 4, 2025
62f9fb0
Add CBD note
g-despot Nov 4, 2025
91f095f
Update generative model parameters and documentation for clarity
databyjp Nov 5, 2025
7cf4da9
Merge pull request #232 from weaviate/v1-34/contextualai-knowledge
databyjp Nov 5, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions _includes/code/howto/configure-rq/rq-compression-v4.py
Original file line number Diff line number Diff line change
Expand Up @@ -101,6 +101,10 @@
quantizer=Configure.VectorIndex.Quantizer.rq(
bits=8, # Optional: Number of bits
rescore_limit=20, # Optional: Number of candidates to fetch before rescoring
cache=True, # Optional: Enable caching for flat index (enabled by default for for HNSW)
),
vector_index_config=Configure.VectorIndex.flat(
vector_cache_max_objects=100000, # Optional: Maximum number of objects in the memory cache
),
# highlight-end
),
Expand Down
2 changes: 1 addition & 1 deletion _includes/code/howto/manage-data.collections.py
Original file line number Diff line number Diff line change
Expand Up @@ -202,7 +202,7 @@
vector_index_config=Configure.VectorIndex.hnsw(
ef_construction=300,
distance_metric=VectorDistances.COSINE,
filter_strategy=VectorFilterStrategy.SWEEPING, # or ACORN (Available from Weaviate v1.27.0)
filter_strategy=VectorFilterStrategy.ACORN,
),
# highlight-end
),
Expand Down
84 changes: 42 additions & 42 deletions _includes/code/howto/manage-data.import.py
Original file line number Diff line number Diff line change
Expand Up @@ -558,48 +558,48 @@ def add_object(obj) -> None:
client.collections.delete("NewCollection")


# # ==================================================
# # ===== Server-side (automatic) batch import =====
# # ==================================================

# # Re-create the collection
# client.collections.delete("MyCollection")
# client.collections.create(
# "MyCollection",
# vector_config=Configure.Vectors.self_provided()
# )

# # START ServerSideBatchImportExample
# data_rows = [
# {"title": f"Object {i+1}"} for i in range(5)
# ]

# collection = client.collections.get("MyCollection")

# # highlight-start
# # Use `automatic` for server-side batching. The client will send data
# # in chunks and the server will dynamically manage the import process.
# with collection.batch.automatic() as batch:
# for data_row in data_rows:
# batch.add_object(
# properties=data_row,
# )
# # highlight-end
# if batch.number_errors > 10:
# print("Batch import stopped due to excessive errors.")
# break

# failed_objects = collection.batch.failed_objects
# if failed_objects:
# print(f"Number of failed imports: {len(failed_objects)}")
# print(f"First failed object: {failed_objects[0]}")
# # END ServerSideBatchImportExample

# result = collection.aggregate.over_all(total_count=True)
# assert result.total_count == 5

# # Clean up
# client.collections.delete(collection.name)
# ==================================================
# ===== Server-side (automatic) batch import =====
# ==================================================

# Re-create the collection
client.collections.delete("MyCollection")
client.collections.create(
"MyCollection",
vector_config=Configure.Vectors.self_provided()
)

# START ServerSideBatchImportExample
data_rows = [
{"title": f"Object {i+1}"} for i in range(5)
]

collection = client.collections.get("MyCollection")

# highlight-start
# Use `experimental` for server-side batching. The client will send data
# in chunks and the server will dynamically manage the import process.
with collection.batch.experimental() as batch:
for data_row in data_rows:
batch.add_object(
properties=data_row,
)
# highlight-end
if batch.number_errors > 10:
print("Batch import stopped due to excessive errors.")
break

failed_objects = collection.batch.failed_objects
if failed_objects:
print(f"Number of failed imports: {len(failed_objects)}")
print(f"First failed object: {failed_objects[0]}")
# END ServerSideBatchImportExample

result = collection.aggregate.over_all(total_count=True)
assert result.total_count == 5

# Clean up
client.collections.delete(collection.name)


client.close()
10 changes: 6 additions & 4 deletions _includes/configuration/rq-compression-parameters.mdx
Original file line number Diff line number Diff line change
@@ -1,4 +1,6 @@
| Parameter | Type | Default | Details |
| :------------------- | :------ | :------ | :--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
| `rq`: `bits` | integer | `8` | The number of bits used to quantize each data point. Value can be `8` or `1`. <br/> <br/>Learn more about [8-bit](/weaviate/concepts/vector-quantization#8-bit-rq) and [1-bit](/weaviate/concepts/vector-quantization#1-bit-rq) RQ. |
| `rq`: `rescoreLimit` | integer | `-1` | The minimum number of candidates to fetch before rescoring. |
| Parameter | Type | Default | Details |
| :---------------------- | :------ | :------ | :-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
| `rq`: `bits` | integer | `8` | The number of bits used to quantize each data point. Value can be `8` or `1`. <br/> <br/>Learn more about [8-bit](/weaviate/concepts/vector-quantization#8-bit-rq) and [1-bit](/weaviate/concepts/vector-quantization#1-bit-rq) RQ. |
| `rq`: `rescoreLimit` | integer | `-1` | The minimum number of candidates to fetch before rescoring. |
| `rq` : `cache` | boolean | `false` | Whether to use the vector cache.<br/> (only when using the `flat` vector index type) |
| `vectorCacheMaxObjects` | integer | `1e12` | Maximum number of objects in the memory cache. By default, this limit is set to one trillion (`1e12`) objects when a new collection is created. For sizing recommendations, see [Vector cache considerations](/weaviate/concepts/vector-index#vector-cache-considerations). |
3 changes: 0 additions & 3 deletions _includes/named-vector-compress.mdx
Original file line number Diff line number Diff line change
@@ -1,4 +1 @@
:::info Added in `v1.24`
:::

Collections can have multiple [named vectors](/weaviate/config-refs/collections#named-vectors). The vectors in a collection can have their own configurations, and compression must be enabled independently for each vector. Every vector is independent and can use [PQ](/weaviate/configuration/compression/pq-compression), [BQ](/weaviate/configuration/compression/bq-compression), [RQ](/weaviate/configuration/compression/rq-compression), [SQ](/weaviate/configuration/compression/sq-compression), or no compression.
2 changes: 1 addition & 1 deletion docs/deploy/configuration/env-vars/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ import APITable from '@site/src/components/APITable';
| `QUERY_MAXIMUM_RESULTS` | Sets the maximum total number of objects that can be retrieved. | `string - number` | `10000` |
| `QUERY_SLOW_LOG_ENABLED` | Log slow queries for debugging. Requires a restart to update. <br/> (New in 1.24.16, 1.25.3) | `boolean` | `False` |
| `QUERY_SLOW_LOG_THRESHOLD` | Set a threshold time for slow query logging. Requires a restart to update. <br/> (New in 1.24.16, 1.25.3) | `string` | `2s` <br/> Values are times: `3h`, `2s`, `100ms` |
| `REINDEX_SET_TO_ROARINGSET_AT_STARTUP` | Allow Weaviate to perform a one-off re-indexing to [use Roaring Bitmaps](/weaviate/concepts/filtering.md#migration-to-indexFilterable). <br/><br/>Available in versions `1.18` and higher. | `boolean` | `true` |
| `REINDEX_SET_TO_ROARINGSET_AT_STARTUP` | Allow Weaviate to perform a one-off re-indexing to use Roaring Bitmaps. <br/><br/>Available in versions `1.18` and higher. | `boolean` | `true` |
| `TOKENIZER_CONCURRENCY_COUNT` | Limit the combined number of GSE and Kagome tokenizers running at the same time. Default: `GOMAXPROCS` | `string - number` | `NUMBER_OF_CPU_CORES` |
| `TOMBSTONE_DELETION_CONCURRENCY` | The maximum number of cores to use for tombstone deletion. Set this to limit the number of cores used for cleanup. Default: Half of the available cores. (New in `v1.24.0`) | `string - int` | `4` |
| `TOMBSTONE_DELETION_MAX_PER_CYCLE` | Maximum number of tombstones to delete per cleanup cycle. Set this to limit cleanup cycles, as they are resource-intensive. As an example, set a maximum of 10000000 (10M) for a cluster with 300 million-object shards. Default: none | `string - int` (New in `v1.24.15` / `v1.25.2`) | `10000000` |
Expand Down
Loading
Loading