# Collections
In Typesense, every record you index is called a Document
and a group of documents with similar fields is called a Collection
.
A Collection is roughly equivalent to a table in a relational database.
# Create a collection
Before we can add documents to Typesense, we need to first create a Collection
- we give it a name and describe the fields that will be indexed from our Documents
.
We call this definition the collection's schema
, which is just a fancy term to describe the fields (and their data types) in your documents.
TIP
It might help to think of defining a collection "schema" as being similar to defining "types" in a strongly-typed programming language like Typescript, C, Java, Dart, Rust, etc. This ensures that the documents you add to your collection have consistent data types and are validated, and this helps prevent a whole class of errors you might typically see with mis-matched or inconsistent data types across documents.
Organizing Collections
Read more on how to organize data into collections in this dedicated guide article: Organizing Collections.
There are two ways to specify a schema:
- Pre-define all the fields to be indexed from your documents OR
- Have Typesense automatically detect your fields and data types based on the documents you index.
The simplest option is #2 where you don't have to worry about defining an explicit schema. But if you need more fine-grained control and/or validation, you want to use #1 or even mix both together.
# With pre-defined schema
Let's first create a collection with an explicit, pre-defined schema.
This option gives you fine-grained control over your document fields' data types and configures your collection to reject documents that don't match the data types defined in your schema (by default).
If you want Typesense to automatically detect your schema for you, skip over to auto-schema detection.
See Schema Parameters for all available options, and Field Types for all available data types.
Sample Response
Definition
POST ${TYPESENSE_HOST}/collections
IMPORTANT NOTE & TIP
All fields you mention in a collection's schema will be indexed in memory.
There might be cases where you don't intend to search / filter / facet / group by a particular field and just want it to be stored (on disk) and returned as is when a document is a search hit. For eg: you can store image URLs in every document that you might use when displaying search results, but you might not want to text-search the actual URLs.
You want to NOT mention these fields in the collection's schema or mark these fields as index: false
(see fields
schema parameter below) to mark it as an unindexed field.
You can have any number of these additional unindexed fields in the documents when adding them to a collection - they will just be stored on disk, and will not take up any memory.
# With auto schema detection
If your field names are dynamic and not known upfront, or if you just want to keep things simple and index all fields you send in your documents by default, auto-schema detection should help you.
You can define a wildcard field with the name .*
and type auto
to let Typesense automatically
detect the type of the fields when you add documents to the collection.
In fact, you can use any RegEx expression to define a field name.
When a .*
field is defined this way, all the fields in a document are automatically indexed for searching and filtering.
# Data Coercion
Say you've set type: auto
for a particular field (or fields) (eg: popularity_score
) in a collection and send the first document as:
Since popularity_score
has type: auto
, the data-type will automatically be set to int64
internally.
What happens when the next document's popularity_score
field is not an integer field, but a string? For eg:
By default, Typesense will try to coerce (convert) the value to the previously inferred type.
So in this example, since the first document had a numeric data-type for popularity_score
, the second document's popularity_score
field will be coerced to an integer from string.
However, this may not always work - (for eg: say the value has alphabets, it can't be coerced to an integer). In such cases, when Typesense is unable to coerce the field value to the previously inferred type, the indexing will fail with the appropriate error.
TIP
You can control this default coercion behavior at write-time with the dirty_values
parameter.
# Faceting fields with auto-schema detection
Faceting is not enabled for a wildcard field {"name": ".*" , ...}
, since that can consume a lot of memory,
especially for large text fields. However, you can still explicitly define specific fields (with or without RegEx names) to facet by
setting facet: true
for them.
For e.g, when you define a schema like this:
{
"name": "companies",
"fields": [
{
"name": ".*_facet",
"type": "auto",
"facet": true
}
]
}
This will only set field names that end with _facet
in the document, as a facet.
# Geopoint
and auto-schema detection
A geopoint
field requires an explicit type definition, as the geo field value is represented as a 2-element
float field and we cannot differentiate between a lat/long definition and an actual float array.
# Indexing all but some fields
If you have a case where you do want to index all fields in the document, except for a few fields, you can use the {"index": false, "optional": true}
settings to exclude fields.
Note: it is not currently possible to have a mandatory field excluded from the indexing, hence the setting to optional.
For eg, if you want to index all fields, except for fields that start with description_
, you can use a schema like this:
TIP
You can mix auto-schema detection with explicit field definitions.
If an explicit definition is available for a field (country
in the example above), Typesense will give
preference to that before falling back to the wildcard definition.
When such an explicit field definition is not available, the first document that contains a field with a given name determines the type of that field.
For example, if you index a document with a field named title
and it is a
string, then the next document that contains the field named title
will be expected to have a string too.
# Schema parameters
Parameter | Required | Description |
---|---|---|
name | yes | Name of the collection you wish to create. |
fields | yes | A list of fields that you wish to index for querying, filtering, faceting, grouping and sorting. For each field, you have to specify at least it's name and type .Eg: {"name": "title", "type": "string", "facet": false, "index": true} name can be a simple string like "name": "score" . Or you can also use a RegEx to specify field names matching a pattern. For eg: if you want to specify that all fields starting with score_ should be an integer, you can set name as "name": "score_.*" .Declaring a field as optional A field can be declared as optional by setting "optional": true .Declaring a field as a facet A field can be declared as a facetable field by setting "facet": true . Faceted fields are indexed verbatim without any tokenization or preprocessing. For example, if you are building a product search, color and brand could be defined as facet fields. Once a field is enabled for faceting in the schema, it can be used in the facet_by search parameter..Enabling stemming Stemming allows you to handle common word variations (singular / plurals, tense changes) of the same root word. For eg: searching for walking , will also return results with walk , walked , walks , etc when stemming is enabled. Enable stemming on the contents of the field during indexing and querying by setting "stem": true . The actual value stored on disk is not affected. We use the Snowball stemmer (opens new window). Language selection for stemmer is automatically made from the value of the locale property associated with the field.Declaring a field as un-indexed You can set a field as un-indexed (you can't search/sort/filter/facet on it) by setting "index": false . This is useful when used along with auto schema detection and you need to exclude certain fields from indexing.Prevent field from being stored on disk: Set "store": false to ensure that a field value is removed from the document before the document is saved to disk. Configuring language-specific tokenization: The default tokenizer that Typesense uses works for most languages, especially ones that separate words by spaces. However, based on feedback from users, we've added locale specific customizations for the following languages. You can enable these customizations for a field, by setting a field called locale inside the field definition. Eg: {name: 'title', type: 'string', locale: 'ja'} will enable the Japanese locale customizations for the field named title . Here's the list of all language-specific customizations:
|
token_separators | no | List of symbols or special characters to be used for splitting the text into individual words in addition to space and new-line characters. For e.g. you can add - (hyphen) to this list to make a word like non-stick to be split on hyphen and indexed as two separate words. Read this guide article for more examples on how to use this setting. |
symbols_to_index | no | List of symbols or special characters to be indexed. For e.g. you can add + to this list to make the word c++ indexable verbatim. Read this guide article for more examples on how to use this setting. |
default_sorting_field | no | The name of an int32 / float field that determines the order in which the search results are ranked when a sort_by clause is not provided during searching. This field must indicate some kind of popularity. For example, in a product search application, you could define num_reviews field as the default_sorting_field to rank products that have the most reviews higher by default.Additionally, when a word in a search query matches multiple possible words (either during a prefix (partial word) search or because of a typo), this parameter is used to rank such equally matching records. For e.g. Searching for "ap", will match records with "apple", "apply", "apart", "apron", or any of hundreds of similar words that start with "ap" in your dataset. Also, searching for "jofn", will match records with "john", "joan" and all similar variations that are 1-typo away in your dataset. For performance reasons though, Typesense will only consider the top 4 prefixes or typo variations by default (the 4 is configurable using the max_candidates search parameter, which defaults to 4 ). If default_sorting_field is NOT specified in the collection schema, then "top" is defined as the prefixes or typo variations with the most number of matching records. But let's say you have a field called popularity in each record, and you want Typesense to use the value in that field to define the "top" records, you'd set that field as default_sorting_field: popularity . Typesense will then use the value of that field to fetch the top max_candidates number of terms that are most popular, and as users type in more characters, it will refine the search further to always rank the most popular prefixes highest. |
# Field parameters
Parameter | Required | Description |
---|---|---|
name | yes | Name of the field. |
type | yes | The data type of the field (see the section below for a list of types). |
facet | no | Enables faceting on the field. Default: false . |
optional | no | When set to true , the field can have empty, null or missing values. Default: false . |
index | no | When set to false , the field will not be indexed in any in-memory index (e.g. search/sort/filter/facet). Default: true . |
store | no | When set to false , the field value will not be stored on disk. Default: true . |
sort | no | When set to true , the field will be sortable. Default: true for numbers, false otherwise. |
infix | no | When set to true , the field value can be infix-searched. Incurs significant memory overhead. Default: false . |
locale | no | For configuring language specific tokenization, e.g. jp for Japanese. Default: en which also broadly supports most European languages. |
num_dim | no | Set this to a non-zero value to treat a field of type float[] as a vector field. |
vec_dist | no | The distance metric to be used for vector search. Default: cosine . You can also use ip for inner product. |
reference | no | Name of a field in another collection that should be linked to this collection so that it can be joined during query. |
range_index | no | Enables an index optimized for range filtering on numerical fields (e.g. rating:>3.5 ). Default: false . |
stem | no | Values are stemmed before indexing in-memory. Default: false . |
# Field types
Typesense allows you to index the following types of fields:
type | Description |
---|---|
string | String values |
string[] | Array of strings |
int32 | Integer values up to 2,147,483,647 |
int32[] | Array of int32 |
int64 | Integer values larger than 2,147,483,647 |
int64[] | Array of int64 |
float | Floating point / decimal numbers |
float[] | Array of floating point / decimal numbers |
bool | true or false |
bool[] | Array of booleans |
geopoint | Latitude and longitude specified as [lat, lng] . Read more here. |
geopoint[] | Arrays of Latitude and longitude specified as [[lat1, lng1], [lat2, lng2]] . Read more here. |
object | Nested objects. Read more here. |
object[] | Arrays of nested objects. Read more here. |
string* | Special type that automatically converts values to a string or string[] . |
image | Special type that is used to indicate a base64 encoded string of an image used for Image search. |
auto | Special type that automatically attempts to infer the data type based on the documents added to the collection. See automatic schema detection. |
# Cloning a collection schema
Here's how you can clone an existing collection's schema (documents are not copied), overrides and synonyms.
curl -k "http://localhost:8108/collections?src_name=existing_coll" -X POST -H "Content-Type: application/json" \
-H "X-TYPESENSE-API-KEY: ${TYPESENSE_API_KEY}" -d '{
"name": "new_coll"
}'
The above API call will create a new collection called new_coll
that contains the schema, overrides and synonyms of
the collection existing_coll
. The actual documents in the existing_coll
collection are not copied,
so this is primarily useful for creating new collections from an existing reference template.
TIP
Cloning a collection this way, does not copy the data.
# Adding metadata to schema
If you wish, you could populate a metadata
object to the schema while creating a collection to add
details about the collection, for e.g. when it was created, who created it etc.
{
"name": "docs",
"enable_nested_fields": true,
"fields": [
{
"name": "title",
"type": "string",
"facet": true
}
],
"metadata": {
"batch_job": 325,
"indexed_from": "2023-04-20T00:00:00.000Z"
}
}
The fields within the metadata
object are persisted and returned in the GET /collections
end-point.
# Notes on indexing common types of data
Here's how to index other common types of data, using the basic primitives in the table above:
# Indexing nested fields
Typesense supports indexing nested objects (and array of objects) from v0.24
.
You must first enable nested fields at a collection level via the enable_nested_fields
schema property, and the object
or object[]
data type:
curl -k "http://localhost:8108/collections" -X POST -H "Content-Type: application/json" \
-H "X-TYPESENSE-API-KEY: ${TYPESENSE_API_KEY}" -d '{
"name": "docs",
"enable_nested_fields": true,
"fields": [
{"name": "person", "type": "object"},
{"name": "details", "type": "object[]"}
]
}'
When you now search on an object field name, all sub-fields will be automatically searched. Use a dot notation to refer to specific sub-fields, e.g. person.last_name
or person.school.name
.
You can also index specific sub-fields within a nested object. For eg, if your document looks like this:
{
"id": 1,
"name": "Jack",
"address": {
"line_1": "111 1st Street",
"city": "Alpha",
"zip": "98765"
}
}
And say you only want to index the zip
field inside the address
nested object, without indexing other fields like line_1
, you can specify this in the schema like this:
curl -k "http://localhost:8108/collections" -X POST -H "Content-Type: application/json" \
-H "X-TYPESENSE-API-KEY: ${TYPESENSE_API_KEY}" -d '{
"name": "docs",
"enable_nested_fields": true,
"fields": [
{"name": "name", "type": "string"},
{"name": "address.zip", "type": "string"}
]
}'
To index specific fields inside an array of objects, you want to specify an array data type. For eg, if your document looks like this:
{
"id": 1,
"name": "Jack",
"addresses": [
{
"line_1": "111 1st Street",
"city": "Alpha",
"zip": "98765"
},
{
"line_1": "222 2nd Street",
"city": "Zeta",
"zip": "45678"
}
]
}
And say you only want to index the zip
field inside each address object in the addresses
array of objects, without indexing other fields, you can specify this in the schema like this:
curl -k "http://localhost:8108/collections" -X POST -H "Content-Type: application/json" \
-H "X-TYPESENSE-API-KEY: ${TYPESENSE_API_KEY}" -d '{
"name": "docs",
"enable_nested_fields": true,
"fields": [
{"name": "name", "type": "string"},
{"name": "addresses.zip", "type": "string[]"}
]
}'
Note
If there are overlapping definitions for nested fields at different levels of the nested hierarchy, the broader definition will take precedence over the definition for a sub-field.
Indexing nested objects via flattening
You can also flatten objects and arrays of objects into top-level keys before sending the data into Typesense.
For example, a document like this containing nested objects:
{
"nested_field": {
"field1": "value1",
"field2": ["value2", "value3", "value4"],
"field3": {
"fieldA": "valueA",
"fieldB": ["valueB", "valueC", "valueD"]
}
}
}
would need to be flattened as:
{
"nested_field.field1": "value1",
"nested_field.field2": ["value2", "value3", "value4"],
"nested_field.field3.fieldA": "valueA",
"nested_field.field3.fieldB": ["valueB", "valueC", "valueD"]
}
before indexing it into Typesense.
To simplify traversing the data in the results, you might want to send both the flattened and unflattened version of the nested fields into Typesense, and only set the flattened keys as indexed in the collection's schema and use them for search/filtering/faceting. At display time when parsing the results, you can then use the nested version.
# Indexing Dates
Dates need to be converted into Unix timestamps (opens new window) and stored as int64
fields in Typesense.
Most languages have libraries that help do this conversion for you.
You'll then be able to use numerical operators like <
, >
, etc to filter records that are before or after or between dates.
# Indexing other types of data
Read our dedicated guide article on how to index other common types of data like emails, phone numbers, SKUs, model numbers, etc here.
# Retrieve a collection
Retrieve the details of a collection, given its name.
Sample Response
Definition GET ${TYPESENSE_HOST}/collections/:collection
# List all collections
Returns a summary of all your collections. The collections are returned sorted by creation date, with the most recent collections appearing first.
NOTE: By default, ALL collections are returned, but you can use the offset
and limit
parameters to paginate on the collection listing.
Sample Response
Definition GET ${TYPESENSE_HOST}/collections
# Drop a collection
Permanently drops a collection. This action cannot be undone. For large collections, this might have an impact on read latencies.
Sample Response
When a collection is dropped, we perform a disk compaction operation which reclaims the disk space used by the
deleted collection. However, if you are managing a large number of collections and need to drop collections
frequently, you can disable this compaction to improve the performance of the drop operation
via the compact_store
GET parameter. You can still do a full database compaction
via the API.
Definition DELETE ${TYPESENSE_HOST}/collections/:collection
# Update or alter a collection
Typesense supports adding or removing fields to a collection's schema in-place.
TIP
Typesense supports updating all fields except the id
field (since it's a special field within Typesense).
Let's see how we can add a new company_category
field to the companies
collection and also drop the existing
num_employees
field.
Sample Response
TIP
Dropping a field affects only the schema and the in-memory index; it does not remove the data associated with that field from the existing documents stored on disk. In Typesense, documents can contain additional fields not explicitly defined in the schema.
To permanently remove a field and its data from a document, you must update the document directly, by passing null
to the field you want to remove.
Definition PATCH ${TYPESENSE_HOST}/collections/:collection
WARNING
The schema update operation is a synchronous blocking operation.
When the update is in progress, all incoming writes to that particular collection will wait for the schema update to finish. So, we recommend updating fields one at a time, especially for large collections and during off-peak hours.
Reads will be serviced as usual, without blocking.
Alternatively, you can also use the alias feature to do zero downtime schema changes.
WARNING
Only one alter operation can be in progress at a time for a given cluster.
Since the alter operation can take a long time, this ensures that a client with a short default time-out does not end up retrying the same alter request.
The update operation consists of an initial validation step where the records on-disk are assessed to ensure
that they are compatible with the proposed schema change. For example, let's say there is a string
field A
which
is already present in the documents on-disk but is not part of the schema. If you try to update the collection
schema by adding a field A
with type integer
, the validation step will reject this change as it's incompatible
with the type of data already present.
If the validation is successful, the actual schema change is done and the records are indexed / re-indexed / dropped as per the requested change. The process is complete as soon as the API call returns (make sure you use a large client timeout value). Because of the blocking nature of the update, we recommend doing the change during off-peak hours. Alternatively, you can also use the alias feature to do zero downtime schema changes.
# Modifying an existing field
Since Typesense currently only supports adding/deleting a field, any modifications to an existing field should be
expressed as a drop + add operation. All fields except the id
field can be modified.
For example, to add a facet
property to the company_category
field, we will drop + add it in the same change set:
curl "http://localhost:8108/collections/companies" \
-X PATCH \
-H "Content-Type: application/json" \
-H "X-TYPESENSE-API-KEY: ${TYPESENSE_API_KEY}" \
-d '{
"fields": [
{"name": "company_category", "drop": true },
{"name": "company_category", "type": "string", "facet": true }
]
}'
# Using an alias
If you need to do zero-downtime schema changes, you could also re-create the collection fully with the updated schema and use the Collection Alias feature to do a zero-downtime switch over to the new collection:
Let's say you have a collection called movies_jan_1
that you want to change the schema for.
- First create an alias pointing to your collection. For eg: create an alias called
movies
pointing tomovies_jan_1
. Use this collection alias in your application to search / index documents in your collection. - Create a new timestamped collection with the updated collection schema. For eg:
movies_feb_1
. - In your application, temporarily write to both the old collection and also to the new collection in parallel - essentially dual writes. Eg: Send writes both to the
movies_jan_1
collection and also to the newmovies_feb_1
collection. - Now, upsert the data from your primary database into the new collection. Eg
movies_feb_1
. - Update the collection alias to now point to the new collection. Eg: Update
movies
to now point tomovies_feb_1
. - Stop your application from sending writes to the old collection and drop the old collection,
movies_jan_1
in our example.
Once you update the alias, any search / indexing operations will go to the new collection (eg: movies_feb_1
)
without you having to do any additional application-side changes.
# Dynamic field additions
If you only need to add new fields to the schema on the fly, we recommend using auto-schema detection when creating the collection. You can essentially define RegEx field names and when documents containing new field names that match the RegEx come in, the new fields will automatically be added to the schema.