Search Index JSON Properties
- reference
Use a JSON payload to control the settings for a Search index.
When you create a Search index with the REST API, you must give a JSON payload with the settings for the new Search index.
Your JSON payload must contain the properties described in Initial Settings, including the Params Object.
Initial Settings
The start of the JSON payload for a Search index contains important settings for your index:
{
"name": "gfx",
"type": "fulltext-index",
"uuid": "28b999e9e17dd4a7",
"sourceType": "gocbcore",
"sourceName": "travel-sample",
"sourceUUID": "d3604c0ec4792b4c6c5f7f2cc8207b80",
"sourceParams": {},
"planParams": {
"maxPartitionsPerPIndex": 1024,
"indexPartitions": 1,
"numReplicas": 0
},
"params": {
To view the entire JSON payload, click View. |
All Search index payloads have the following properties:
Property | Type | Required? | Description |
---|---|---|---|
name |
String |
Yes |
The name of the Search index. A Search index name must be unique for each cluster. |
type |
String |
Yes |
The type of index you want to create:
|
String |
No |
The Search Service automatically generates a UUID for a Search index. If you use an existing UUID, the Search Service updates the existing Search index.
Don’t include the |
|
sourceType |
String |
Yes |
The |
sourceName |
String |
Yes |
The name of the bucket where you want to create the Search index. The Search Service automatically finds the UUID for the bucket. |
String |
No |
The UUID of the bucket where you want to create the Search index. The Search Service automatically finds the UUID for the bucket. Don’t include the |
|
sourceParams |
Object |
No |
This object contains advanced settings for index behavior. Don’t add content into this object unless instructed by Couchbase Support. |
planParams |
Object |
Yes |
An object that sets the Search index’s partitions and replications. For more information, see planParams Object. |
params |
Object |
Yes |
An object that sets the Search index’s type identifier, type mappings, and analyzers. For more information, see Params Object. |
planParams Object
The planParams
object sets a Search index’s partition and replication settings:
"planParams": {
"maxPartitionsPerPIndex": 1024,
"indexPartitions": 1,
"numReplicas": 0
},
To view the entire JSON payload, click View. |
The planParams
object contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
maxPartitionsPerPIndex |
n/a |
No |
This setting is deprecated. Use |
indexPartitions |
Number |
Yes |
The number of partitions to split the Search index into, across the nodes you have available in your database with the Search Service enabled. Use index partitions to increase index and query performance on large datasets. |
numReplicas |
Number |
Yes |
For high-availability, set the number of replicas the Search Service creates for the Search index. You can create up to three replicas for a Search index.
Each replica creates a full copy of the Search index to increase high-availability.
To turn off replication for the Search index, set The number of replicas you can create depends on the number of nodes you have available with the Search Service enabled. |
Params Object
The params
object sets a Search index’s type identifier, type mappings, and analyzers.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
doc_config |
Object |
Yes |
An object that sets how the Search index sets a document’s type. For more information, see Doc_config Object. |
mapping |
Object |
Yes |
An object that sets the analyzers and type mappings for a Search index. For more information, see Mapping Object. |
Doc_config Object
The doc_config
object sets how the Search index sets a document’s type:
"doc_config": {
"docid_prefix_delim": "",
"docid_regexp": "",
"mode": "scope.collection.type_field",
"type_field": "type"
},
To view the entire JSON payload, click View. |
The doc_config
object is a child object of the Params Object. It contains the following properties:
Property | Type | Required? | Description | ||
---|---|---|---|---|---|
mode |
String |
Yes |
Set a type identifier for the Search index to filter documents from search results:
|
||
docid_prefix_delim |
String |
Yes |
If For example, to filter documents based on the characters before a |
||
docid_regexp |
String |
Yes |
If For example, to filter documents that contain the characters |
||
type_field |
String |
Yes |
If For example, to filter documents based on the value of their |
Mapping Object
The mapping
object contains a Search index’s analyzers and other advanced settings from the UI:
"mapping": {
"analysis": {
"analyzers": {
"My_Analyzer": {
"token_filters": [
"apostrophe",
"My_Token_Filter"
],
"char_filters": [
"asciifolding",
"html",
"My_Char_Filter"
],
"type": "custom",
"tokenizer": "My_Tokenizer_Excep"
}
},
"char_filters": {
"My_Char_Filter": {
"regexp": "[']",
"replace": " ",
"type": "regexp"
}
},
"tokenizers": {
"My_Tokenizer_Excep": {
"exceptions": [
"[*]"
],
"tokenizer": "unicode",
"type": "exception"
},
"My_Tokenizer_RegExp": {
"regexp": "[*]",
"type": "regexp"
}
},
"token_filters": {
To view the entire JSON payload, click View. |
The mapping
object is a child object of the Params Object. It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
Object |
Yes |
An object that contains the following child objects: |
|
default_analyzer |
String |
Yes |
The name of the default analyzer to use for the Search index. For more information about analyzers, see Analyzers. |
default_datetime_parser |
String |
Yes |
The name of the default date/time parser to use for the Search index. For more information about date/time parsers, see Date/Time Parsers. |
String |
Yes |
Set a name for the If you enable the include_in_all property for a child field, the contents of that child field can be searched without specifying a field name or by specifying the default field’s name in your Search query. |
|
default_mapping |
Object |
No |
An object that contains settings for the default type mapping on the Search index. The default type mapping contains all documents under the This type mapping is included for compatibility only. For more information about the properties inside the |
default_type |
String |
No |
This setting is included for compatibility with earlier indexes only. |
docvalues_dynamic |
Boolean |
Yes |
To include the values for an indexed field in the Search index, set To exclude the values for an indexed field in the index, set |
index_dynamic |
Boolean |
Yes |
To index any fields in the Search index where To exclude dynamic fields from the index, set |
store_dynamic |
Boolean |
Yes |
To return the content from an indexed field in the Search index, set To exclude field content from the index, set |
type_field |
String |
No |
Use the same value assigned to the |
types |
Object |
No |
An object that contains any user-defined type mappings for the Search index, as For more information, see Types Object. |
Analyzers Object
The analyzers
object contains any custom analyzers defined for a Search index.
"analyzers": {
"My_Analyzer": {
"token_filters": [
"apostrophe",
"My_Token_Filter"
],
"char_filters": [
"asciifolding",
"html",
"My_Char_Filter"
],
"type": "custom",
"tokenizer": "My_Tokenizer_Excep"
}
},
To view the entire JSON payload, click View. |
The analyzers
object is a child object of the analysis object. It contains any number of {analyzer_name} objects:
Property | Type | Required? | Description |
---|---|---|---|
{analyzer_name} |
Object |
Yes |
Set the name of this object to the name you want for your custom analyzer. You can reference the For more information about the properties in an |
{Analyzer_name} Object
The {analyzer_name}
object defines a custom analyzer for a Search index:
"My_Analyzer": {
"token_filters": [
"apostrophe",
"My_Token_Filter"
],
"char_filters": [
"asciifolding",
"html",
"My_Char_Filter"
],
"type": "custom",
"tokenizer": "My_Tokenizer_Excep"
}
An {analyzer_name}
object is a child object of the Analyzers Object. It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
token_filters |
Array |
Yes |
An array of strings that contains the token filters for the custom analyzer. For more information about the token filters you can define in a Search index JSON payload, see Token_filters Object. You can also use one of the default token filters. |
char_filters |
Array |
Yes |
An array of strings that contains the character filters for the custom analyzer. For more information about the character filters you can define in a Search index JSON payload, see Char_filters Object. You can also use one of the default character filters available. |
type |
String |
Yes |
The |
tokenizer |
String |
Yes |
The selected tokenizer for the custom analyzer. |
Char_filters Object
The char_filters
object contains any custom character filters defined for a Search index:
"char_filters": {
"My_Char_Filter": {
"regexp": "[']",
"replace": " ",
"type": "regexp"
}
},
To view the entire JSON payload, click View. |
The char_filters
object is a child object of the analysis object. It contains any number of {char_filter_name} objects:
Property | Type | Required? | Description |
---|---|---|---|
{char_filter_name} |
Object |
Yes |
Set the name of this object to the name you want for your custom character filter. You can reference the For more information about the properties in an |
{Char_filter_name} Object
The {char_filter_name}
object defines a specific custom character filter for a Search index:
"My_Char_Filter": {
"regexp": "[']",
"replace": " ",
"type": "regexp"
}
A {char_filter_name}
object is a child object of the Char_filters Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
regexp |
String |
Yes |
The regular expression to use to filter characters from search queries and documents. |
replace |
String |
No |
The content to insert instead of the content in the |
type |
String |
Yes |
The |
Tokenizers Object
The tokenizers
object contains any custom tokenizers defined for a Search index:
"tokenizers": {
"My_Tokenizer_Excep": {
"exceptions": [
"[*]"
],
"tokenizer": "unicode",
"type": "exception"
},
"My_Tokenizer_RegExp": {
"regexp": "[*]",
"type": "regexp"
}
},
To view the entire JSON payload, click View. |
The tokenizers
object is a child object of the analysis object.
It contains any number of {tokenizer_name objects}:
Property | Type | Required? | Description |
---|---|---|---|
{tokenizer_name} |
Object |
Yes |
Set the name of this object to the name you want for your custom tokenizer. You can reference the For more information about the properties in an |
{Tokenizer_name} Object
The {tokenizer_name}
object defines a specific custom tokenizer for a Search index.
For example, the following My_Tokenizer_Excep
object defines an exception
tokenizer:
"My_Tokenizer_Excep": {
"exceptions": [
"[*]"
],
"tokenizer": "unicode",
"type": "exception"
},
A {tokenizer_name}
object is a child object of the Tokenizers Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
exceptions |
Array |
Yes |
If the tokenizer’s For example, if you add the characters |
regexp |
String |
Yes |
If the tokenizer’s The tokenizer takes any matches for the regular expression from the input text stream and uses them as tokens. For example, if you use the regular expression |
tokenizer |
String |
Yes |
If the tokenizer’s You can choose a default tokenizer or use a tokenizer defined in the |
String |
Yes |
The tokenizer’s type. Can be one of:
|
Token_filters Object
The token_filters
object contains any custom token filters defined for a Search index.
"token_filters": {
"My_Token_Filter": {
"min": 3,
"max": 255,
"type": "length"
}
},
To view the entire JSON payload, click View. |
The token_filters
object is a child object of the analysis object.
It contains any number of {token_filter_name} objects:
Property | Type | Required? | Description |
---|---|---|---|
{token_filter_name} |
Object |
Yes |
Set the name of this object to the name you want for your custom token filter. You can reference the For more information about the properties in an |
{Token_filter_name} Object
The {token_filter_name}
object defines a custom token filter for a Search index.
For example, the following My_Token_Filter
object defines a custom length
token filter:
"My_Token_Filter": {
"min": 3,
"max": 255,
"type": "length"
}
A {token_filter_name}
object is a child object of the Token_filters Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
type |
String |
Yes |
The token filter’s type. Can be one of:
|
Dict_compound Token Filters
A dict_compound
token filter uses a wordlist to find subwords inside an input token.
If the token filter finds a subword inside a compound word, it turns it into a separate token.
"My_Dict_Compound_Filter": {
"dict_token_map": "articles_ca",
"type": "dict_compound"
},
For example, if you had a wordlist that contained play
and jump
, the token filter converts playful jumping
into two tokens: play
and jump
.
Property | Type | Required? | Description |
---|---|---|---|
dict_token_map |
String |
Yes |
The wordlist to use to find subwords in existing tokens. You can use a default wordlist or one defined in the Token_maps Object. |
Edge_ngram Token Filters
An edge_ngram
token filter uses a specified range to create new tokens.
You can also choose whether to create the new token from the start or backward from the end of the input token.
"My_Edge_ngram_Filter": {
"back": false,
"min": 4,
"max": 5,
"type": "edge_ngram"
},
For example, if you had a miminum of four and a maximum of five with an input token of breweries
, the token filter creates the tokens brew
and brewe
.
Property | Type | Required? | Description |
---|---|---|---|
back |
Boolean |
Yes |
To create new tokens starting from the end and moving backward in an input token, set To create new tokens starting from the beginning and moving forward in an input token, set |
min |
Integer |
Yes |
Set the minimum character length for a new token. |
max |
Integer |
Yes |
Set the maximum character length for a new token. |
Elision Token Filters
An elision
token filter removes elisions from input tokens.
"My_Elision_Filter": {
"articles_token_map": "stop_fr",
"type": "elision"
},
For example, if you had the stop_fr
wordlist in an elision token filter, the token je m’appelle John
becomes the tokens je
, appelle
, and John
.
Property | Type | Required? | Description |
---|---|---|---|
articles_token_map |
String |
Yes |
The wordlist to use to find and remove elisions in existing tokens. You can use a default wordlist or one defined in the Token_maps Object. |
Keyword_marker Token Filters
A keyword_marker
token filter finds keywords in an input token and turns them into tokens.
"My_Keyword_Marker_Filter": {
"keywords_token_map": "articles_ca",
"type": "keyword_marker"
},
For example, if you had a wordlist that contained the keyword beer
, the token beer and breweries
becomes the token beer
.
Property | Type | Required? | Description |
---|---|---|---|
keywords_token_map |
String |
Yes |
The wordlist to use to find keywords in existing tokens. You can use a default wordlist or one defined in the Token_maps Object. |
Length Token Filters
A length
token filter removes tokens that are shorter or longer than a set character length.
"My_Length_Filter": {
"min": 2,
"max": 4,
"type": "length"
},
For example, if you had a range with a minimum of two and a maximum of four, the token beer and breweries
becomes the tokens beer
and and
.
Property | Type | Required? | Description |
---|---|---|---|
min |
Integer |
Yes |
The minimum character length for a new token from the token filter. |
max |
Integer |
Yes |
The maximum character length for a new token from the token filter. |
Ngram Token Filters
An ngram
token filter uses a specified character length to split an input token into new tokens.
"My_Ngram_Filter": {
"min": 4,
"max": 5,
"type": "ngram"
},
For example, if you had a range with a minimum of four and a maximum of five, the token beers
becomes the tokens beer
, beers
, and eers
.
Property | Type | Required? | Description |
---|---|---|---|
min |
Integer |
Yes |
The minimum character length for a new token from the token filter. |
max |
Integer |
Yes |
The maximum character length for a new token from the token filter. |
Normalize_unicode Token Filters
A normalize_unicode
token filter uses a specified Unicode Normalization form to create new tokens.
"My_Normalize_Unicode_Filter": {
"form": "nfkd",
"type": "normalize_unicode"
},
Property | Type | Required? | Description |
---|---|---|---|
form |
String |
Yes |
Select the form of Unicode Normalization to use on input tokens:
For more information about Unicode Normalization, see the Unicode Consortium’s Unicode Normalization Forms report. |
Shingle Token Filters
A shingle
token filter uses a specified character length and separator to create new tokens.
"My_Shingle_Filter":{
"min": 2,
"max": 3,
"output_original": true,
"separator": " ",
"filler": "x",
"type": "shingle"
},
For example, if you use a whitespace tokenizer, a range with a minimum of two and a maximum of three, and a space as a separator, the token abc def
becomes abc
, def
, and abc def
.
Property | Type | Required? | Description |
---|---|---|---|
min |
Integer |
Yes |
The minimum character length for a new token before concatenation. |
max |
Integer |
Yes |
The maximum character length for a new token before concatenation. |
output_original |
Boolean |
Yes |
To add the original token to the token filter’s output, set To exclude the original token from the token filter’s output, set |
separator |
String |
No |
Set a |
filler |
String |
No |
If another token filter removes a token from the input for this token filter, set a |
Stop_tokens Token Filters
A stop_tokens
token filter uses a wordlist to remove specific tokens from input.
"My_Stop_Tokens_Filter":{
"stop_token_map": "articles_ca",
"type": "stop_tokens"
},
For example, if you have a wordlist that contains the word and
, the token beers and breweries
becomes beers
and breweries
.
Property | Type | Required? | Description |
---|---|---|---|
stop_token_map |
String |
Yes |
The wordlist to use to filter tokens. The token filter removes any tokens from input that match an entry in the wordlist. You can use a default wordlist or one defined in the Token_maps Object. |
Truncate_token Token Filters
A truncate_token
token filter uses a specified character length to shorten any input tokens that are too long.
"My_Truncate_Token_Filter":{
"length": 4,
"type": "truncate_token"
}
For example, if you had a length
of four, the token beer and breweries
becomes beer
, and
, and brewe
.
Property | Type | Required? | Description |
---|---|---|---|
length |
Integer |
Yes |
The maximum character length for an output token. |
Token_maps Object
The token_maps
object contains any custom wordlists defined for a Search index:
"token_maps": {
"My_Wordlist": {
"type": "custom",
"tokens": [
"the",
"is",
"and"
]
}
},
To view the entire JSON payload, click View. |
The token_maps
object is a child object of the analysis object.
It contains any number of {wordlist_name} objects:
Property | Type | Required? | Description |
---|---|---|---|
{wordlist_name} |
Object |
Yes |
Set the name of this object to the name you want for your custom wordlist. You can reference the For more information about the properties in an |
{Wordlist_name} Object
The {wordlist_name}
object defines a custom wordlist for a Search index:
"My_Wordlist": {
"type": "custom",
"tokens": [
"the",
"is",
"and"
]
}
A {wordlist_name}
object is a child object of the Token_maps Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
type |
String |
Yes |
The |
tokens |
Array |
Yes |
An array of strings that contains each word added to the wordlist. |
Date_time_parsers Object
The date_time_parsers
object contains any custom date/time parsers defined for a Search index:
"date_time_parsers": {
"My_Date_Time_Parser": {
"type": "flexiblego",
"layouts": [
"RFC850"
]
}
}
To view the entire JSON payload, click View. |
The date_time_parsers
object is a child object of the analysis object.
It contains any number of {date_time_parser_name} objects:
Property | Type | Required? | Description |
---|---|---|---|
{date_time_parser_name} |
Object |
Yes |
Set the name of this object to the name you want for your custom date/time parser. You can reference the For more information about the properties in an |
{date_time_parser_name} Object
The {date_time_parser_name}
object defines a custom date/time parser for a Search index:
"My_Date_Time_Parser": {
"type": "flexiblego",
"layouts": [
"RFC850"
]
}
A {date_time_parser_name}
object is a child object of the Date_time_parsers Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
type |
String |
Yes |
The |
layouts |
Array |
Yes |
An array of strings that contains layouts for date and time fields. Use a layout from the Go Programming Language Time Package’s Layout Constant. |
Default_mapping Object
The default_mapping
object contains settings for the default type mapping on the Search index.
The default type mapping is a legacy feature and only included for compatibility.
"default_mapping": {
"dynamic": false,
"enabled": false
},
To view the entire JSON payload, click View. |
The default_mapping
object is a child object of the Mapping Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
dynamic |
Boolean |
Yes |
To index all available fields in a document with the default type mapping, set To only index the fields you specify in the type mapping, set |
enabled |
Boolean |
Yes |
To enable the Search Service’s default type mapping, set The default type mapping includes all documents in the bucket in the Search index, even if they don’t match another configured type mapping. This can increase index size and indexing time. To disable the default type mapping, set |
Types Object
The types
object contains any additional user-defined type mappings for a Search index.
"types": {
"inventory.hotel": {
"dynamic": false,
"enabled": true,
"properties": {
"reviews": {
"dynamic": false,
"enabled": true,
"properties": {
"content": {
"enabled": true,
"dynamic": false,
"fields": [
{
"docvalues": true,
"include_in_all": true,
"include_term_vectors": true,
"index": true,
"name": "content",
"store": true,
"type": "text",
"analyzer": "My_Analyzer"
}
]
}
}
},
"city": {
"enabled": true,
"dynamic": false,
"fields": [
{
"docvalues": true,
"include_in_all": true,
"include_term_vectors": true,
"index": true,
"name": "city",
"store": true,
"type": "text"
}
]
}
}
}
}
To view the entire JSON payload, click View. |
The types
object is a child object of the Mapping Object.
It contains any number of {scope}.{collection} objects:
Property | Type | Required? | Description | ||
---|---|---|---|---|---|
{scope}.{collection} |
Object |
Yes |
The name of the type mapping.
Corresponds to the selected scope and collection where the type mapping applies.
For example, For more information about the properties in an
|
{Scope}.{collection} Object
The {scope}.{collection}
object defines a custom type mapping for a Search index:
"inventory.hotel": {
"dynamic": false,
"enabled": true,
"properties": {
"reviews": {
"dynamic": false,
"enabled": true,
"properties": {
"content": {
"enabled": true,
"dynamic": false,
"fields": [
{
"docvalues": true,
"include_in_all": true,
"include_term_vectors": true,
"index": true,
"name": "content",
"store": true,
"type": "text",
"analyzer": "My_Analyzer"
}
]
}
}
},
"city": {
"enabled": true,
"dynamic": false,
"fields": [
{
"docvalues": true,
"include_in_all": true,
"include_term_vectors": true,
"index": true,
"name": "city",
"store": true,
"type": "text"
}
]
}
}
}
A {scope}.{collection}
object is a child object of the Types Object.
It contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
dynamic |
Boolean |
Yes |
To index all fields under the specified scope and collection, set To only index the fields you specify and enable the |
enabled |
Boolean |
Yes |
To enable the type mapping and include any documents that match it in the Search index, set To remove any documents that match this type mapping from the Search index, set |
Object |
No |
The Specifies properties for the fields to index in the type mapping. Contains any number of For more information, see {field_name} Object |
{field_name} Object
The {field_name}
object contains properties and an array for a child field in a type mapping.
You can have multiple {field_name}
objects in a properties object.
"reviews": {
"dynamic": false,
"enabled": true,
"properties": {
"content": {
"enabled": true,
"dynamic": false,
"fields": [
{
"docvalues": true,
"include_in_all": true,
"include_term_vectors": true,
"index": true,
"name": "content",
"store": true,
"type": "text",
"analyzer": "My_Analyzer"
}
]
}
}
},
To view the entire JSON payload, click View. |
The name of the object corresponds to the name of the field you want to include or exclude from your Search index.
A {field_name}
object contains the following properties:
Property | Type | Required? | Description |
---|---|---|---|
enabled |
Boolean |
Yes |
To add this child field to the Search index, set To remove this child field from the index, set |
dynamic |
Boolean |
No |
This field is included for legacy compatibility only. |
fields |
Array |
Yes |
An array that contains objects with settings for each child field to index in the type mapping. For more information, see Fields Array. |
Fields Array
The fields
array contains objects with settings for each child field to index in the type mapping:
"fields": [
{
"docvalues": true,
"include_in_all": true,
"include_term_vectors": true,
"index": true,
"name": "content",
"store": true,
"type": "text",
"analyzer": "My_Analyzer"
}
]
To view the entire JSON payload, click View. |
The fields
array is located inside a {field_name} object.
It contains the following properties:
Property | Type | Required? | Description | ||
---|---|---|---|---|---|
docvalues |
Boolean |
Yes |
To include the value for each instance of the field in the Search index to support facets and sorting search results, set To exclude the values for each instance of this field from the index, set |
||
Boolean |
Yes |
To allow this field to be searched without specifying the specific field’s name in the search, set When enabled, you can search this field through the specified To only search this field by specifying the field name, set |
|||
Boolean |
Yes |
To allow the Search Service to highlight matching search terms in search results for this field, set You must also enable term vectors to use To disable term highlighting and reduce index size, set |
|||
index |
Boolean |
Yes |
To include the child field in the Search index, set To exclude the child field from the index, set |
||
name |
String |
Yes |
The child field’s name. |
||
store |
Boolean |
Yes |
To include the content of the child field in the Search index and allow its content to be viewed in search results, set To exclude the content of the child field from the index, set |
||
type |
String |
Yes |
The child field’s type. Can be one of:
For more information about the available field data types, see Field Data Types. |
||
analyzer |
String |
No |
If the child field’s
|