diff --git a/output/openapi/elasticsearch-openapi.json b/output/openapi/elasticsearch-openapi.json index 1322e7da29..6e2aac9e96 100644 --- a/output/openapi/elasticsearch-openapi.json +++ b/output/openapi/elasticsearch-openapi.json @@ -53193,6 +53193,9 @@ "description": "The HTTP status code returned for the operation.", "type": "number" }, + "failure_store": { + "$ref": "#/components/schemas/_global.bulk:FailureStoreStatus" + }, "error": { "$ref": "#/components/schemas/_types:ErrorCause" }, @@ -53225,6 +53228,15 @@ "status" ] }, + "_global.bulk:FailureStoreStatus": { + "type": "string", + "enum": [ + "not_applicable_or_unknown", + "used", + "not_enabled", + "failed" + ] + }, "_types:InlineGetDictUserDefined": { "type": "object", "properties": { @@ -73943,7 +73955,7 @@ "index": { "type": "number" }, - "score": { + "relevance_score": { "type": "number" }, "text": { @@ -73952,7 +73964,7 @@ }, "required": [ "index", - "score" + "relevance_score" ] }, "_types:StreamResult": { @@ -88742,7 +88754,7 @@ } }, "ignored_fields": { - "description": "A list of the fields that would be ignored at the indexing step. For example, a field whose\nvalue is larger than the allowed limit would make it through all of the pipoelines, but\nwould not be indexed into Elasticsearch.", + "description": "A list of the fields that would be ignored at the indexing step. For example, a field whose\nvalue is larger than the allowed limit would make it through all of the pipelines, but\nwould not be indexed into Elasticsearch.", "type": "array", "items": { "type": "object", diff --git a/output/openapi/elasticsearch-serverless-openapi.json b/output/openapi/elasticsearch-serverless-openapi.json index cedc5cfb3e..85dce10084 100644 --- a/output/openapi/elasticsearch-serverless-openapi.json +++ b/output/openapi/elasticsearch-serverless-openapi.json @@ -33461,6 +33461,9 @@ "description": "The HTTP status code returned for the operation.", "type": "number" }, + "failure_store": { + "$ref": "#/components/schemas/_global.bulk:FailureStoreStatus" + }, "error": { "$ref": "#/components/schemas/_types:ErrorCause" }, @@ -33493,6 +33496,15 @@ "status" ] }, + "_global.bulk:FailureStoreStatus": { + "type": "string", + "enum": [ + "not_applicable_or_unknown", + "used", + "not_enabled", + "failed" + ] + }, "_types:InlineGetDictUserDefined": { "type": "object", "properties": { diff --git a/output/schema/schema-serverless.json b/output/schema/schema-serverless.json index 00aed97a76..34269f495f 100644 --- a/output/schema/schema-serverless.json +++ b/output/schema/schema-serverless.json @@ -92,11 +92,16 @@ "stability": "stable" } }, - "description": "Get the async search status.\n\nGet the status of a previously submitted async search request given its identifier, without retrieving search results.\nIf the Elasticsearch security features are enabled, use of this API is restricted to the `monitoring_user` role.", + "description": "Get the async search status.\n\nGet the status of a previously submitted async search request given its identifier, without retrieving search results.\nIf the Elasticsearch security features are enabled, the access to the status of a specific async search is restricted to:\n\n* The user or API key that submitted the original async search request.\n* Users that have the `monitor` cluster privilege or greater privileges.", "docId": "async-search", "docTag": "search", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/async-search.html", "name": "async_search.status", + "privileges": { + "cluster": [ + "monitor" + ] + }, "request": { "name": "Request", "namespace": "async_search.status" @@ -5208,7 +5213,7 @@ "stability": "stable" } }, - "description": "Close anomaly detection jobs.\nA job can be opened and closed multiple times throughout its lifecycle. A closed job cannot receive data or perform analysis operations, but you can still explore and navigate results.\nWhen you close a job, it runs housekeeping tasks such as pruning the model history, flushing buffers, calculating final results and persisting the model snapshots. Depending upon the size of the job, it could take several minutes to close and the equivalent time to re-open. After it is closed, the job has a minimal overhead on the cluster except for maintaining its meta data. Therefore it is a best practice to close jobs that are no longer required to process data.\nIf you close an anomaly detection job whose datafeed is running, the request first tries to stop the datafeed. This behavior is equivalent to calling stop datafeed API with the same timeout and force parameters as the close job request.\nWhen a datafeed that has a specified end date stops, it automatically closes its associated job.", + "description": "Close anomaly detection jobs.\n\nA job can be opened and closed multiple times throughout its lifecycle. A closed job cannot receive data or perform analysis operations, but you can still explore and navigate results.\nWhen you close a job, it runs housekeeping tasks such as pruning the model history, flushing buffers, calculating final results and persisting the model snapshots. Depending upon the size of the job, it could take several minutes to close and the equivalent time to re-open. After it is closed, the job has a minimal overhead on the cluster except for maintaining its meta data. Therefore it is a best practice to close jobs that are no longer required to process data.\nIf you close an anomaly detection job whose datafeed is running, the request first tries to stop the datafeed. This behavior is equivalent to calling stop datafeed API with the same timeout and force parameters as the close job request.\nWhen a datafeed that has a specified end date stops, it automatically closes its associated job.", "docId": "ml-close-job", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-close-job.html", @@ -5253,7 +5258,7 @@ "stability": "stable" } }, - "description": "Delete a calendar.\nRemoves all scheduled events from a calendar, then deletes it.", + "description": "Delete a calendar.\n\nRemove all scheduled events from a calendar, then delete it.", "docId": "ml-delete-calendar", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-delete-calendar.html", @@ -5458,7 +5463,7 @@ "stability": "stable" } }, - "description": "Delete a filter.\nIf an anomaly detection job references the filter, you cannot delete the\nfilter. You must update or delete the job before you can delete the filter.", + "description": "Delete a filter.\n\nIf an anomaly detection job references the filter, you cannot delete the\nfilter. You must update or delete the job before you can delete the filter.", "docId": "ml-delete-filter", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-delete-filter.html", @@ -5500,7 +5505,7 @@ "stability": "stable" } }, - "description": "Delete an anomaly detection job.\nAll job configuration, model state and results are deleted.\nIt is not currently possible to delete multiple jobs using wildcards or a\ncomma separated list. If you delete a job that has a datafeed, the request\nfirst tries to delete the datafeed. This behavior is equivalent to calling\nthe delete datafeed API with the same timeout and force parameters as the\ndelete job request.", + "description": "Delete an anomaly detection job.\n\nAll job configuration, model state and results are deleted.\nIt is not currently possible to delete multiple jobs using wildcards or a\ncomma separated list. If you delete a job that has a datafeed, the request\nfirst tries to delete the datafeed. This behavior is equivalent to calling\nthe delete datafeed API with the same timeout and force parameters as the\ndelete job request.", "docId": "ml-delete-job", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-delete-job.html", @@ -5542,7 +5547,7 @@ "stability": "stable" } }, - "description": "Delete an unreferenced trained model.\nThe request deletes a trained inference model that is not referenced by an ingest pipeline.", + "description": "Delete an unreferenced trained model.\n\nThe request deletes a trained inference model that is not referenced by an ingest pipeline.", "docId": "delete-trained-models", "docTag": "ml trained model", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/delete-trained-models.html", @@ -5584,7 +5589,7 @@ "stability": "stable" } }, - "description": "Delete a trained model alias.\nThis API deletes an existing model alias that refers to a trained model. If\nthe model alias is missing or refers to a model other than the one identified\nby the `model_id`, this API returns an error.", + "description": "Delete a trained model alias.\n\nThis API deletes an existing model alias that refers to a trained model. If\nthe model alias is missing or refers to a model other than the one identified\nby the `model_id`, this API returns an error.", "docId": "delete-trained-models-aliases", "docTag": "ml trained model", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/delete-trained-models-aliases.html", @@ -5629,7 +5634,7 @@ "stability": "stable" } }, - "description": "Estimate job model memory usage.\nMakes an estimation of the memory usage for an anomaly detection job model.\nIt is based on analysis configuration details for the job and cardinality\nestimates for the fields it references.", + "description": "Estimate job model memory usage.\n\nMake an estimation of the memory usage for an anomaly detection job model.\nThe estimate is based on analysis configuration details for the job and cardinality\nestimates for the fields it references.", "docId": "ml-estimate-memory", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-estimate-model-memory.html", @@ -5674,7 +5679,7 @@ "stability": "stable" } }, - "description": "Evaluate data frame analytics.\nThe API packages together commonly used evaluation metrics for various types\nof machine learning features. This has been designed for use on indexes\ncreated by data frame analytics. Evaluation requires both a ground truth\nfield and an analytics result field to be present.", + "description": "Evaluate data frame analytics.\n\nThe API packages together commonly used evaluation metrics for various types\nof machine learning features. This has been designed for use on indexes\ncreated by data frame analytics. Evaluation requires both a ground truth\nfield and an analytics result field to be present.", "docId": "evaluate-dfanalytics", "docTag": "ml data frame", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/evaluate-dfanalytics.html", @@ -6387,7 +6392,7 @@ "stability": "stable" } }, - "description": "Open anomaly detection jobs.\nAn anomaly detection job must be opened to be ready to receive and analyze\ndata. It can be opened and closed multiple times throughout its lifecycle.\nWhen you open a new job, it starts with an empty model.\nWhen you open an existing job, the most recent model state is automatically\nloaded. The job is ready to resume its analysis from where it left off, once\nnew data is received.", + "description": "Open anomaly detection jobs.\n\nAn anomaly detection job must be opened to be ready to receive and analyze\ndata. It can be opened and closed multiple times throughout its lifecycle.\nWhen you open a new job, it starts with an empty model.\nWhen you open an existing job, the most recent model state is automatically\nloaded. The job is ready to resume its analysis from where it left off, once\nnew data is received.", "docId": "ml-open-job", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-open-job.html", @@ -6818,7 +6823,7 @@ "stability": "stable" } }, - "description": "Create an anomaly detection job.\nIf you include a `datafeed_config`, you must have read index privileges on the source index.\nIf you include a `datafeed_config` but do not provide a query, the datafeed uses `{\"match_all\": {\"boost\": 1}}`.", + "description": "Create an anomaly detection job.\n\nIf you include a `datafeed_config`, you must have read index privileges on the source index.\nIf you include a `datafeed_config` but do not provide a query, the datafeed uses `{\"match_all\": {\"boost\": 1}}`.", "docId": "ml-put-job", "docTag": "ml anomaly", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/ml-put-job.html", @@ -10046,7 +10051,7 @@ "stability": "stable" } }, - "description": "Get transform stats.\nGet usage information for transforms.", + "description": "Get transform stats.\n\nGet usage information for transforms.", "docId": "get-transform-stats", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/get-transform-stats.html", "name": "transform.get_transform_stats", @@ -10197,7 +10202,7 @@ "stability": "stable" } }, - "description": "Reset a transform.\nResets a transform.\nBefore you can reset it, you must stop it; alternatively, use the `force` query parameter.\nIf the destination index was created by the transform, it is deleted.", + "description": "Reset a transform.\n\nBefore you can reset it, you must stop it; alternatively, use the `force` query parameter.\nIf the destination index was created by the transform, it is deleted.", "docId": "reset-transform", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/reset-transform.html", "name": "transform.reset_transform", @@ -10238,7 +10243,7 @@ "stability": "stable" } }, - "description": "Schedule a transform to start now.\nInstantly runs a transform to process data.\n\nIf you _schedule_now a transform, it will process the new data instantly,\nwithout waiting for the configured frequency interval. After _schedule_now API is called,\nthe transform will be processed again at now + frequency unless _schedule_now API\nis called again in the meantime.", + "description": "Schedule a transform to start now.\n\nInstantly run a transform to process data.\nIf you run this API, the transform will process the new data instantly,\nwithout waiting for the configured frequency interval. After the API is called,\nthe transform will be processed again at `now + frequency` unless the API\nis called again in the meantime.", "docId": "schedule-now-transform", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/schedule-now-transform.html", "name": "transform.schedule_now_transform", @@ -10282,7 +10287,7 @@ "stability": "stable" } }, - "description": "Start a transform.\nStarts a transform.\n\nWhen you start a transform, it creates the destination index if it does not already exist. The `number_of_shards` is\nset to `1` and the `auto_expand_replicas` is set to `0-1`. If it is a pivot transform, it deduces the mapping\ndefinitions for the destination index from the source indices and the transform aggregations. If fields in the\ndestination index are derived from scripts (as in the case of `scripted_metric` or `bucket_script` aggregations),\nthe transform uses dynamic mappings unless an index template exists. If it is a latest transform, it does not deduce\nmapping definitions; it uses dynamic mappings. To use explicit mappings, create the destination index before you\nstart the transform. Alternatively, you can create an index template, though it does not affect the deduced mappings\nin a pivot transform.\n\nWhen the transform starts, a series of validations occur to ensure its success. If you deferred validation when you\ncreated the transform, they occur when you start the transform—​with the exception of privilege checks. When\nElasticsearch security features are enabled, the transform remembers which roles the user that created it had at the\ntime of creation and uses those same roles. If those roles do not have the required privileges on the source and\ndestination indices, the transform fails when it attempts unauthorized operations.", + "description": "Start a transform.\n\nWhen you start a transform, it creates the destination index if it does not already exist. The `number_of_shards` is\nset to `1` and the `auto_expand_replicas` is set to `0-1`. If it is a pivot transform, it deduces the mapping\ndefinitions for the destination index from the source indices and the transform aggregations. If fields in the\ndestination index are derived from scripts (as in the case of `scripted_metric` or `bucket_script` aggregations),\nthe transform uses dynamic mappings unless an index template exists. If it is a latest transform, it does not deduce\nmapping definitions; it uses dynamic mappings. To use explicit mappings, create the destination index before you\nstart the transform. Alternatively, you can create an index template, though it does not affect the deduced mappings\nin a pivot transform.\n\nWhen the transform starts, a series of validations occur to ensure its success. If you deferred validation when you\ncreated the transform, they occur when you start the transform—​with the exception of privilege checks. When\nElasticsearch security features are enabled, the transform remembers which roles the user that created it had at the\ntime of creation and uses those same roles. If those roles do not have the required privileges on the source and\ndestination indices, the transform fails when it attempts unauthorized operations.", "docId": "start-transform", "docUrl": "https://www.elastic.co/guide/en/elasticsearch/reference/{branch}/start-transform.html", "name": "transform.start_transform", @@ -10587,7 +10592,7 @@ ], "query": [ { - "description": "Specifies how long the async search should be available in the cluster.\nWhen not specified, the `keep_alive` set with the corresponding submit async request will be used.\nOtherwise, it is possible to override the value and extend the validity of the request.\nWhen this period expires, the search, if still running, is cancelled.\nIf the search is completed, its saved results are deleted.", + "description": "The length of time that the async search should be available in the cluster.\nWhen not specified, the `keep_alive` set with the corresponding submit async request will be used.\nOtherwise, it is possible to override the value and extend the validity of the request.\nWhen this period expires, the search, if still running, is cancelled.\nIf the search is completed, its saved results are deleted.", "name": "keep_alive", "required": false, "type": { @@ -10665,7 +10670,7 @@ "body": { "kind": "no_body" }, - "description": "Get the async search status.\n\nGet the status of a previously submitted async search request given its identifier, without retrieving search results.\nIf the Elasticsearch security features are enabled, use of this API is restricted to the `monitoring_user` role.", + "description": "Get the async search status.\n\nGet the status of a previously submitted async search request given its identifier, without retrieving search results.\nIf the Elasticsearch security features are enabled, the access to the status of a specific async search is restricted to:\n\n* The user or API key that submitted the original async search request.\n* Users that have the `monitor` cluster privilege or greater privileges.", "inherits": { "type": { "name": "RequestBase", @@ -10693,7 +10698,7 @@ ], "query": [ { - "description": "Specifies how long the async search needs to be available.\nOngoing async searches and any saved search results are deleted after this period.", + "description": "The length of time that the async search needs to be available.\nOngoing async searches and any saved search results are deleted after this period.", "name": "keep_alive", "required": false, "serverDefault": "5d", @@ -10706,7 +10711,7 @@ } } ], - "specLocation": "async_search/status/AsyncSearchStatusRequest.ts#L24-L54" + "specLocation": "async_search/status/AsyncSearchStatusRequest.ts#L24-L58" }, { "body": { @@ -18202,8 +18207,10 @@ } }, { + "description": "Allow query execution also in case of shard failures.\nIf true, the query will keep running and will return results based on the available shards.\nFor sequences, the behavior can be further refined using allow_partial_sequence_results", "name": "allow_partial_search_results", "required": false, + "serverDefault": true, "type": { "kind": "instance_of", "type": { @@ -18213,8 +18220,10 @@ } }, { + "description": "This flag applies only to sequences and has effect only if allow_partial_search_results=true.\nIf true, the sequence query will return results based on the available shards, ignoring the others.\nIf false, the sequence query will return successfully, but will always have empty results.", "name": "allow_partial_sequence_results", "required": false, + "serverDefault": false, "type": { "kind": "instance_of", "type": { @@ -18437,7 +18446,7 @@ } } ], - "specLocation": "eql/search/EqlSearchRequest.ts#L28-L149" + "specLocation": "eql/search/EqlSearchRequest.ts#L28-L161" }, { "body": { @@ -18590,9 +18599,10 @@ } }, { - "description": "If provided and `true`, the response will include an extra `_clusters` object with information\nabout the clusters that participated in the search along with info such as shards count.", + "description": "When set to `true` and performing a cross-cluster query, the response will include an extra `_clusters`\nobject with information about the clusters that participated in the search along with info such as shards\ncount.", "name": "include_ccs_metadata", "required": false, + "serverDefault": false, "type": { "kind": "instance_of", "type": { @@ -18655,7 +18665,7 @@ } } ], - "specLocation": "esql/query/QueryRequest.ts#L27-L103" + "specLocation": "esql/query/QueryRequest.ts#L27-L105" }, { "body": { @@ -27360,7 +27370,7 @@ } ] }, - "description": "Close anomaly detection jobs.\nA job can be opened and closed multiple times throughout its lifecycle. A closed job cannot receive data or perform analysis operations, but you can still explore and navigate results.\nWhen you close a job, it runs housekeeping tasks such as pruning the model history, flushing buffers, calculating final results and persisting the model snapshots. Depending upon the size of the job, it could take several minutes to close and the equivalent time to re-open. After it is closed, the job has a minimal overhead on the cluster except for maintaining its meta data. Therefore it is a best practice to close jobs that are no longer required to process data.\nIf you close an anomaly detection job whose datafeed is running, the request first tries to stop the datafeed. This behavior is equivalent to calling stop datafeed API with the same timeout and force parameters as the close job request.\nWhen a datafeed that has a specified end date stops, it automatically closes its associated job.", + "description": "Close anomaly detection jobs.\n\nA job can be opened and closed multiple times throughout its lifecycle. A closed job cannot receive data or perform analysis operations, but you can still explore and navigate results.\nWhen you close a job, it runs housekeeping tasks such as pruning the model history, flushing buffers, calculating final results and persisting the model snapshots. Depending upon the size of the job, it could take several minutes to close and the equivalent time to re-open. After it is closed, the job has a minimal overhead on the cluster except for maintaining its meta data. Therefore it is a best practice to close jobs that are no longer required to process data.\nIf you close an anomaly detection job whose datafeed is running, the request first tries to stop the datafeed. This behavior is equivalent to calling stop datafeed API with the same timeout and force parameters as the close job request.\nWhen a datafeed that has a specified end date stops, it automatically closes its associated job.", "inherits": { "type": { "name": "RequestBase", @@ -27427,7 +27437,7 @@ } } ], - "specLocation": "ml/close_job/MlCloseJobRequest.ts#L24-L84" + "specLocation": "ml/close_job/MlCloseJobRequest.ts#L24-L85" }, { "body": { @@ -27460,7 +27470,7 @@ "body": { "kind": "no_body" }, - "description": "Delete a calendar.\nRemoves all scheduled events from a calendar, then deletes it.", + "description": "Delete a calendar.\n\nRemove all scheduled events from a calendar, then delete it.", "inherits": { "type": { "name": "RequestBase", @@ -27487,7 +27497,7 @@ } ], "query": [], - "specLocation": "ml/delete_calendar/MlDeleteCalendarRequest.ts#L23-L44" + "specLocation": "ml/delete_calendar/MlDeleteCalendarRequest.ts#L23-L45" }, { "body": { @@ -27825,7 +27835,7 @@ "body": { "kind": "no_body" }, - "description": "Delete a filter.\nIf an anomaly detection job references the filter, you cannot delete the\nfilter. You must update or delete the job before you can delete the filter.", + "description": "Delete a filter.\n\nIf an anomaly detection job references the filter, you cannot delete the\nfilter. You must update or delete the job before you can delete the filter.", "inherits": { "type": { "name": "RequestBase", @@ -27852,7 +27862,7 @@ } ], "query": [], - "specLocation": "ml/delete_filter/MlDeleteFilterRequest.ts#L23-L47" + "specLocation": "ml/delete_filter/MlDeleteFilterRequest.ts#L23-L48" }, { "body": { @@ -27879,7 +27889,7 @@ "body": { "kind": "no_body" }, - "description": "Delete an anomaly detection job.\nAll job configuration, model state and results are deleted.\nIt is not currently possible to delete multiple jobs using wildcards or a\ncomma separated list. If you delete a job that has a datafeed, the request\nfirst tries to delete the datafeed. This behavior is equivalent to calling\nthe delete datafeed API with the same timeout and force parameters as the\ndelete job request.", + "description": "Delete an anomaly detection job.\n\nAll job configuration, model state and results are deleted.\nIt is not currently possible to delete multiple jobs using wildcards or a\ncomma separated list. If you delete a job that has a datafeed, the request\nfirst tries to delete the datafeed. This behavior is equivalent to calling\nthe delete datafeed API with the same timeout and force parameters as the\ndelete job request.", "inherits": { "type": { "name": "RequestBase", @@ -27945,7 +27955,7 @@ } } ], - "specLocation": "ml/delete_job/MlDeleteJobRequest.ts#L23-L71" + "specLocation": "ml/delete_job/MlDeleteJobRequest.ts#L23-L72" }, { "body": { @@ -27972,7 +27982,7 @@ "body": { "kind": "no_body" }, - "description": "Delete an unreferenced trained model.\nThe request deletes a trained inference model that is not referenced by an ingest pipeline.", + "description": "Delete an unreferenced trained model.\n\nThe request deletes a trained inference model that is not referenced by an ingest pipeline.", "inherits": { "type": { "name": "RequestBase", @@ -28025,7 +28035,7 @@ } } ], - "specLocation": "ml/delete_trained_model/MlDeleteTrainedModelRequest.ts#L24-L56" + "specLocation": "ml/delete_trained_model/MlDeleteTrainedModelRequest.ts#L24-L57" }, { "body": { @@ -28052,7 +28062,7 @@ "body": { "kind": "no_body" }, - "description": "Delete a trained model alias.\nThis API deletes an existing model alias that refers to a trained model. If\nthe model alias is missing or refers to a model other than the one identified\nby the `model_id`, this API returns an error.", + "description": "Delete a trained model alias.\n\nThis API deletes an existing model alias that refers to a trained model. If\nthe model alias is missing or refers to a model other than the one identified\nby the `model_id`, this API returns an error.", "inherits": { "type": { "name": "RequestBase", @@ -28091,7 +28101,7 @@ } ], "query": [], - "specLocation": "ml/delete_trained_model_alias/MlDeleteTrainedModelAliasRequest.ts#L23-L52" + "specLocation": "ml/delete_trained_model_alias/MlDeleteTrainedModelAliasRequest.ts#L23-L53" }, { "body": { @@ -28178,7 +28188,7 @@ } ] }, - "description": "Estimate job model memory usage.\nMakes an estimation of the memory usage for an anomaly detection job model.\nIt is based on analysis configuration details for the job and cardinality\nestimates for the fields it references.", + "description": "Estimate job model memory usage.\n\nMake an estimation of the memory usage for an anomaly detection job model.\nThe estimate is based on analysis configuration details for the job and cardinality\nestimates for the fields it references.", "inherits": { "type": { "name": "RequestBase", @@ -28192,7 +28202,7 @@ }, "path": [], "query": [], - "specLocation": "ml/estimate_model_memory/MlEstimateModelMemoryRequest.ts#L26-L70" + "specLocation": "ml/estimate_model_memory/MlEstimateModelMemoryRequest.ts#L26-L71" }, { "body": { @@ -28265,7 +28275,7 @@ } ] }, - "description": "Evaluate data frame analytics.\nThe API packages together commonly used evaluation metrics for various types\nof machine learning features. This has been designed for use on indexes\ncreated by data frame analytics. Evaluation requires both a ground truth\nfield and an analytics result field to be present.", + "description": "Evaluate data frame analytics.\n\nThe API packages together commonly used evaluation metrics for various types\nof machine learning features. This has been designed for use on indexes\ncreated by data frame analytics. Evaluation requires both a ground truth\nfield and an analytics result field to be present.", "inherits": { "type": { "name": "RequestBase", @@ -28279,7 +28289,7 @@ }, "path": [], "query": [], - "specLocation": "ml/evaluate_data_frame/MlEvaluateDataFrameRequest.ts#L25-L60" + "specLocation": "ml/evaluate_data_frame/MlEvaluateDataFrameRequest.ts#L25-L61" }, { "body": { @@ -30233,7 +30243,7 @@ } ] }, - "description": "Open anomaly detection jobs.\nAn anomaly detection job must be opened to be ready to receive and analyze\ndata. It can be opened and closed multiple times throughout its lifecycle.\nWhen you open a new job, it starts with an empty model.\nWhen you open an existing job, the most recent model state is automatically\nloaded. The job is ready to resume its analysis from where it left off, once\nnew data is received.", + "description": "Open anomaly detection jobs.\n\nAn anomaly detection job must be opened to be ready to receive and analyze\ndata. It can be opened and closed multiple times throughout its lifecycle.\nWhen you open a new job, it starts with an empty model.\nWhen you open an existing job, the most recent model state is automatically\nloaded. The job is ready to resume its analysis from where it left off, once\nnew data is received.", "inherits": { "type": { "name": "RequestBase", @@ -30274,7 +30284,7 @@ } } ], - "specLocation": "ml/open_job/MlOpenJobRequest.ts#L24-L66" + "specLocation": "ml/open_job/MlOpenJobRequest.ts#L24-L67" }, { "body": { @@ -31980,7 +31990,7 @@ } ] }, - "description": "Create an anomaly detection job.\nIf you include a `datafeed_config`, you must have read index privileges on the source index.\nIf you include a `datafeed_config` but do not provide a query, the datafeed uses `{\"match_all\": {\"boost\": 1}}`.", + "description": "Create an anomaly detection job.\n\nIf you include a `datafeed_config`, you must have read index privileges on the source index.\nIf you include a `datafeed_config` but do not provide a query, the datafeed uses `{\"match_all\": {\"boost\": 1}}`.", "inherits": { "type": { "name": "RequestBase", @@ -32064,7 +32074,7 @@ } } ], - "specLocation": "ml/put_job/MlPutJobRequest.ts#L30-L156" + "specLocation": "ml/put_job/MlPutJobRequest.ts#L30-L157" }, { "body": { @@ -43930,7 +43940,7 @@ "body": { "kind": "no_body" }, - "description": "Get transform stats.\nGet usage information for transforms.", + "description": "Get transform stats.\n\nGet usage information for transforms.", "inherits": { "type": { "name": "RequestBase", @@ -44009,7 +44019,7 @@ } } ], - "specLocation": "transform/get_transform_stats/GetTransformStatsRequest.ts#L25-L76" + "specLocation": "transform/get_transform_stats/GetTransformStatsRequest.ts#L25-L77" }, { "body": { @@ -44465,7 +44475,7 @@ "body": { "kind": "no_body" }, - "description": "Reset a transform.\nResets a transform.\nBefore you can reset it, you must stop it; alternatively, use the `force` query parameter.\nIf the destination index was created by the transform, it is deleted.", + "description": "Reset a transform.\n\nBefore you can reset it, you must stop it; alternatively, use the `force` query parameter.\nIf the destination index was created by the transform, it is deleted.", "inherits": { "type": { "name": "RequestBase", @@ -44546,7 +44556,7 @@ "body": { "kind": "no_body" }, - "description": "Schedule a transform to start now.\nInstantly runs a transform to process data.\n\nIf you _schedule_now a transform, it will process the new data instantly,\nwithout waiting for the configured frequency interval. After _schedule_now API is called,\nthe transform will be processed again at now + frequency unless _schedule_now API\nis called again in the meantime.", + "description": "Schedule a transform to start now.\n\nInstantly run a transform to process data.\nIf you run this API, the transform will process the new data instantly,\nwithout waiting for the configured frequency interval. After the API is called,\nthe transform will be processed again at `now + frequency` unless the API\nis called again in the meantime.", "inherits": { "type": { "name": "RequestBase", @@ -44614,7 +44624,7 @@ "body": { "kind": "no_body" }, - "description": "Start a transform.\nStarts a transform.\n\nWhen you start a transform, it creates the destination index if it does not already exist. The `number_of_shards` is\nset to `1` and the `auto_expand_replicas` is set to `0-1`. If it is a pivot transform, it deduces the mapping\ndefinitions for the destination index from the source indices and the transform aggregations. If fields in the\ndestination index are derived from scripts (as in the case of `scripted_metric` or `bucket_script` aggregations),\nthe transform uses dynamic mappings unless an index template exists. If it is a latest transform, it does not deduce\nmapping definitions; it uses dynamic mappings. To use explicit mappings, create the destination index before you\nstart the transform. Alternatively, you can create an index template, though it does not affect the deduced mappings\nin a pivot transform.\n\nWhen the transform starts, a series of validations occur to ensure its success. If you deferred validation when you\ncreated the transform, they occur when you start the transform—​with the exception of privilege checks. When\nElasticsearch security features are enabled, the transform remembers which roles the user that created it had at the\ntime of creation and uses those same roles. If those roles do not have the required privileges on the source and\ndestination indices, the transform fails when it attempts unauthorized operations.", + "description": "Start a transform.\n\nWhen you start a transform, it creates the destination index if it does not already exist. The `number_of_shards` is\nset to `1` and the `auto_expand_replicas` is set to `0-1`. If it is a pivot transform, it deduces the mapping\ndefinitions for the destination index from the source indices and the transform aggregations. If fields in the\ndestination index are derived from scripts (as in the case of `scripted_metric` or `bucket_script` aggregations),\nthe transform uses dynamic mappings unless an index template exists. If it is a latest transform, it does not deduce\nmapping definitions; it uses dynamic mappings. To use explicit mappings, create the destination index before you\nstart the transform. Alternatively, you can create an index template, though it does not affect the deduced mappings\nin a pivot transform.\n\nWhen the transform starts, a series of validations occur to ensure its success. If you deferred validation when you\ncreated the transform, they occur when you start the transform—​with the exception of privilege checks. When\nElasticsearch security features are enabled, the transform remembers which roles the user that created it had at the\ntime of creation and uses those same roles. If those roles do not have the required privileges on the source and\ndestination indices, the transform fails when it attempts unauthorized operations.", "inherits": { "type": { "name": "RequestBase", @@ -44667,7 +44677,7 @@ } } ], - "specLocation": "transform/start_transform/StartTransformRequest.ts#L24-L73" + "specLocation": "transform/start_transform/StartTransformRequest.ts#L24-L72" }, { "body": { @@ -46222,6 +46232,28 @@ }, "specLocation": "_global/update_by_query/UpdateByQueryResponse.ts#L26-L67" }, + { + "kind": "enum", + "members": [ + { + "name": "not_applicable_or_unknown" + }, + { + "name": "used" + }, + { + "name": "not_enabled" + }, + { + "name": "failed" + } + ], + "name": { + "name": "FailureStoreStatus", + "namespace": "_global.bulk" + }, + "specLocation": "_global/bulk/types.ts#L86-L91" + }, { "kind": "enum", "members": [ @@ -46242,7 +46274,7 @@ "name": "OperationType", "namespace": "_global.bulk" }, - "specLocation": "_global/bulk/types.ts#L85-L90" + "specLocation": "_global/bulk/types.ts#L93-L98" }, { "kind": "enum", @@ -93649,7 +93681,7 @@ "name": "FollowerIndexStatus", "namespace": "ccr.follow_info" }, - "specLocation": "ccr/follow_info/types.ts#L32-L35" + "specLocation": "ccr/follow_info/types.ts#L37-L40" }, { "kind": "enum", @@ -93989,7 +94021,7 @@ } }, { - "description": "This field presents and has value of ::es_redacted:: only when the remote cluster is configured with the API key based model. Otherwise, the field is not present.", + "description": "This field is present and has a value of `::es_redacted::` only when the remote cluster is configured with the API key based model. Otherwise, the field is not present.", "name": "cluster_credentials", "required": false, "type": { @@ -108675,7 +108707,7 @@ } } ], - "specLocation": "async_search/_types/AsyncSearchResponseBase.ts#L50-L54" + "specLocation": "async_search/_types/AsyncSearchResponseBase.ts#L52-L56" }, { "kind": "interface", @@ -108708,7 +108740,7 @@ } }, { - "description": "Indicates whether the search is still running or has completed.\nNOTE: If the search failed after some shards returned their results or the node that is coordinating the async search dies, results may be partial even though `is_running` is `false`.", + "description": "Indicates whether the search is still running or has completed.\n\n> info\n> If the search failed after some shards returned their results or the node that is coordinating the async search dies, results may be partial even though `is_running` is `false`.", "name": "is_running", "required": true, "type": { @@ -108783,7 +108815,7 @@ } }, { - "description": "Indicates when the async search completed. Only present\nwhen the search has completed.", + "description": "Indicates when the async search completed.\nIt is present only when the search has completed.", "name": "completion_time", "required": false, "type": { @@ -108815,7 +108847,7 @@ } } ], - "specLocation": "async_search/_types/AsyncSearchResponseBase.ts#L24-L49" + "specLocation": "async_search/_types/AsyncSearchResponseBase.ts#L24-L51" }, { "generics": [ @@ -108831,7 +108863,7 @@ }, "properties": [ { - "description": "Partial aggregations results, coming from the shards that have already completed the execution of the query.", + "description": "Partial aggregations results, coming from the shards that have already completed running the query.", "name": "aggregations", "required": false, "type": { @@ -109054,7 +109086,7 @@ }, "properties": [ { - "description": "Indicates how many shards have run the query so far.", + "description": "The number of shards that have run the query so far.", "name": "_shards", "required": true, "type": { @@ -109066,7 +109098,7 @@ } }, { - "description": "Metadata about clusters involved in the cross-cluster search.\nNot shown for local-only searches.", + "description": "Metadata about clusters involved in the cross-cluster search.\nIt is not shown for local-only searches.", "name": "_clusters", "required": false, "type": { @@ -109078,7 +109110,7 @@ } }, { - "description": "If the async search completed, this field shows the status code of the search.\nFor example, 200 indicates that the async search was successfully completed.\n503 indicates that the async search was completed with an error.", + "description": "If the async search completed, this field shows the status code of the search.\nFor example, `200` indicates that the async search was successfully completed.\n`503` indicates that the async search was completed with an error.", "name": "completion_status", "required": false, "type": { @@ -109191,7 +109223,7 @@ } } ], - "specLocation": "_global/bulk/types.ts#L150-L172", + "specLocation": "_global/bulk/types.ts#L158-L180", "variants": { "kind": "container" } @@ -109209,7 +109241,7 @@ "namespace": "_global.bulk" }, "properties": [], - "specLocation": "_global/bulk/types.ts#L134-L134" + "specLocation": "_global/bulk/types.ts#L142-L142" }, { "inherits": { @@ -109273,7 +109305,7 @@ } } ], - "specLocation": "_global/bulk/types.ts#L111-L130" + "specLocation": "_global/bulk/types.ts#L119-L138" }, { "kind": "interface", @@ -109363,7 +109395,7 @@ } } ], - "specLocation": "_global/bulk/types.ts#L92-L109" + "specLocation": "_global/bulk/types.ts#L100-L117" }, { "inherits": { @@ -109378,7 +109410,7 @@ "namespace": "_global.bulk" }, "properties": [], - "specLocation": "_global/bulk/types.ts#L132-L132" + "specLocation": "_global/bulk/types.ts#L140-L140" }, { "inherits": { @@ -109419,7 +109451,7 @@ } } ], - "specLocation": "_global/bulk/types.ts#L138-L148" + "specLocation": "_global/bulk/types.ts#L146-L156" }, { "inherits": { @@ -109434,7 +109466,7 @@ "namespace": "_global.bulk" }, "properties": [], - "specLocation": "_global/bulk/types.ts#L136-L136" + "specLocation": "_global/bulk/types.ts#L144-L144" }, { "generics": [ @@ -109542,7 +109574,7 @@ } } ], - "specLocation": "_global/bulk/types.ts#L174-L209" + "specLocation": "_global/bulk/types.ts#L182-L217" }, { "kind": "interface", @@ -109599,6 +109631,17 @@ } } }, + { + "name": "failure_store", + "required": false, + "type": { + "kind": "instance_of", + "type": { + "name": "FailureStoreStatus", + "namespace": "_global.bulk" + } + } + }, { "description": "Additional information about the failed operation.\nThe property is returned only for failed operations.", "name": "error", @@ -109710,7 +109753,7 @@ } } ], - "specLocation": "_global/bulk/types.ts#L37-L83" + "specLocation": "_global/bulk/types.ts#L37-L84" }, { "attachedBehaviors": [ @@ -124097,7 +124140,7 @@ "specLocation": "inference/_types/Results.ts#L60-L65" }, { - "description": "The rerank result object representing a single ranked document\nid: the original index of the document in the request\nscore: the score of the document relative to the query\ntext: Optional, the text of the document, if requested", + "description": "The rerank result object representing a single ranked document\nid: the original index of the document in the request\nrelevance_score: the relevance_score of the document relative to the query\ntext: Optional, the text of the document, if requested", "kind": "interface", "name": { "name": "RankedDocument", @@ -124116,7 +124159,7 @@ } }, { - "name": "score", + "name": "relevance_score", "required": true, "type": { "kind": "instance_of", diff --git a/output/schema/schema.json b/output/schema/schema.json index 4e16da93b8..778cba0007 100644 --- a/output/schema/schema.json +++ b/output/schema/schema.json @@ -208230,7 +208230,7 @@ } }, { - "description": "A list of the fields that would be ignored at the indexing step. For example, a field whose\nvalue is larger than the allowed limit would make it through all of the pipoelines, but\nwould not be indexed into Elasticsearch.", + "description": "A list of the fields that would be ignored at the indexing step. For example, a field whose\nvalue is larger than the allowed limit would make it through all of the pipelines, but\nwould not be indexed into Elasticsearch.", "name": "ignored_fields", "required": false, "type": { diff --git a/output/schema/validation-errors.json b/output/schema/validation-errors.json index 90af09b8e8..9bde8d02d1 100644 --- a/output/schema/validation-errors.json +++ b/output/schema/validation-errors.json @@ -39,6 +39,7 @@ "async_search.submit": { "request": [ "Request: query parameter 'min_compatible_shard_node' does not exist in the json spec", + "Request: missing json spec query parameter 'keep_alive'", "interface definition _types:QueryVectorBuilder - Property text_embedding is a single-variant and must be required" ], "response": [] @@ -394,6 +395,14 @@ ], "response": [] }, + "indices.resolve_cluster": { + "request": [ + "Request: different number of urls in the json spec", + "Request: path parameter 'name' is required in the json spec", + "Request: query parameter 'timeout' does not exist in the json spec" + ], + "response": [] + }, "indices.rollover": { "request": [ "Request: missing json spec query parameter 'lazy'" @@ -413,6 +422,13 @@ ], "response": [] }, + "inference.update": { + "request": [ + "/_inference/{inference_id}/_update: different http methods in the json spec", + "/_inference/{task_type}/{inference_id}/_update: different http methods in the json spec" + ], + "response": [] + }, "ingest.delete_geoip_database": { "request": [ "Request: query parameter 'master_timeout' does not exist in the json spec", diff --git a/specification/_json_spec/async_search.submit.json b/specification/_json_spec/async_search.submit.json index 2d5ad25831..5bff2b1606 100644 --- a/specification/_json_spec/async_search.submit.json +++ b/specification/_json_spec/async_search.submit.json @@ -39,6 +39,11 @@ "description": "Control whether the response should be stored in the cluster if it completed within the provided [wait_for_completion] time (default: false)", "default": false }, + "keep_alive": { + "type": "time", + "description": "Update the time interval in which the results (partial or final) for this search will be available", + "default": "5d" + }, "batched_reduce_size": { "type": "number", "description": "The number of shard results that should be reduced at once on the coordinating node. This value should be used as the granularity at which progress results will be made available.", diff --git a/specification/_json_spec/eql.search.json b/specification/_json_spec/eql.search.json index 3818508320..1d9831a6e8 100644 --- a/specification/_json_spec/eql.search.json +++ b/specification/_json_spec/eql.search.json @@ -42,7 +42,7 @@ "allow_partial_search_results": { "type": "boolean", "description": "Control whether the query should keep running in case of shard failures, and return partial results", - "default": false + "default": true }, "allow_partial_sequence_results": { "type": "boolean", diff --git a/specification/_json_spec/indices.cancel_migrate_reindex.json b/specification/_json_spec/indices.cancel_migrate_reindex.json index 9c333c2736..26d8881201 100644 --- a/specification/_json_spec/indices.cancel_migrate_reindex.json +++ b/specification/_json_spec/indices.cancel_migrate_reindex.json @@ -1,7 +1,7 @@ { "indices.cancel_migrate_reindex": { "documentation": { - "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex.html", + "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex-cancel-api.html", "description": "This API returns the status of a migration reindex attempt for a data stream or index" }, "stability": "experimental", diff --git a/specification/_json_spec/indices.create_from.json b/specification/_json_spec/indices.create_from.json index 76fb06d072..342f560d1c 100644 --- a/specification/_json_spec/indices.create_from.json +++ b/specification/_json_spec/indices.create_from.json @@ -1,7 +1,7 @@ { "indices.create_from": { "documentation": { - "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex.html", + "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/indices-create-index-from-source.html", "description": "This API creates a destination from a source index. It copies the mappings and settings from the source index while allowing request settings and mappings to override the source values." }, "stability": "experimental", diff --git a/specification/_json_spec/indices.get_migrate_reindex_status.json b/specification/_json_spec/indices.get_migrate_reindex_status.json index 309d44f137..7e027dec38 100644 --- a/specification/_json_spec/indices.get_migrate_reindex_status.json +++ b/specification/_json_spec/indices.get_migrate_reindex_status.json @@ -1,7 +1,7 @@ { "indices.get_migrate_reindex_status": { "documentation": { - "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex.html", + "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex-status-api.html", "description": "This API returns the status of a migration reindex attempt for a data stream or index" }, "stability": "experimental", diff --git a/specification/_json_spec/indices.migrate_reindex.json b/specification/_json_spec/indices.migrate_reindex.json index 1f69103497..d1689f7eb9 100644 --- a/specification/_json_spec/indices.migrate_reindex.json +++ b/specification/_json_spec/indices.migrate_reindex.json @@ -1,7 +1,7 @@ { "indices.migrate_reindex": { "documentation": { - "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex.html", + "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/data-stream-reindex-api.html", "description": "This API reindexes all legacy backing indices for a data stream. It does this in a persistent task. The persistent task id is returned immediately, and the reindexing work is completed in that task" }, "stability": "experimental", diff --git a/specification/_json_spec/indices.resolve_cluster.json b/specification/_json_spec/indices.resolve_cluster.json index 0469c903fa..fddbc513a0 100644 --- a/specification/_json_spec/indices.resolve_cluster.json +++ b/specification/_json_spec/indices.resolve_cluster.json @@ -2,7 +2,7 @@ "indices.resolve_cluster": { "documentation": { "url": "https://www.elastic.co/guide/en/elasticsearch/reference/master/indices-resolve-cluster-api.html", - "description": "Resolves the specified index expressions to return information about each cluster. If no index expression is provided, this endpoint will return information about all the remote clusters that are configured on the local cluster." + "description": "Resolves the specified index expressions to return information about each cluster, including the local cluster, if included." }, "stability": "stable", "visibility": "public", @@ -11,10 +11,6 @@ }, "url": { "paths": [ - { - "path": "/_resolve/cluster", - "methods": ["GET"] - }, { "path": "/_resolve/cluster/{name}", "methods": ["GET"], @@ -30,25 +26,21 @@ "params": { "ignore_unavailable": { "type": "boolean", - "description": "Whether specified concrete indices should be ignored when unavailable (missing or closed). Only allowed when providing an index expression." + "description": "Whether specified concrete indices should be ignored when unavailable (missing or closed)" }, "ignore_throttled": { "type": "boolean", - "description": "Whether specified concrete, expanded or aliased indices should be ignored when throttled. Only allowed when providing an index expression." + "description": "Whether specified concrete, expanded or aliased indices should be ignored when throttled" }, "allow_no_indices": { "type": "boolean", - "description": "Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes `_all` string or when no indices have been specified). Only allowed when providing an index expression." + "description": "Whether to ignore if a wildcard indices expression resolves into no concrete indices. (This includes `_all` string or when no indices have been specified)" }, "expand_wildcards": { "type": "enum", "options": ["open", "closed", "hidden", "none", "all"], "default": "open", - "description": "Whether wildcard expressions should get expanded to open or closed indices (default: open). Only allowed when providing an index expression." - }, - "timeout": { - "type": "time", - "description": "The maximum time to wait for remote clusters to respond" + "description": "Whether wildcard expressions should get expanded to open or closed indices (default: open)" } } } diff --git a/specification/_json_spec/inference.update.json b/specification/_json_spec/inference.update.json index 6c458ce080..133354e3ec 100644 --- a/specification/_json_spec/inference.update.json +++ b/specification/_json_spec/inference.update.json @@ -14,7 +14,7 @@ "paths": [ { "path": "/_inference/{inference_id}/_update", - "methods": ["POST"], + "methods": ["PUT"], "parts": { "inference_id": { "type": "string", @@ -24,7 +24,7 @@ }, { "path": "/_inference/{task_type}/{inference_id}/_update", - "methods": ["POST"], + "methods": ["PUT"], "parts": { "task_type": { "type": "string",