mirror of
https://github.com/elastic/elasticsearch.git
synced 2025-04-24 15:17:30 -04:00
This adds support for GET and DELETE and the ids query and Elasticsearch's standard document versioning to TSDB. So you can do things like: ``` POST /tsdb_idx/_doc?filter_path=_id { "@timestamp": "2021-12-29T19:25:05Z", "uid": "adsfadf", "v": 1.2 } ``` That'll return `{"_id" : "BsYQJjqS3TnsUlF3aDKnB34BAAA"}` which you can turn around and fetch with ``` GET /tsdb_idx/_doc/BsYQJjqS3TnsUlF3aDKnB34BAAA ``` just like any other document in any other index. You can delete it too! Or fetch it. The ID comes from the dimensions and the `@timestamp`. So you can overwrite the document: ``` POST /tsdb_idx/_bulk {"index": {}} {"@timestamp": "2021-12-29T19:25:05Z", "uid": "adsfadf", "v": 1.2} ``` Or you can write only if it doesn't already exist: ``` POST /tsdb_idx/_bulk {"create": {}} {"@timestamp": "2021-12-29T19:25:05Z", "uid": "adsfadf", "v": 1.2} ``` This works by generating an id from the dimensions and the `@timestamp` when parsing the document. The id looks like: * 4 bytes of hash from the routing calculated from routing_path fields * 8 bytes of hash from the dimensions * 8 bytes of timestamp All that's base 64 encoded so that `Uid` can chew on it fairly efficiently. When it comes time to fetch or delete documents we base 64 decode the id and grab the routing from the first four bytes. We use that hash to pick the shard. Then we use the entire ID to perform the fetch or delete. We don't implement update actions because we haven't written the infrastructure to make sure the dimensions don't change. It's possible to do, but feels like more than we need now. There *ton* of compromises with this. The long term sad thing is that it locks us into *indexing* the id of the sample. It'll index fairly efficiently because the each time series will have the same first eight bytes. It's also possible we'd share many of the first few bytes in the timestamp as well. In our tsdb rally track this costs 8.75 bytes per document. It's substantial, but not overwhelming. In the short term there are lots of problems that I'd like to save for a follow up change: 1. ~~We still generate the automatic `_id` for the document but we don't use it. We should stop generating it.~~ Included in this PR based on review comments. 2. We generated the time series `_id` on each shard and when replaying the translog. It'd be the good kind of paranoid to generate it once on the primary and then keep it forever. 3. We have to encode the `_id` as a string to pass it around Elasticsearch internally. And Elasticsearch assumes that when an id is loaded we always store as bytes encoded the `Uid` - which *does* have nice encoding for base 64 bytes. But this whole thing requires us to make the bytes, base 64 encode them, and then hand them back to `Uid` to base 64 decode them into bytes. It's a bit hacky. And, it's a small thing, but if the first byte of the routing hash encodes to 254 or 255 we `Uid` spends an extra byte to encode it. One that'll always be a common prefix for tsdb indices, but still, it hurts my heart. It's just hard to fix. 4. We store the `_id` in Lucene stored fields for tsdb indices. Now that we're building it from the dimensions and the `@timestamp` we really don't *need* to store it. We could recalculate it when fetching documents. In the tsdb rall ytrick this'd save us 6 bytes per document at the cost of marginally slower fetches. Which is *fine*. 5. There are several error messages that try to use `_id` right now during parsing but the `_id` isn't available until after the parsing is complete. And, if parsing fails, it may not be possible to know the id at all. All of these error messages will have to change, at least in tsdb mode. 6. ~~If you specify an `_id` on the request right now we just overwrite it. We should send you an error.~~ Included in this PR after review comments. 7. We have to entirely disable the append-only optimization that allows Elasticsearch to skip looking up the ids in lucene. This *halves* indexing speed. It's substantial. We have to claw that optimization back *somehow*. Something like sliding bloom filters or relying on the increasing timestamps. 8. We parse the source from json when building the routing hash when parsing fields. We should just build it from to parsed field values. It looks like that'd improve indexing speed by about 20%. 9. Right now we write the `@timestamp` little endian. This is likely bad the prefix encoded inverted index. It'll prefer big endian. Might shrink it. 10. Improve error message on version conflict to include tsid and timestamp. 11. Improve error message when modifying dimensions or timestamp in update_by_query 12. Make it possible to modify dimension or timestamp in reindex. 13. Test TSDB's `_id` in `RecoverySourceHandlerTests.java` and `EngineTests.java`. I've had to make some changes as part of this that don't feel super expected. The biggest one is changing `Engine.Result` to include the `id`. When the `id` comes from the dimensions it is calculated by the document parsing infrastructure which is happens in `IndexShard#pepareIndex`. Which returns an `Engine.IndexResult`. To make everything clean I made it so `id` is available on all `Engine.Result`s and I made all of the "outer results classes" read from `Engine.Results#id`. I'm not excited by it. But it works and it's what we're going with. I've opted to create two subclasses of `IdFieldMapper`, one for standard indices and one for tsdb indices. This feels like the right way to introduce the distinction, especially if we don't want tsdb to cary around it's old fielddata support. Honestly if we *need* to aggregate on `_id` in tsdb mode we have doc values for the `tsdb` and the `@timestamp` - we could build doc values for `_id` on the fly. But I'm not expecting folks will need to do this. Also! I'd like to stop storing tsdb'd `_id` field (see number 4 above) and the new subclass feels like a good place to put that too. |
||
---|---|---|
.. | ||
src/test/java/org/elasticsearch/upgrades | ||
build.gradle |