github.com/yankunsam/loki/v2@v2.6.3-0.20220817130409-389df5235c27/docs/sources/release-notes/v2-5.md (about) 1 --- 2 title: V2.5 3 weight: 77 4 --- 5 6 # Version 2.5 release notes 7 8 It has been nearly 6 months since Loki 2.4 was released, and we’ve been busy making Loki better than ever. Here's a summary of new enhancements and important fixes. 9 10 ## Features and enhancements 11 12 - Thanks to @bboreham for taking a deep dive on the Go `regexp` library, and for creating a fork which greatly **[improves the performance of common regex uses cases in Loki.](https://github.com/grafana/loki/pull/5315)** 13 - **[Binary operations are now significantly faster](https://github.com/grafana/loki/pull/5317)**, taking full advantage of Loki's parallelism. 14 - **[A new schema is available](https://github.com/grafana/loki/pull/5054)**, which uses more path prefixes to avoid rate limits on S3. 15 - That same schema change **[was also added to the filesystem store](https://github.com/grafana/loki/pull/5291)**, which avoids using one directory to store every chunk. 16 - A new capability for **[hedging requests to storage](https://github.com/grafana/loki/pull/4826)** improves performance on highly parallelized queries. Refer to the [hedging configuration](../../configuration/#hedging) block for more information. 17 - Promtail has several new ways to ingest logs: 18 - The **[ability to do service discovery and tailing directly from the Docker daemon](https://github.com/grafana/loki/pull/4911)**. 19 - **[Fetching logs directly from Cloudflare](https://github.com/grafana/loki/pull/4813)**. 20 - **[Receiving logs directly in Graylog Extended Log Format (GELF)](https://github.com/grafana/loki/pull/4744)**. 21 - Promtail permits **[client-side global rate limiting](https://github.com/grafana/loki/pull/5031)** and **[pipeline-configurable rate limiting](https://github.com/grafana/loki/pull/5051)**. 22 23 For a full list of all changes please look at the [CHANGELOG](https://github.com/grafana/loki/blob/main/CHANGELOG.md). 24 25 ## Upgrade Considerations 26 27 As always, please read the [upgrade guide](../../upgrading/#250) before upgrading Loki. 28 29 ### Changes to the config `split_queries_by_interval` 30 The most likely impact many people will see is Loki failing to start because of a change in the YAML configuration for `split_queries_by_interval`. It was previously possible to define this value in two places. 31 32 ```yaml 33 query_range: 34 split_queries_by_interval: 10m 35 ``` 36 37 and/or 38 39 ```yaml 40 limits_config: 41 split_queries_by_interval: 10m 42 ``` 43 44 In 2.5.0 it can **only be defined in the `limits_config` section**. 45 46 **Loki will fail to start if you do not remove the `split_queries_by_interval` configuration parameter from the `query_range` section.** 47 48 Additionally, it has a new default value of `30m` rather than `0`. 49 50 The CLI flag is not changed and remains `querier.split-queries-by-interval`. 51 52 ### More parallelism by default 53 54 We continue to move defaults in Loki to leverage parallelism in all configurations, including the single binary. 55 56 In v2.5, all queries will be split and sharded by default. This will likely result in increased memory and CPU usage for Loki processes during queries, if you didn’t previously have these values enabled. 57 58 ## Usage reporting 59 60 Loki 2.5 includes code we added to report anonymous usage statistics back to Grafana Labs, [an issue was created to outline the intent](https://github.com/grafana/loki/issues/5062), and what went into the final implementation [can be seen here in the source](https://github.com/grafana/loki/blob/v2.5.0/pkg/usagestats/stats.go#L75). 61 62 Usage reporting helps provide anonymous information on how people use Loki and what the Loki team should focus on for features and documentation. No private information is collected, and all reports are completely anonymous. 63 64 If possible, we ask you to leave the usage reporting feature enabled and help us understand more about Loki! We are also working to figure out how we can share this info with the community so everyone can watch Loki grow. 65 66 If you would rather not participate in usage stats reporting, [the feature can be disabled in config](https://grafana.com/docs/loki/latest/configuration/#analytics) 67 68 ``` 69 analytics: 70 reporting_enabled: false 71 ``` 72 73 ## Bug fixes 74 75 ### 2.5.0 bug fixes 76 77 V2.5.0 fixes numerous bugs. The [CHANGELOG](https://github.com/grafana/loki/blob/main/CHANGELOG.md) has the complete list. 78 79 A summary of some of the more important fixes: 80 81 * Query correctness: 82 * [PR 5474](https://github.com/grafana/loki/pull/5474) **cyriltovena**: Disable sharding of count/avg when labels are mutated 83 * [PR 5444](https://github.com/grafana/loki/pull/5444) **cyriltovena**: Do not insert missing point when sharding 84 * [PR 5423](https://github.com/grafana/loki/pull/5423) **cyriltovena**: Correctly sets hash value for headblock iterator 85 * [PR 5289](https://github.com/grafana/loki/pull/5289) **cyriltovena**: Fixes log deduplication when mutating Labels using LogQL 86 * [PR 5006](https://github.com/grafana/loki/pull/5006) **sandeepsukhani**: Fix splitting of queries when step is larger than split interval 87 * [PR 5413](https://github.com/grafana/loki/pull/5413) **owen-d**: A deadlock in the Azure blob client 88 * [PR 5334](https://github.com/grafana/loki/pull/5334) **kavirajk**: A fix in live tailing that could lead to exploding memory usage 89 * [PR 5144](https://github.com/grafana/loki/pull/5144) **owen-d**: A fix for the ruler when trying to use Basic Auth for remote write 90 * Query cancellation: 91 * [PR 5113](https://github.com/grafana/loki/pull/5113) **kavirajk**: Fix cancel issue between query-frontend and query scheduler 92 * [PR 5080](https://github.com/grafana/loki/pull/5080) **kavirajk**: Handle `context` cancellation in some of the `querier` downstream requests 93 * [PR 5075](https://github.com/grafana/loki/pull/5075) **cyriltovena**: Fixes a possible cancellation issue in the frontend 94 * [PR 4741](https://github.com/grafana/loki/pull/4741) **sandeepsukhani**: An issue where retention didn't cleanup the index completely