Documentation

InfluxDB shards and shard groups

InfluxDB organizes time series data into shards when storing data to disk. Shards are grouped into shard groups. Learn the relationships between buckets, shards, and shard groups.

Shards

A shard contains encoded and compressed time series data for a given time range defined by the shard group duration. All points in a series within the specified shard group duration are stored in the same shard. A single shard contains multiple series, one or more TSM files on disk, and belongs to a shard group.

Shard groups

A shard group belongs to an InfluxDB bucket and contains time series data for a specific time range defined by the shard group duration.

In InfluxDB OSS, a shard group typically contains only a single shard. In an InfluxDB Enterprise 1.x cluster, shard groups contain multiple shards distributed across multiple data nodes.

Shard group duration

The shard group duration specifies the time range for each shard group and determines how often to create a new shard group. By default, InfluxDB sets the shard group duration according to the retention period of the bucket:

Bucket retention period Default shard group duration
less than 2 days 1h
between 2 days and 6 months 1d
greater than 6 months 7d
Shard group duration configuration options

To configure a custom bucket shard group duration, use the --shard-group-duration flag with the influx bucket create and influx bucket update commands.

Shard group durations must be shorter than the bucket’s retention period.

To view your bucket’s shard group duration, use the influx bucket list command.

Shard group diagram

The following diagram represents a bucket with a 4d retention period and a 1d shard group duration:


0d

1d

2d

3d

4d

Shard group

Shard

Shard group

Shard

Shard group

Shard

Shard group

Shard


Shard life cycle

Shard precreation

The InfluxDB shard precreation service pre-creates shards with future start and end times for each shard group based on the shard group duration.

The precreator service does not pre-create shards for past time ranges. When backfilling historical data, InfluxDB creates shards for past time ranges as needed, resulting in temporarily lower write throughput.

Shard writes

InfluxDB writes time series data to un-compacted or “hot” shards. When a shard is no longer actively written to, InfluxDB compacts shard data, resulting in a “cold” shard.

Typically, InfluxDB writes data to the most recent shard group, but when backfilling historical data, InfluxDB writes to older shards that must first be un-compacted. When the backfill is complete, InfluxDB re-compacts the older shards.

Shard compaction

InfluxDB compacts shards at regular intervals to compress time series data and optimize disk usage. When compactions are enabled, InfluxDB checks to see whether shard compactions are needed every second. If there haven’t been writes during the compact-full-write-cold-duration period (by default, 4h), InfluxDB compacts all TSM files. Otherwise, InfluxDB groups TSM files into compaction levels (determined by the number of times the file have been compacted), and attempts to combine files and compress them more efficiently.

InfluxDB uses the following four compaction levels:

  • Level 0 (L0): The log file (LogFile) is considered level 0 (L0). Once this file exceeds a 5MB threshold, InfluxDB creates a new active log file, and the previous one begins compacting into an IndexFile. This first index file is at level 1 (L1).
  • Level 1 (L1): InfluxDB flushes all newly written data held in an in-memory cache to disk into an IndexFile.
  • Level 2 (L2): InfluxDB compacts up to eight L1-compacted files into one or more L2 files by combining multiple blocks containing the same series into fewer blocks in one or more new files.
  • Level 3 (L3): InfluxDB iterates over L2-compacted file blocks (over a certain size) and combines multiple blocks containing the same series into one block in a new file.
  • Level 4 (L4): Full compaction InfluxDB iterates over L3-compacted file blocks and combines multiple blocks containing the same series into one block in a new file.

InfluxDB schedules compactions preferentially, using the following guidelines:

  • The lower the level (fewer times the file has been compacted), the more weight is given to compacting the file.
  • The more compactible files in a level, the higher the priority given to compacting that level. If the number of files in each level is equal, lower levels are compacted first.
  • If a higher level has more candidates for compaction, it may be compacted before a lower level. InfluxDB multiplies the number of collection groups (collections of files to compact into a single next-generation file) by a specified weight (0.4, 0.3, 0.2, and 0.1) per level, to determine the compaction priority.

The following configuration settings are especially beneficial for systems with irregular loads, because they limit compactions during periods of high usage, and let compactions catch up during periods of lower load:

In systems with stable loads, if compactions interfere with other operations, typically, the system is undersized for its load, and configuration changes won’t help much.

Shard deletion

The InfluxDB retention enforcement service routinely checks for shard groups older than their bucket’s retention period. Once the start time of a shard group is beyond the bucket’s retention period, InfluxDB deletes the shard group and associated shards and TSM files.

In buckets with an infinite retention period, shards remain on disk indefinitely.

InfluxDB only deletes cold shards

InfluxDB only deletes cold shards. If backfilling data beyond a bucket’s retention period, the backfilled data will remain on disk until the following occurs:

  1. The shard returns to a cold state.
  2. The retention enforcement service deletes the shard group.

Was this page helpful?

Thank you for your feedback!


The future of Flux

Flux is going into maintenance mode. You can continue using it as you currently are without any changes to your code.

Flux is going into maintenance mode and will not be supported in InfluxDB 3.0. This was a decision based on the broad demand for SQL and the continued growth and adoption of InfluxQL. We are continuing to support Flux for users in 1.x and 2.x so you can continue using it with no changes to your code. If you are interested in transitioning to InfluxDB 3.0 and want to future-proof your code, we suggest using InfluxQL.

For information about the future of Flux, see the following: