BTRFS(5) BTRFS BTRFS(5)
NAME
btrfs - topics about the BTRFS filesystem (mount options, supported
file attributes and other)
DESCRIPTION
This document describes topics related to BTRFS that are not specific
to the tools. Currently covers:
1. mount options
2. filesystem features
3. checksum algorithms
4. compression
5. sysfs interface
6. filesystem exclusive operations
7. filesystem limits
8. bootloader support
9. file attributes
10. zoned mode
11. control device
12. filesystems with multiple block group profiles
13. seeding device
14. RAID56 status and recommended practices
15. storage model, hardware considerations
MOUNT OPTIONS
BTRFS SPECIFIC MOUNT OPTIONS
This section describes mount options specific to BTRFS. For the
generic mount options please refer to mount(8) manual page. The options
are sorted alphabetically (discarding the no prefix).
NOTE:
Most mount options apply to the whole filesystem and only options in
the first mounted subvolume will take effect. This is due to lack of
implementation and may change in the future. This means that (for
example) you can't set per-subvolume nodatacow, nodatasum, or com-
press using mount options. This should eventually be fixed, but it
has proved to be difficult to implement correctly within the Linux
VFS framework.
Mount options are processed in order, only the last occurrence of an
option takes effect and may disable other options due to constraints
(see e.g. nodatacow and compress). The output of mount command shows
which options have been applied.
acl, noacl
(default: on)
Enable/disable support for POSIX Access Control Lists (ACLs).
See the acl(5) manual page for more information about ACLs.
The support for ACL is build-time configurable
(BTRFS_FS_POSIX_ACL) and mount fails if acl is requested but the
feature is not compiled in.
autodefrag, noautodefrag
(since: 3.0, default: off)
Enable automatic file defragmentation. When enabled, small ran-
dom writes into files (in a range of tens of kilobytes, cur-
rently it's 64KiB) are detected and queued up for the defragmen-
tation process. May not be well suited for large database work-
loads.
The read latency may increase due to reading the adjacent blocks
that make up the range for defragmentation, successive write
will merge the blocks in the new location.
WARNING:
Defragmenting with Linux kernel versions < 3.9 or ≥ 3.14-rc2
as well as with Linux stable kernel versions ≥ 3.10.31, ≥
3.12.12 or ≥ 3.13.4 will break up the reflinks of COW data
(for example files copied with cp --reflink, snapshots or
de-duplicated data). This may cause considerable increase of
space usage depending on the broken up reflinks.
barrier, nobarrier
(default: on)
Ensure that all IO write operations make it through the device
cache and are stored permanently when the filesystem is at its
consistency checkpoint. This typically means that a flush com-
mand is sent to the device that will synchronize all pending
data and ordinary metadata blocks, then writes the superblock
and issues another flush.
The write flushes incur a slight hit and also prevent the IO
block scheduler to reorder requests in a more effective way.
Disabling barriers gets rid of that penalty but will most cer-
tainly lead to a corrupted filesystem in case of a crash or
power loss. The ordinary metadata blocks could be yet unwritten
at the time the new superblock is stored permanently, expecting
that the block pointers to metadata were stored permanently be-
fore.
On a device with a volatile battery-backed write-back cache, the
nobarrier option will not lead to filesystem corruption as the
pending blocks are supposed to make it to the permanent storage.
check_int, check_int_data, check_int_print_mask=<value>
(since: 3.0, default: off)
These debugging options control the behavior of the integrity
checking module (the BTRFS_FS_CHECK_INTEGRITY config option re-
quired). The main goal is to verify that all blocks from a given
transaction period are properly linked.
check_int enables the integrity checker module, which examines
all block write requests to ensure on-disk consistency, at a
large memory and CPU cost.
check_int_data includes extent data in the integrity checks, and
implies the check_int option.
check_int_print_mask takes a bitmask of BTRFSIC_PRINT_MASK_*
values as defined in fs/btrfs/check-integrity.c, to control the
integrity checker module behavior.
See comments at the top of fs/btrfs/check-integrity.c for more
information.
clear_cache
Force clearing and rebuilding of the free space cache if some-
thing has gone wrong.
For free space cache v1, this only clears (and, unless no-
space_cache is used, rebuilds) the free space cache for block
groups that are modified while the filesystem is mounted with
that option. To actually clear an entire free space cache v1,
see btrfs check --clear-space-cache v1.
For free space cache v2, this clears the entire free space
cache. To do so without requiring to mounting the filesystem,
see btrfs check --clear-space-cache v2.
See also: space_cache.
commit=<seconds>
(since: 3.12, default: 30)
Set the interval of periodic transaction commit when data are
synchronized to permanent storage. Higher interval values lead
to larger amount of unwritten data, which has obvious conse-
quences when the system crashes. The upper bound is not forced,
but a warning is printed if it's more than 300 seconds (5 min-
utes). Use with care.
compress, compress=<type[:level]>, compress-force, com-
press-force=<type[:level]>
(default: off, level support since: 5.1)
Control BTRFS file data compression. Type may be specified as
zlib, lzo, zstd or no (for no compression, used for remounting).
If no type is specified, zlib is used. If compress-force is
specified, then compression will always be attempted, but the
data may end up uncompressed if the compression would make them
larger.
Both zlib and zstd (since version 5.1) expose the compression
level as a tunable knob with higher levels trading speed and
memory (zstd) for higher compression ratios. This can be set by
appending a colon and the desired level. ZLIB accepts the range
[1, 9] and ZSTD accepts [1, 15]. If no level is set, both cur-
rently use a default level of 3. The value 0 is an alias for the
default level.
Otherwise some simple heuristics are applied to detect an incom-
pressible file. If the first blocks written to a file are not
compressible, the whole file is permanently marked to skip com-
pression. As this is too simple, the compress-force is a
workaround that will compress most of the files at the cost of
some wasted CPU cycles on failed attempts. Since kernel 4.15, a
set of heuristic algorithms have been improved by using fre-
quency sampling, repeated pattern detection and Shannon entropy
calculation to avoid that.
NOTE:
If compression is enabled, nodatacow and nodatasum are dis-
abled.
datacow, nodatacow
(default: on)
Enable data copy-on-write for newly created files. Nodatacow
implies nodatasum, and disables compression. All files created
under nodatacow are also set the NOCOW file attribute (see
chattr(1)).
NOTE:
If nodatacow or nodatasum are enabled, compression is dis-
abled.
Updates in-place improve performance for workloads that do fre-
quent overwrites, at the cost of potential partial writes, in
case the write is interrupted (system crash, device failure).
datasum, nodatasum
(default: on)
Enable data checksumming for newly created files. Datasum im-
plies datacow, i.e. the normal mode of operation. All files cre-
ated under nodatasum inherit the "no checksums" property, how-
ever there's no corresponding file attribute (see chattr(1)).
NOTE:
If nodatacow or nodatasum are enabled, compression is dis-
abled.
There is a slight performance gain when checksums are turned
off, the corresponding metadata blocks holding the checksums do
not need to updated. The cost of checksumming of the blocks in
memory is much lower than the IO, modern CPUs feature hardware
support of the checksumming algorithm.
degraded
(default: off)
Allow mounts with fewer devices than the RAID profile con-
straints require. A read-write mount (or remount) may fail when
there are too many devices missing, for example if a stripe mem-
ber is completely missing from RAID0.
Since 4.14, the constraint checks have been improved and are
verified on the chunk level, not at the device level. This al-
lows degraded mounts of filesystems with mixed RAID profiles for
data and metadata, even if the device number constraints would
not be satisfied for some of the profiles.
Example: metadata -- raid1, data -- single, devices -- /dev/sda,
/dev/sdb
Suppose the data are completely stored on sda, then missing sdb
will not prevent the mount, even if 1 missing device would nor-
mally prevent (any) single profile to mount. In case some of the
data chunks are stored on sdb, then the constraint of sin-
gle/data is not satisfied and the filesystem cannot be mounted.
device=<devicepath>
Specify a path to a device that will be scanned for BTRFS
filesystem during mount. This is usually done automatically by a
device manager (like udev) or using the btrfs device scan com-
mand (e.g. run from the initial ramdisk). In cases where this is
not possible the device mount option can help.
NOTE:
Booting e.g. a RAID1 system may fail even if all filesystem's
device paths are provided as the actual device nodes may not
be discovered by the system at that point.
discard, discard=sync, discard=async, nodiscard
(default: async when devices support it since 6.2, async support
since: 5.6)
Enable discarding of freed file blocks. This is useful for SSD
devices, thinly provisioned LUNs, or virtual machine images;
however, every storage layer must support discard for it to
work.
In the synchronous mode (sync or without option value), lack of
asynchronous queued TRIM on the backing device TRIM can severely
degrade performance, because a synchronous TRIM operation will
be attempted instead. Queued TRIM requires newer than SATA revi-
sion 3.1 chipsets and devices.
The asynchronous mode (async) gathers extents in larger chunks
before sending them to the devices for TRIM. The overhead and
performance impact should be negligible compared to the previous
mode and it's supposed to be the preferred mode if needed.
If it is not necessary to immediately discard freed blocks, then
the fstrim tool can be used to discard all free blocks in a
batch. Scheduling a TRIM during a period of low system activity
will prevent latent interference with the performance of other
operations. Also, a device may ignore the TRIM command if the
range is too small, so running a batch discard has a greater
probability of actually discarding the blocks.
enospc_debug, noenospc_debug
(default: off)
Enable verbose output for some ENOSPC conditions. It's safe to
use but can be noisy if the system reaches near-full state.
fatal_errors=<action>
(since: 3.4, default: bug)
Action to take when encountering a fatal error.
bug BUG() on a fatal error, the system will stay in the
crashed state and may be still partially usable, but re-
boot is required for full operation
panic panic() on a fatal error, depending on other system con-
figuration, this may be followed by a reboot. Please re-
fer to the documentation of kernel boot parameters, e.g.
panic, oops or crashkernel.
flushoncommit, noflushoncommit
(default: off)
This option forces any data dirtied by a write in a prior trans-
action to commit as part of the current commit, effectively a
full filesystem sync.
This makes the committed state a fully consistent view of the
file system from the application's perspective (i.e. it includes
all completed file system operations). This was previously the
behavior only when a snapshot was created.
When off, the filesystem is consistent but buffered writes may
last more than one transaction commit.
fragment=<type>
(depends on compile-time option CONFIG_BTRFS_DEBUG, since: 4.4,
default: off)
A debugging helper to intentionally fragment given type of block
groups. The type can be data, metadata or all. This mount option
should not be used outside of debugging environments and is not
recognized if the kernel config option CONFIG_BTRFS_DEBUG is not
enabled.
nologreplay
(default: off, even read-only)
The tree-log contains pending updates to the filesystem until
the full commit. The log is replayed on next mount, this can be
disabled by this option. See also treelog. Note that nologre-
play is the same as norecovery.
WARNING:
Currently, the tree log is replayed even with a read-only
mount! To disable that behaviour, mount also with nologre-
play.
max_inline=<bytes>
(default: min(2048, page size) )
Specify the maximum amount of space, that can be inlined in a
metadata b-tree leaf. The value is specified in bytes, option-
ally with a K suffix (case insensitive). In practice, this
value is limited by the filesystem block size (named sectorsize
at mkfs time), and memory page size of the system. In case of
sectorsize limit, there's some space unavailable due to b-tree
leaf headers. For example, a 4KiB sectorsize, maximum size of
inline data is about 3900 bytes.
Inlining can be completely turned off by specifying 0. This will
increase data block slack if file sizes are much smaller than
block size but will reduce metadata consumption in return.
NOTE:
The default value has changed to 2048 in kernel 4.6.
metadata_ratio=<value>
(default: 0, internal logic)
Specifies that 1 metadata chunk should be allocated after every
value data chunks. Default behaviour depends on internal logic,
some percent of unused metadata space is attempted to be main-
tained but is not always possible if there's not enough space
left for chunk allocation. The option could be useful to over-
ride the internal logic in favor of the metadata allocation if
the expected workload is supposed to be metadata intense (snap-
shots, reflinks, xattrs, inlined files).
norecovery
(since: 4.5, default: off)
Do not attempt any data recovery at mount time. This will dis-
able logreplay and avoids other write operations. Note that this
option is the same as nologreplay.
NOTE:
The opposite option recovery used to have different meaning
but was changed for consistency with other filesystems, where
norecovery is used for skipping log replay. BTRFS does the
same and in general will try to avoid any write operations.
rescan_uuid_tree
(since: 3.12, default: off)
Force check and rebuild procedure of the UUID tree. This should
not normally be needed.
rescue (since: 5.9)
Modes allowing mount with damaged filesystem structures.
• usebackuproot (since: 5.9, replaces standalone option useback-
uproot)
• nologreplay (since: 5.9, replaces standalone option nologre-
play)
• ignorebadroots, ibadroots (since: 5.11)
• ignoredatacsums, idatacsums (since: 5.11)
• all (since: 5.9)
skip_balance
(since: 3.3, default: off)
Skip automatic resume of an interrupted balance operation. The
operation can later be resumed with btrfs balance resume, or the
paused state can be removed with btrfs balance cancel. The de-
fault behaviour is to resume an interrupted balance immediately
after a volume is mounted.
space_cache, space_cache=<version>, nospace_cache
(nospace_cache since: 3.2, space_cache=v1 and space_cache=v2
since 4.5, default: space_cache=v2)
Options to control the free space cache. The free space cache
greatly improves performance when reading block group free space
into memory. However, managing the space cache consumes some re-
sources, including a small amount of disk space.
There are two implementations of the free space cache. The orig-
inal one, referred to as v1, used to be a safe default but has
been superseded by v2. The v1 space cache can be disabled at
mount time with nospace_cache without clearing.
On very large filesystems (many terabytes) and certain work-
loads, the performance of the v1 space cache may degrade drasti-
cally. The v2 implementation, which adds a new b-tree called the
free space tree, addresses this issue. Once enabled, the v2
space cache will always be used and cannot be disabled unless it
is cleared. Use clear_cache,space_cache=v1 or clear_cache,no-
space_cache to do so. If v2 is enabled, and v1 space cache will
be cleared (at the first mount) and kernels without v2 support
will only be able to mount the filesystem in read-only mode. On
an unmounted filesystem the caches (both versions) can be
cleared by "btrfs check --clear-space-cache".
The btrfs-check(8) and :doc:`mkfs.btrfs commands have full v2
free space cache support since v4.19.
If a version is not explicitly specified, the default implemen-
tation will be chosen, which is v2.
ssd, ssd_spread, nossd, nossd_spread
(default: SSD autodetected)
Options to control SSD allocation schemes. By default, BTRFS
will enable or disable SSD optimizations depending on status of
a device with respect to rotational or non-rotational type. This
is determined by the contents of /sys/block/DEV/queue/rota-
tional). If it is 0, the ssd option is turned on. The option
nossd will disable the autodetection.
The optimizations make use of the absence of the seek penalty
that's inherent for the rotational devices. The blocks can be
typically written faster and are not offloaded to separate
threads.
NOTE:
Since 4.14, the block layout optimizations have been dropped.
This used to help with first generations of SSD devices.
Their FTL (flash translation layer) was not effective and the
optimization was supposed to improve the wear by better
aligning blocks. This is no longer true with modern SSD de-
vices and the optimization had no real benefit. Furthermore
it caused increased fragmentation. The layout tuning has been
kept intact for the option ssd_spread.
The ssd_spread mount option attempts to allocate into bigger and
aligned chunks of unused space, and may perform better on
low-end SSDs. ssd_spread implies ssd, enabling all other SSD
heuristics as well. The option nossd will disable all SSD op-
tions while nossd_spread only disables ssd_spread.
subvol=<path>
Mount subvolume from path rather than the toplevel subvolume.
The path is always treated as relative to the toplevel subvol-
ume. This mount option overrides the default subvolume set for
the given filesystem.
subvolid=<subvolid>
Mount subvolume specified by a subvolid number rather than the
toplevel subvolume. You can use btrfs subvolume list of btrfs
subvolume show to see subvolume ID numbers. This mount option
overrides the default subvolume set for the given filesystem.
NOTE:
If both subvolid and subvol are specified, they must point at
the same subvolume, otherwise the mount will fail.
thread_pool=<number>
(default: min(NRCPUS + 2, 8) )
The number of worker threads to start. NRCPUS is number of
on-line CPUs detected at the time of mount. Small number leads
to less parallelism in processing data and metadata, higher num-
bers could lead to a performance hit due to increased locking
contention, process scheduling, cache-line bouncing or costly
data transfers between local CPU memories.
treelog, notreelog
(default: on)
Enable the tree logging used for fsync and O_SYNC writes. The
tree log stores changes without the need of a full filesystem
sync. The log operations are flushed at sync and transaction
commit. If the system crashes between two such syncs, the pend-
ing tree log operations are replayed during mount.
WARNING:
Currently, the tree log is replayed even with a read-only
mount! To disable that behaviour, also mount with nologre-
play.
The tree log could contain new files/directories, these would
not exist on a mounted filesystem if the log is not replayed.
usebackuproot
(since: 4.6, default: off)
Enable autorecovery attempts if a bad tree root is found at
mount time. Currently this scans a backup list of several pre-
vious tree roots and tries to use the first readable. This can
be used with read-only mounts as well.
NOTE:
This option has replaced recovery.
user_subvol_rm_allowed
(default: off)
Allow subvolumes to be deleted by their respective owner. Other-
wise, only the root user can do that.
NOTE:
Historically, any user could create a snapshot even if he was
not owner of the source subvolume, the subvolume deletion has
been restricted for that reason. The subvolume creation has
been restricted but this mount option is still required. This
is a usability issue. Since 4.18, the rmdir(2) syscall can
delete an empty subvolume just like an ordinary directory.
Whether this is possible can be detected at runtime, see
rmdir_subvol feature in FILESYSTEM FEATURES.
DEPRECATED MOUNT OPTIONS
List of mount options that have been removed, kept for backward compat-
ibility.
recovery
(since: 3.2, default: off, deprecated since: 4.5)
NOTE:
This option has been replaced by usebackuproot and should not
be used but will work on 4.5+ kernels.
inode_cache, noinode_cache
(removed in: 5.11, since: 3.0, default: off)
NOTE:
The functionality has been removed in 5.11, any stale data
created by previous use of the inode_cache option can be re-
moved by btrfs rescue clear-ino-cache.
NOTES ON GENERIC MOUNT OPTIONS
Some of the general mount options from mount(8) that affect BTRFS and
are worth mentioning.
noatime
under read intensive work-loads, specifying noatime signifi-
cantly improves performance because no new access time informa-
tion needs to be written. Without this option, the default is
relatime, which only reduces the number of inode atime updates
in comparison to the traditional strictatime. The worst case for
atime updates under relatime occurs when many files are read
whose atime is older than 24 h and which are freshly snapshot-
ted. In that case the atime is updated and COW happens - for
each file - in bulk. See also https://lwn.net/Articles/499293/ -
Atime and btrfs: a bad combination? (LWN, 2012-05-31).
Note that noatime may break applications that rely on atime up-
times like the venerable Mutt (unless you use maildir mail-
boxes).
FILESYSTEM FEATURES
The basic set of filesystem features gets extended over time. The back-
ward compatibility is maintained and the features are optional, need to
be explicitly asked for so accidental use will not create incompatibil-
ities.
There are several classes and the respective tools to manage the fea-
tures:
at mkfs time only
This is namely for core structures, like the b-tree nodesize or
checksum algorithm, see mkfs.btrfs(8) for more details.
after mkfs, on an unmounted filesystem
Features that may optimize internal structures or add new struc-
tures to support new functionality, see btrfstune(8). The com-
mand btrfs inspect-internal dump-super /dev/sdx will dump a su-
perblock, you can map the value of incompat_flags to the fea-
tures listed below
after mkfs, on a mounted filesystem
The features of a filesystem (with a given UUID) are listed in
/sys/fs/btrfs/UUID/features/, one file per feature. The status
is stored inside the file. The value 1 is for enabled and ac-
tive, while 0 means the feature was enabled at mount time but
turned off afterwards.
Whether a particular feature can be turned on a mounted filesys-
tem can be found in the directory /sys/fs/btrfs/features/, one
file per feature. The value 1 means the feature can be enabled.
List of features (see also mkfs.btrfs(8) section FILESYSTEM FEATURES):
big_metadata
(since: 3.4)
the filesystem uses nodesize for metadata blocks, this can be
bigger than the page size
block_group_tree
(since: 6.1)
block group item representation using a dedicated b-tree, this
can greatly reduce mount time for large filesystems
compress_lzo
(since: 2.6.38)
the lzo compression has been used on the filesystem, either as a
mount option or via btrfs filesystem defrag.
compress_zstd
(since: 4.14)
the zstd compression has been used on the filesystem, either as
a mount option or via btrfs filesystem defrag.
default_subvol
(since: 2.6.34)
the default subvolume has been set on the filesystem
extended_iref
(since: 3.7)
increased hardlink limit per file in a directory to 65536, older
kernels supported a varying number of hardlinks depending on the
sum of all file name sizes that can be stored into one metadata
block
free_space_tree
(since: 4.5)
free space representation using a dedicated b-tree, successor of
v1 space cache
metadata_uuid
(since: 5.0)
the main filesystem UUID is the metadata_uuid, which stores the
new UUID only in the superblock while all metadata blocks still
have the UUID set at mkfs time, see btrfstune(8) for more
mixed_backref
(since: 2.6.31)
the last major disk format change, improved backreferences, now
default
mixed_groups
(since: 2.6.37)
mixed data and metadata block groups, i.e. the data and metadata
are not separated and occupy the same block groups, this mode is
suitable for small volumes as there are no constraints how the
remaining space should be used (compared to the split mode,
where empty metadata space cannot be used for data and vice
versa)
on the other hand, the final layout is quite unpredictable and
possibly highly fragmented, which means worse performance
no_holes
(since: 3.14)
improved representation of file extents where holes are not ex-
plicitly stored as an extent, saves a few percent of metadata if
sparse files are used
raid1c34
(since: 5.5)
extended RAID1 mode with copies on 3 or 4 devices respectively
RAID56 (since: 3.9)
the filesystem contains or contained a RAID56 profile of block
groups
rmdir_subvol
(since: 4.18)
indicate that rmdir(2) syscall can delete an empty subvolume
just like an ordinary directory. Note that this feature only de-
pends on the kernel version.
skinny_metadata
(since: 3.10)
reduced-size metadata for extent references, saves a few percent
of metadata
send_stream_version
(since: 5.10)
number of the highest supported send stream version
supported_checksums
(since: 5.5)
list of checksum algorithms supported by the kernel module, the
respective modules or built-in implementing the algorithms need
to be present to mount the filesystem, see section CHECKSUM AL-
GORITHMS.
supported_sectorsizes
(since: 5.13)
list of values that are accepted as sector sizes (mkfs.btrfs
--sectorsize) by the running kernel
supported_rescue_options
(since: 5.11)
list of values for the mount option rescue that are supported by
the running kernel, see btrfs(5)
zoned (since: 5.12)
zoned mode is allocation/write friendly to host-managed zoned
devices, allocation space is partitioned into fixed-size zones
that must be updated sequentially, see section ZONED MODE
SWAPFILE SUPPORT
A swapfile, when active, is a file-backed swap area. It is supported
since kernel 5.0. Use swapon(8) to activate it, until then (respec-
tively again after deactivating it with swapoff(8)) it's just a normal
file (with NODATACOW set), for which the special restrictions for ac-
tive swapfiles don't apply.
There are some limitations of the implementation in BTRFS and Linux
swap subsystem:
• filesystem - must be only single device
• filesystem - must have only single data profile
• subvolume - cannot be snapshotted if it contains any active swapfiles
• swapfile - must be preallocated (i.e. no holes)
• swapfile - must be NODATACOW (i.e. also NODATASUM, no compression)
The limitations come namely from the COW-based design and mapping layer
of blocks that allows the advanced features like relocation and
multi-device filesystems. However, the swap subsystem expects simpler
mapping and no background changes of the file block location once
they've been assigned to swap.
With active swapfiles, the following whole-filesystem operations will
skip swapfile extents or may fail:
• balance - block groups with extents of any active swapfiles are
skipped and reported, the rest will be processed normally
• resize grow - unaffected
• resize shrink - works as long as the extents of any active swapfiles
are outside of the shrunk range
• device add - if the new devices do not interfere with any already ac-
tive swapfiles this operation will work, though no new swapfile can
be activated afterwards
• device delete - if the device has been added as above, it can be also
deleted
• device replace - ditto
When there are no active swapfiles and a whole-filesystem exclusive op-
eration is running (e.g. balance, device delete, shrink), the swapfiles
cannot be temporarily activated. The operation must finish first.
To create and activate a swapfile run the following commands:
# truncate -s 0 swapfile
# chattr +C swapfile
# fallocate -l 2G swapfile
# chmod 0600 swapfile
# mkswap swapfile
# swapon swapfile
Since version 6.1 it's possible to create the swapfile in a single com-
mand (except the activation):
# btrfs filesystem mkswapfile --size 2G swapfile
# swapon swapfile
Please note that the UUID returned by the mkswap utility identifies the
swap "filesystem" and because it's stored in a file, it's not generally
visible and usable as an identifier unlike if it was on a block device.
Once activated the file will appear in /proc/swaps:
# cat /proc/swaps
Filename Type Size Used Priority
/path/swapfile file 2097152 0 -2
The swapfile can be created as one-time operation or, once properly
created, activated on each boot by the swapon -a command (usually
started by the service manager). Add the following entry to /etc/fstab,
assuming the filesystem that provides the /path has been already
mounted at this point. Additional mount options relevant for the swap-
file can be set too (like priority, not the BTRFS mount options).
/path/swapfile none swap defaults 0 0
From now on the subvolume with the active swapfile cannot be snapshot-
ted until the swapfile is deactivated again by swapoff. Then the swap-
file is a regular file and the subvolume can be snapshotted again,
though this would prevent another activation any swapfile that has been
snapshotted. New swapfiles (not snapshotted) can be created and acti-
vated.
Otherwise, an inactive swapfile does not affect the containing subvol-
ume. Activation creates a temporary in-memory status and prevents some
file operations, but is not stored permanently.
HIBERNATION
A swapfile can be used for hibernation but it's not straightforward.
Before hibernation a resume offset must be written to file
/sys/power/resume_offset or the kernel command line parameter re-
sume_offset must be set.
The value is the physical offset on the device. Note that this is not
the same value that filefrag prints as physical offset!
Btrfs filesystem uses mapping between logical and physical addresses
but here the physical can still map to one or more device-specific
physical block addresses. It's the device-specific physical offset that
is suitable as resume offset.
Since version 6.1 there's a command btrfs inspect-internal map-swapfile
that will print the device physical offset and the adjusted value for
/sys/power/resume_offset. Note that the value is divided by page size,
i.e. it's not the offset itself.
# btrfs filesystem mkswapfile swapfile
# btrfs inspect-internal map-swapfile swapfile
Physical start: 811511726080
Resume offset: 198122980
For scripting and convenience the option -r will print just the offset:
# btrfs inspect-internal map-swapfile -r swapfile
198122980
The command map-swapfile also verifies all the requirements, i.e. no
holes, single device, etc.
TROUBLESHOOTING
If the swapfile activation fails please verify that you followed all
the steps above or check the system log (e.g. dmesg or journalctl) for
more information.
Notably, the swapon utility exits with a message that does not say what
failed:
# swapon /path/swapfile
swapon: /path/swapfile: swapon failed: Invalid argument
The specific reason is likely to be printed to the system log by the
btrfs module:
# journalctl -t kernel | grep swapfile
kernel: BTRFS warning (device sda): swapfile must have single data profile
CHECKSUM ALGORITHMS
Data and metadata are checksummed by default, the checksum is calcu-
lated before write and verified after reading the blocks from devices.
The whole metadata block has a checksum stored inline in the b-tree
node header, each data block has a detached checksum stored in the
checksum tree.
There are several checksum algorithms supported. The default and back-
ward compatible is crc32c. Since kernel 5.5 there are three more with
different characteristics and trade-offs regarding speed and strength.
The following list may help you to decide which one to select.
CRC32C (32bit digest)
default, best backward compatibility, very fast, modern CPUs
have instruction-level support, not collision-resistant but
still good error detection capabilities
XXHASH (64bit digest)
can be used as CRC32C successor, very fast, optimized for modern
CPUs utilizing instruction pipelining, good collision resistance
and error detection
SHA256 (256bit digest)
a cryptographic-strength hash, relatively slow but with possible
CPU instruction acceleration or specialized hardware cards, FIPS
certified and in wide use
BLAKE2b (256bit digest)
a cryptographic-strength hash, relatively fast with possible CPU
acceleration using SIMD extensions, not standardized but based
on BLAKE which was a SHA3 finalist, in wide use, the algorithm
used is BLAKE2b-256 that's optimized for 64bit platforms
The digest size affects overall size of data block checksums stored in
the filesystem. The metadata blocks have a fixed area up to 256 bits
(32 bytes), so there's no increase. Each data block has a separate
checksum stored, with additional overhead of the b-tree leaves.
Approximate relative performance of the algorithms, measured against
CRC32C using reference software implementations on a 3.5GHz intel CPU:
┌─────────┬─────────────┬───────┬─────────────────┐
│ Digest │ Cycles/4KiB │ Ratio │ Implementation │
├─────────┼─────────────┼───────┼─────────────────┤
│ CRC32C │ 1700 │ 1.00 │ CPU instruction │
├─────────┼─────────────┼───────┼─────────────────┤
│ XXHASH │ 2500 │ 1.44 │ reference impl. │
├─────────┼─────────────┼───────┼─────────────────┤
│ SHA256 │ 105000 │ 61 │ reference impl. │
├─────────┼─────────────┼───────┼─────────────────┤
│ SHA256 │ 36000 │ 21 │ libgcrypt/AVX2 │
├─────────┼─────────────┼───────┼─────────────────┤
│ SHA256 │ 63000 │ 37 │ libsodium/AVX2 │
├─────────┼─────────────┼───────┼─────────────────┤
│ BLAKE2b │ 22000 │ 13 │ reference impl. │
├─────────┼─────────────┼───────┼─────────────────┤
│ BLAKE2b │ 19000 │ 11 │ libgcrypt/AVX2 │
├─────────┼─────────────┼───────┼─────────────────┤
│ BLAKE2b │ 19000 │ 11 │ libsodium/AVX2 │
└─────────┴─────────────┴───────┴─────────────────┘
Many kernels are configured with SHA256 as built-in and not as a mod-
ule. The accelerated versions are however provided by the modules and
must be loaded explicitly (modprobe sha256) before mounting the
filesystem to make use of them. You can check in
/sys/fs/btrfs/FSID/checksum which one is used. If you see
sha256-generic, then you may want to unmount and mount the filesystem
again, changing that on a mounted filesystem is not possible. Check
the file /proc/crypto, when the implementation is built-in, you'd find
name : sha256
driver : sha256-generic
module : kernel
priority : 100
...
while accelerated implementation is e.g.
name : sha256
driver : sha256-avx2
module : sha256_ssse3
priority : 170
...
COMPRESSION
Btrfs supports transparent file compression. There are three algorithms
available: ZLIB, LZO and ZSTD (since v4.14), with various levels. The
compression happens on the level of file extents and the algorithm is
selected by file property, mount option or by a defrag command. You
can have a single btrfs mount point that has some files that are uncom-
pressed, some that are compressed with LZO, some with ZLIB, for in-
stance (though you may not want it that way, it is supported).
Once the compression is set, all newly written data will be compressed,
i.e. existing data are untouched. Data are split into smaller chunks
(128KiB) before compression to make random rewrites possible without a
high performance hit. Due to the increased number of extents the meta-
data consumption is higher. The chunks are compressed in parallel.
The algorithms can be characterized as follows regarding the speed/ra-
tio trade-offs:
ZLIB
• slower, higher compression ratio
• levels: 1 to 9, mapped directly, default level is 3
• good backward compatibility
LZO
• faster compression and decompression than ZLIB, worse compres-
sion ratio, designed to be fast
• no levels
• good backward compatibility
ZSTD
• compression comparable to ZLIB with higher compression/decom-
pression speeds and different ratio
• levels: 1 to 15, mapped directly (higher levels are not avail-
able)
• since 4.14, levels since 5.1
The differences depend on the actual data set and cannot be expressed
by a single number or recommendation. Higher levels consume more CPU
time and may not bring a significant improvement, lower levels are
close to real time.
HOW TO ENABLE COMPRESSION
Typically the compression can be enabled on the whole filesystem, spec-
ified for the mount point. Note that the compression mount options are
shared among all mounts of the same filesystem, either bind mounts or
subvolume mounts. Please refer to btrfs(5) section MOUNT OPTIONS.
$ mount -o compress=zstd /dev/sdx /mnt
This will enable the zstd algorithm on the default level (which is 3).
The level can be specified manually too like zstd:3. Higher levels com-
press better at the cost of time. This in turn may cause increased
write latency, low levels are suitable for real-time compression and on
reasonably fast CPU don't cause noticeable performance drops.
$ btrfs filesystem defrag -czstd file
The command above will start defragmentation of the whole file and ap-
ply the compression, regardless of the mount option. (Note: specifying
level is not yet implemented). The compression algorithm is not persis-
tent and applies only to the defragmentation command, for any other
writes other compression settings apply.
Persistent settings on a per-file basis can be set in two ways:
$ chattr +c file
$ btrfs property set file compression zstd
The first command is using legacy interface of file attributes inher-
ited from ext2 filesystem and is not flexible, so by default the zlib
compression is set. The other command sets a property on the file with
the given algorithm. (Note: setting level that way is not yet imple-
mented.)
COMPRESSION LEVELS
The level support of ZLIB has been added in v4.14, LZO does not support
levels (the kernel implementation provides only one), ZSTD level sup-
port has been added in v5.1.
There are 9 levels of ZLIB supported (1 to 9), mapping 1:1 from the
mount option to the algorithm defined level. The default is level 3,
which provides the reasonably good compression ratio and is still rea-
sonably fast. The difference in compression gain of levels 7, 8 and 9
is comparable but the higher levels take longer.
The ZSTD support includes levels 1 to 15, a subset of full range of
what ZSTD provides. Levels 1-3 are real-time, 4-8 slower with improved
compression and 9-15 try even harder though the resulting size may not
be significantly improved.
Level 0 always maps to the default. The compression level does not af-
fect compatibility.
INCOMPRESSIBLE DATA
Files with already compressed data or with data that won't compress
well with the CPU and memory constraints of the kernel implementations
are using a simple decision logic. If the first portion of data being
compressed is not smaller than the original, the compression of the
file is disabled -- unless the filesystem is mounted with com-
press-force. In that case compression will always be attempted on the
file only to be later discarded. This is not optimal and subject to op-
timizations and further development.
If a file is identified as incompressible, a flag is set (NOCOMPRESS)
and it's sticky. On that file compression won't be performed unless
forced. The flag can be also set by chattr +m (since e2fsprogs 1.46.2)
or by properties with value no or none. Empty value will reset it to
the default that's currently applicable on the mounted filesystem.
There are two ways to detect incompressible data:
• actual compression attempt - data are compressed, if the result is
not smaller, it's discarded, so this depends on the algorithm and
level
• pre-compression heuristics - a quick statistical evaluation on the
data is performed and based on the result either compression is per-
formed or skipped, the NOCOMPRESS bit is not set just by the heuris-
tic, only if the compression algorithm does not make an improvement
$ lsattr file
---------------------m file
Using the forcing compression is not recommended, the heuristics are
supposed to decide that and compression algorithms internally detect
incompressible data too.
PRE-COMPRESSION HEURISTICS
The heuristics aim to do a few quick statistical tests on the com-
pressed data in order to avoid probably costly compression that would
turn out to be inefficient. Compression algorithms could have internal
detection of incompressible data too but this leads to more overhead as
the compression is done in another thread and has to write the data
anyway. The heuristic is read-only and can utilize cached memory.
The tests performed based on the following: data sampling, long re-
peated pattern detection, byte frequency, Shannon entropy.
COMPATIBILITY
Compression is done using the COW mechanism so it's incompatible with
nodatacow. Direct IO works on compressed files but will fall back to
buffered writes and leads to recompression. Currently nodatasum and
compression don't work together.
The compression algorithms have been added over time so the version
compatibility should be also considered, together with other tools that
may access the compressed data like bootloaders.
SYSFS INTERFACE
Btrfs has a sysfs interface to provide extra knobs.
The top level path is /sys/fs/btrfs/, and the main directory layout is
the following:
┌──────────────────────────────┬─────────────────────┬─────────┐
│ Relative Path │ Description │ Version │
├──────────────────────────────┼─────────────────────┼─────────┤
│ features/ │ All supported fea- │ 3.14+ │
│ │ tures │ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/ │ Mounted fs UUID │ 3.14+ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/allocation/ │ Space allocation │ 3.14+ │
│ │ info │ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/features/ │ Features of the │ 3.14+ │
│ │ filesystem │ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/devices/<DE- │ Symlink to each │ 5.6+ │
│ VID>/ │ block device sysfs │ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/devinfo/<DE- │ Btrfs specific info │ 5.6+ │
│ VID>/ │ for each device │ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/qgroups/ │ Global qgroup info │ 5.9+ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/qgroups/<LEVEL>_<ID>/ │ Info for each │ 5.9+ │
│ │ qgroup │ │
├──────────────────────────────┼─────────────────────┼─────────┤
│ <UUID>/discard/ │ Discard stats and │ 6.1+ │
│ │ tunables │ │
└──────────────────────────────┴─────────────────────┴─────────┘
For /sys/fs/btrfs/features/ directory, each file means a supported fea-
ture for the current kernel.
For /sys/fs/btrfs/<UUID>/features/ directory, each file means an en-
abled feature for the mounted filesystem.
The features shares the same name in section FILESYSTEM FEATURES.
Files in /sys/fs/btrfs/<UUID>/ directory are:
bg_reclaim_threshold
(RW, since: 5.19)
Used space percentage of total device space to start auto block
group claim. Mostly for zoned devices.
checksum
(RO, since: 5.5)
The checksum used for the mounted filesystem. This includes
both the checksum type (see section CHECKSUM ALGORITHMS) and the
implemented driver (mostly shows if it's hardware accelerated).
clone_alignment
(RO, since: 3.16)
The bytes alignment for clone and dedupe ioctls.
commit_stats
(RW, since: 6.0)
The performance statistics for btrfs transaction commit. Mostly
for debug purposes.
Writing into this file will reset the maximum commit duration to
the input value.
exclusive_operation
(RO, since: 5.10)
Shows the running exclusive operation. Check section FILESYSTEM
EXCLUSIVE OPERATIONS for details.
generation
(RO, since: 5.11)
Show the generation of the mounted filesystem.
label (RW, since: 3.14)
Show the current label of the mounted filesystem.
metadata_uuid
(RO, since: 5.0)
Shows the metadata uuid of the mounted filesystem. Check meta-
data_uuid feature for more details.
nodesize
(RO, since: 3.14)
Show the nodesize of the mounted filesystem.
quota_override
(RW, since: 4.13)
Shows the current quota override status. 0 means no quota over-
ride. 1 means quota override, quota can ignore the existing
limit settings.
read_policy
(RW, since: 5.11)
Shows the current balance policy for reads. Currently only
"pid" (balance using pid value) is supported.
sectorsize
(RO, since: 3.14)
Shows the sectorsize of the mounted filesystem.
Files and directories in /sys/fs/btrfs/<UUID>/allocations directory
are:
global_rsv_reserved
(RO, since: 3.14)
The used bytes of the global reservation.
global_rsv_size
(RO, since: 3.14)
The total size of the global reservation.
data/, metadata/ and system/ directories
(RO, since: 5.14)
Space info accounting for the 3 chunk types. Mostly for debug
purposes.
Files in /sys/fs/btrfs/<UUID>/allocations/data,metadata,system direc-
tory are:
bg_reclaim_threshold
(RW, since: 5.19)
Reclaimable space percentage of block group's size (excluding
permanently unusable space) to reclaim the block group. Can be
used on regular or zoned devices.
chunk_size
(RW, since: 6.0)
Shows the chunk size. Can be changed for data and metadata.
Cannot be set for zoned devices.
Files in /sys/fs/btrfs/<UUID>/devinfo/<DEVID> directory are:
error_stats:
(RO, since: 5.14)
Shows all the history error numbers of the device.
fsid: (RO, since: 5.17)
Shows the fsid which the device belongs to. It can be different
than the <UUID> if it's a seed device.
in_fs_metadata
(RO, since: 5.6)
Shows whether we have found the device. Should always be 1, as
if this turns to 0, the <DEVID> directory would get removed au-
tomatically.
missing
(RO, since: 5.6)
Shows whether the device is missing.
replace_target
(RO, since: 5.6)
Shows whether the device is the replace target. If no dev-re-
place is running, this value should be 0.
scrub_speed_max
(RW, since: 5.14)
Shows the scrub speed limit for this device. The unit is
Bytes/s. 0 means no limit.
writeable
(RO, since: 5.6)
Show if the device is writeable.
Files in /sys/fs/btrfs/<UUID>/qgroups/ directory are:
enabled
(RO, since: 6.1)
Shows if qgroup is enabled. Also, if qgroup is disabled, the
qgroups directory would be removed automatically.
inconsistent
(RO, since: 6.1)
Shows if the qgroup numbers are inconsistent. If 1, it's recom-
mended to do a qgroup rescan.
drop_subtree_threshold
(RW, since: 6.1)
Shows the subtree drop threshold to automatically mark qgroup
inconsistent.
When dropping large subvolumes with qgroup enabled, there would
be a huge load for qgroup accounting. If we have a subtree
whose level is larger than or equal to this value, we will not
trigger qgroup account at all, but mark qgroup inconsistent to
avoid the huge workload.
Default value is 8, where no subtree drop can trigger qgroup.
Lower value can reduce qgroup workload, at the cost of extra
qgroup rescan to re-calculate the numbers.
Files in /sys/fs/btrfs/<UUID>/<LEVEL>_<ID>/ directory are:
exclusive
(RO, since: 5.9)
Shows the exclusively owned bytes of the qgroup.
limit_flags
(RO, since: 5.9)
Shows the numeric value of the limit flags. If 0, means no
limit implied.
max_exclusive
(RO, since: 5.9)
Shows the limits on exclusively owned bytes.
max_referenced
(RO, since: 5.9)
Shows the limits on referenced bytes.
referenced
(RO, since: 5.9)
Shows the referenced bytes of the qgroup.
rsv_data
(RO, since: 5.9)
Shows the reserved bytes for data.
rsv_meta_pertrans
(RO, since: 5.9)
Shows the reserved bytes for per transaction metadata.
rsv_meta_prealloc
(RO, since: 5.9)
Shows the reserved bytes for preallocated metadata.
Files in /sys/fs/btrfs/<UUID>/discard/ directory are:
discardable_bytes
(RO, since: 6.1)
Shows amount of bytes that can be discarded in the async discard
and nodiscard mode.
discardable_extents
(RO, since: 6.1)
Shows number of extents to be discarded in the async discard and
nodiscard mode.
discard_bitmap_bytes
(RO, since: 6.1)
Shows amount of discarded bytes from data tracked as bitmaps.
discard_extent_bytes
(RO, since: 6.1)
Shows amount of discarded extents from data tracked as bitmaps.
discard_bytes_saved
(RO, since: 6.1)
Shows the amount of bytes that were reallocated without being
discarded.
kbps_limit
(RW, since: 6.1)
Tunable limit of kilobytes per second issued as discard IO in
the async discard mode.
iops_limit
(RW, since: 6.1)
Tunable limit of number of discard IO operations to be issued in
the async discard mode.
max_discard_size
(RW, since: 6.1)
Tunable limit for size of one IO discard request.
FILESYSTEM EXCLUSIVE OPERATIONS
There are several operations that affect the whole filesystem and can-
not be run in parallel. Attempt to start one while another is running
will fail (see exceptions below).
Since kernel 5.10 the currently running operation can be obtained from
/sys/fs/UUID/exclusive_operation with following values and operations:
• balance
• balance paused (since 5.17)
• device add
• device delete
• device replace
• resize
• swapfile activate
• none
Enqueuing is supported for several btrfs subcommands so they can be
started at once and then serialized.
There's an exception when a paused balance allows to start a device add
operation as they don't really collide and this can be used to add more
space for the balance to finish.
FILESYSTEM LIMITS
maximum file name length
255
This limit is imposed by Linux VFS, the structures of BTRFS
could store larger file names.
maximum symlink target length
depends on the nodesize value, for 4KiB it's 3949 bytes, for
larger nodesize it's 4095 due to the system limit PATH_MAX
The symlink target may not be a valid path, i.e. the path name
components can exceed the limits (NAME_MAX), there's no content
validation at symlink(3) creation.
maximum number of inodes
264 but depends on the available metadata space as the inodes
are created dynamically
Each subvolume is an independent namespace of inodes and thus
their numbers, so the limit is per subvolume, not for the whole
filesystem.
inode numbers
minimum number: 256 (for subvolumes), regular files and directo-
ries: 257, maximum number: (264 - 256)
The inode numbers that can be assigned to user created files are
from the whole 64bit space except first 256 and last 256 in that
range that are reserved for internal b-tree identifiers.
maximum file length
inherent limit of BTRFS is 264 (16 EiB) but the practical limit
of Linux VFS is 263 (8 EiB)
maximum number of subvolumes
the subvolume ids can go up to 248 but the number of actual sub-
volumes depends on the available metadata space
The space consumed by all subvolume metadata includes bookkeep-
ing of shared extents can be large (MiB, GiB). The range is not
the full 64bit range because of qgroups that use the upper 16
bits for another purposes.
maximum number of hardlinks of a file in a directory
65536 when the extref feature is turned on during mkfs (de-
fault), roughly 100 otherwise and depends on file name length
that fits into one metadata node
minimum filesystem size
the minimal size of each device depends on the mixed-bg feature,
without that (the default) it's about 109MiB, with mixed-bg it's
is 16MiB
BOOTLOADER SUPPORT
GRUB2 (https://www.gnu.org/software/grub) has the most advanced support
of booting from BTRFS with respect to features.
U-Boot (https://www.denx.de/wiki/U-Boot/) has decent support for boot-
ing but not all BTRFS features are implemented, check the documenta-
tion.
In general, the first 1MiB on each device is unused with the exception
of primary superblock that is on the offset 64KiB and spans 4KiB. The
rest can be freely used by bootloaders or for other system information.
Note that booting from a filesystem on zoned device is not supported.
FILE ATTRIBUTES
The btrfs filesystem supports setting file attributes or flags. Note
there are old and new interfaces, with confusing names. The following
list should clarify that:
• attributes: chattr(1) or lsattr(1) utilities (the ioctls are
FS_IOC_GETFLAGS and FS_IOC_SETFLAGS), due to the ioctl names the at-
tributes are also called flags
• xflags: to distinguish from the previous, it's extended flags, with
tunable bits similar to the attributes but extensible and new bits
will be added in the future (the ioctls are FS_IOC_FSGETXATTR and
FS_IOC_FSSETXATTR but they are not related to extended attributes
that are also called xattrs), there's no standard tool to change the
bits, there's support in xfs_io(8) as command xfs_io -c chattr
Attributes
a append only, new writes are always written at the end of the
file
A no atime updates
c compress data, all data written after this attribute is set will
be compressed. Please note that compression is also affected by
the mount options or the parent directory attributes.
When set on a directory, all newly created files will inherit
this attribute. This attribute cannot be set with 'm' at the
same time.
C no copy-on-write, file data modifications are done in-place
When set on a directory, all newly created files will inherit
this attribute.
NOTE:
Due to implementation limitations, this flag can be set/unset
only on empty files.
d no dump, makes sense with 3rd party tools like dump(8), on BTRFS
the attribute can be set/unset but no other special handling is
done
D synchronous directory updates, for more details search open(2)
for O_SYNC and O_DSYNC
i immutable, no file data and metadata changes allowed even to the
root user as long as this attribute is set (obviously the excep-
tion is unsetting the attribute)
m no compression, permanently turn off compression on the given
file. Any compression mount options will not affect this file.
(chattr support added in 1.46.2)
When set on a directory, all newly created files will inherit
this attribute. This attribute cannot be set with c at the same
time.
S synchronous updates, for more details search open(2) for O_SYNC
and O_DSYNC
No other attributes are supported. For the complete list please refer
to the chattr(1) manual page.
XFLAGS
There's an overlap of letters assigned to the bits with the attributes,
this list refers to what xfs_io(8) provides:
i immutable, same as the attribute
a append only, same as the attribute
s synchronous updates, same as the attribute S
A no atime updates, same as the attribute
d no dump, same as the attribute
ZONED MODE
Since version 5.12 btrfs supports so called zoned mode. This is a spe-
cial on-disk format and allocation/write strategy that's friendly to
zoned devices. In short, a device is partitioned into fixed-size zones
and each zone can be updated by append-only manner, or reset. As btrfs
has no fixed data structures, except the super blocks, the zoned mode
only requires block placement that follows the device constraints. You
can learn about the whole architecture at https://zonedstorage.io .
The devices are also called SMR/ZBC/ZNS, in host-managed mode. Note
that there are devices that appear as non-zoned but actually are, this
is drive-managed and using zoned mode won't help.
The zone size depends on the device, typical sizes are 256MiB or 1GiB.
In general it must be a power of two. Emulated zoned devices like
null_blk allow to set various zone sizes.
Requirements, limitations
• all devices must have the same zone size
• maximum zone size is 8GiB
• minimum zone size is 4MiB
• mixing zoned and non-zoned devices is possible, the zone writes are
emulated, but this is namely for testing
• the super block is handled in a special way and is at different loca-
tions than on a non-zoned filesystem:
• primary: 0B (and the next two zones)
• secondary: 512GiB (and the next two zones)
• tertiary: 4TiB (4096GiB, and the next two zones)
Incompatible features
The main constraint of the zoned devices is lack of in-place update of
the data. This is inherently incompatible with some features:
• NODATACOW - overwrite in-place, cannot create such files
• fallocate - preallocating space for in-place first write
• mixed-bg - unordered writes to data and metadata, fixing that means
using separate data and metadata block groups
• booting - the zone at offset 0 contains superblock, resetting the
zone would destroy the bootloader data
Initial support lacks some features but they're planned:
• only single (data, metadata) and DUP (metadata) profile is supported
• fstrim - due to dependency on free space cache v1
Super block
As said above, super block is handled in a special way. In order to be
crash safe, at least one zone in a known location must contain a valid
superblock. This is implemented as a ring buffer in two consecutive
zones, starting from known offsets 0B, 512GiB and 4TiB.
The values are different than on non-zoned devices. Each new super
block is appended to the end of the zone, once it's filled, the zone is
reset and writes continue to the next one. Looking up the latest super
block needs to read offsets of both zones and determine the last writ-
ten version.
The amount of space reserved for super block depends on the zone size.
The secondary and tertiary copies are at distant offsets as the capac-
ity of the devices is expected to be large, tens of terabytes. Maximum
zone size supported is 8GiB, which would mean that e.g. offset 0-16GiB
would be reserved just for the super block on a hypothetical device of
that zone size. This is wasteful but required to guarantee crash
safety.
Devices
Real hardware
The WD Ultrastar series 600 advertises HM-SMR, i.e. the host-managed
zoned mode. There are two more: DA (device managed, no zoned informa-
tion exported to the system), HA (host aware, can be used as regular
disk but zoned writes improve performance). There are not many devices
available at the moment, the information about exact zoned mode is hard
to find, check data sheets or community sources gathering information
from real devices.
Note: zoned mode won't work with DM-SMR disks.
• Ultrastar® DC ZN540 NVMe ZNS SSD (product brief)
Emulated: null_blk
The driver null_blk provides memory backed device and is suitable for
testing. There are some quirks setting up the devices. The module must
be loaded with nr_devices=0 or the numbering of device nodes will be
offset. The configfs must be mounted at /sys/kernel/config and the ad-
ministration of the null_blk devices is done in /sys/kernel/con-
fig/nullb. The device nodes are named like /dev/nullb0 and are numbered
sequentially. NOTE: the device name may be different than the named di-
rectory in sysfs!
Setup:
modprobe configfs
modprobe null_blk nr_devices=0
Create a device mydev, assuming no other previously created devices,
size is 2048MiB, zone size 256MiB. There are more tunable parameters,
this is a minimal example taking defaults:
cd /sys/kernel/config/nullb/
mkdir mydev
cd mydev
echo 2048 > size
echo 1 > zoned
echo 1 > memory_backed
echo 256 > zone_size
echo 1 > power
This will create a device /dev/nullb0 and the value of file index will
match the ending number of the device node.
Remove the device:
rmdir /sys/kernel/config/nullb/mydev
Then continue with mkfs.btrfs /dev/nullb0, the zoned mode is auto-de-
tected.
For convenience, there's a script wrapping the basic null_blk manage-
ment operations https://github.com/kdave/nullb.git, the above commands
become:
nullb setup
nullb create -s 2g -z 256
mkfs.btrfs /dev/nullb0
...
nullb rm nullb0
Emulated: TCMU runner
TCMU is a framework to emulate SCSI devices in userspace, providing
various backends for the storage, with zoned support as well. A
file-backed zoned device can provide more options for larger storage
and zone size. Please follow the instructions at
https://zonedstorage.io/projects/tcmu-runner/ .
Compatibility, incompatibility
• the feature sets an incompat bit and requires new kernel to access
the filesystem (for both read and write)
• superblock needs to be handled in a special way, there are still 3
copies but at different offsets (0, 512GiB, 4TiB) and the 2 consecu-
tive zones are a ring buffer of the superblocks, finding the latest
one needs reading it from the write pointer or do a full scan of the
zones
• mixing zoned and non zoned devices is possible (zones are emulated)
but is recommended only for testing
• mixing zoned devices with different zone sizes is not possible
• zone sizes must be power of two, zone sizes of real devices are e.g.
256MiB or 1GiB, larger size is expected, maximum zone size supported
by btrfs is 8GiB
Status, stability, reporting bugs
The zoned mode has been released in 5.12 and there are still some rough
edges and corner cases one can hit during testing. Please report bugs
to https://github.com/naota/linux/issues/ .
References
• https://zonedstorage.io
• https://zonedstorage.io/projects/libzbc/ -- libzbc is library and
set of tools to directly manipulate devices with ZBC/ZAC support
• https://zonedstorage.io/projects/libzbd/ -- libzbd uses the kernel
provided zoned block device interface based on the ioctl() system
calls
• https://hddscan.com/blog/2020/hdd-wd-smr.html -- some details about
exact device types
• https://lwn.net/Articles/853308/ -- Btrfs on zoned block devices
• https://www.usenix.org/conference/vault20/presentation/bjorling --
Zone Append: A New Way of Writing to Zoned Storage
CONTROL DEVICE
There's a character special device /dev/btrfs-control with major and
minor numbers 10 and 234 (the device can be found under the misc cate-
gory).
$ ls -l /dev/btrfs-control
crw------- 1 root root 10, 234 Jan 1 12:00 /dev/btrfs-control
The device accepts some ioctl calls that can perform following actions
on the filesystem module:
• scan devices for btrfs filesystem (i.e. to let multi-device filesys-
tems mount automatically) and register them with the kernel module
• similar to scan, but also wait until the device scanning process is
finished for a given filesystem
• get the supported features (can be also found under
/sys/fs/btrfs/features)
The device is created when btrfs is initialized, either as a module or
a built-in functionality and makes sense only in connection with that.
Running e.g. mkfs without the module loaded will not register the de-
vice and will probably warn about that.
In rare cases when the module is loaded but the device is not present
(most likely accidentally deleted), it's possible to recreate it by
# mknod --mode=600 /dev/btrfs-control c 10 234
or (since 5.11) by a convenience command
# btrfs rescue create-control-device
The control device is not strictly required but the device scanning
will not work and a workaround would need to be used to mount a
multi-device filesystem. The mount option device can trigger the de-
vice scanning during mount, see also btrfs device scan.
FILESYSTEM WITH MULTIPLE PROFILES
It is possible that a btrfs filesystem contains multiple block group
profiles of the same type. This could happen when a profile conversion
using balance filters is interrupted (see btrfs-balance(8)). Some
btrfs commands perform a test to detect this kind of condition and
print a warning like this:
WARNING: Multiple block group profiles detected, see 'man btrfs(5)'.
WARNING: Data: single, raid1
WARNING: Metadata: single, raid1
The corresponding output of btrfs filesystem df might look like:
WARNING: Multiple block group profiles detected, see 'man btrfs(5)'.
WARNING: Data: single, raid1
WARNING: Metadata: single, raid1
Data, RAID1: total=832.00MiB, used=0.00B
Data, single: total=1.63GiB, used=0.00B
System, single: total=4.00MiB, used=16.00KiB
Metadata, single: total=8.00MiB, used=112.00KiB
Metadata, RAID1: total=64.00MiB, used=32.00KiB
GlobalReserve, single: total=16.25MiB, used=0.00B
There's more than one line for type Data and Metadata, while the pro-
files are single and RAID1.
This state of the filesystem OK but most likely needs the user/adminis-
trator to take an action and finish the interrupted tasks. This cannot
be easily done automatically, also the user knows the expected final
profiles.
In the example above, the filesystem started as a single device and
single block group profile. Then another device was added, followed by
balance with convert=raid1 but for some reason hasn't finished.
Restarting the balance with convert=raid1 will continue and end up with
filesystem with all block group profiles RAID1.
NOTE:
If you're familiar with balance filters, you can use con-
vert=raid1,profiles=single,soft, which will take only the uncon-
verted single profiles and convert them to raid1. This may speed up
the conversion as it would not try to rewrite the already convert
raid1 profiles.
Having just one profile is desired as this also clearly defines the
profile of newly allocated block groups, otherwise this depends on in-
ternal allocation policy. When there are multiple profiles present, the
order of selection is RAID56, RAID10, RAID1, RAID0 as long as the de-
vice number constraints are satisfied.
Commands that print the warning were chosen so they're brought to user
attention when the filesystem state is being changed in that regard.
This is: device add, device delete, balance cancel, balance pause. Com-
mands that report space usage: filesystem df, device usage. The command
filesystem usage provides a line in the overall summary:
Multiple profiles: yes (data, metadata)
SEEDING DEVICE
The COW mechanism and multiple devices under one hood enable an inter-
esting concept, called a seeding device: extending a read-only filesys-
tem on a device with another device that captures all writes. For exam-
ple imagine an immutable golden image of an operating system enhanced
with another device that allows to use the data from the golden image
and normal operation. This idea originated on CD-ROMs with base OS and
allowing to use them for live systems, but this became obsolete. There
are technologies providing similar functionality, like unionmount,
overlayfs or qcow2 image snapshot.
The seeding device starts as a normal filesystem, once the contents is
ready, btrfstune -S 1 is used to flag it as a seeding device. Mounting
such device will not allow any writes, except adding a new device by
btrfs device add. Then the filesystem can be remounted as read-write.
Given that the filesystem on the seeding device is always recognized as
read-only, it can be used to seed multiple filesystems from one device
at the same time. The UUID that is normally attached to a device is au-
tomatically changed to a random UUID on each mount.
Once the seeding device is mounted, it needs the writable device. After
adding it, unmounting and mounting with umount /path; mount
/dev/writable /path or remounting read-write with remount -o remount,rw
makes the filesystem at /path ready for use.
NOTE:
There is a known bug with using remount to make the mount writeable:
remount will leave the filesystem in a state where it is unable to
clean deleted snapshots, so it will leak space until it is unmounted
and mounted properly.
Furthermore, deleting the seeding device from the filesystem can turn
it into a normal filesystem, provided that the writable device can also
contain all the data from the seeding device.
The seeding device flag can be cleared again by btrfstune -f -S 0, e.g.
allowing to update with newer data but please note that this will in-
validate all existing filesystems that use this particular seeding de-
vice. This works for some use cases, not for others, and the forcing
flag to the command is mandatory to avoid accidental mistakes.
Example how to create and use one seeding device:
# mkfs.btrfs /dev/sda
# mount /dev/sda /mnt/mnt1
... fill mnt1 with data
# umount /mnt/mnt1
# btrfstune -S 1 /dev/sda
# mount /dev/sda /mnt/mnt1
# btrfs device add /dev/sdb /mnt/mnt1
# umount /mnt/mnt1
# mount /dev/sdb /mnt/mnt1
... /mnt/mnt1 is now writable
Now /mnt/mnt1 can be used normally. The device /dev/sda can be mounted
again with a another writable device:
# mount /dev/sda /mnt/mnt2
# btrfs device add /dev/sdc /mnt/mnt2
# umount /mnt/mnt2
# mount /dev/sdc /mnt/mnt2
... /mnt/mnt2 is now writable
The writable device (file:/dev/sdb) can be decoupled from the seeding
device and used independently:
# btrfs device delete /dev/sda /mnt/mnt1
As the contents originated in the seeding device, it's possible to turn
/dev/sdb to a seeding device again and repeat the whole process.
A few things to note:
• it's recommended to use only single device for the seeding device, it
works for multiple devices but the single profile must be used in or-
der to make the seeding device deletion work
• block group profiles single and dup support the use cases above
• the label is copied from the seeding device and can be changed by
btrfs filesystem label
• each new mount of the seeding device gets a new random UUID
• umount /path; mount /dev/writable /path can be replaced with mount -o
remount,rw /path but it won't reclaim space of deleted subvolumes un-
til the seeding device is mounted read-write again before making it
seeding again
Chained seeding devices
Though it's not recommended and is rather an obscure and untested use
case, chaining seeding devices is possible. In the first example, the
writable device /dev/sdb can be turned onto another seeding device
again, depending on the unchanged seeding device /dev/sda. Then using
/dev/sdb as the primary seeding device it can be extended with another
writable device, say /dev/sdd, and it continues as before as a simple
tree structure on devices.
# mkfs.btrfs /dev/sda
# mount /dev/sda /mnt/mnt1
... fill mnt1 with data
# umount /mnt/mnt1
# btrfstune -S 1 /dev/sda
# mount /dev/sda /mnt/mnt1
# btrfs device add /dev/sdb /mnt/mnt1
# mount -o remount,rw /mnt/mnt1
... /mnt/mnt1 is now writable
# umount /mnt/mnt1
# btrfstune -S 1 /dev/sdb
# mount /dev/sdb /mnt/mnt1
# btrfs device add /dev/sdc /mnt
# mount -o remount,rw /mnt/mnt1
... /mnt/mnt1 is now writable
# umount /mnt/mnt1
As a result we have:
• sda is a single seeding device, with its initial contents
• sdb is a seeding device but requires sda, the contents are from the
time when sdb is made seeding, i.e. contents of sda with any later
changes
• sdc last writable, can be made a seeding one the same way as was sdb,
preserving its contents and depending on sda and sdb
As long as the seeding devices are unmodified and available, they can
be used to start another branch.
RAID56 STATUS AND RECOMMENDED PRACTICES
The RAID56 feature provides striping and parity over several devices,
same as the traditional RAID5/6. There are some implementation and de-
sign deficiencies that make it unreliable for some corner cases and the
feature should not be used in production, only for evaluation or test-
ing. The power failure safety for metadata with RAID56 is not 100%.
Metadata
Do not use raid5 nor raid6 for metadata. Use raid1 or raid1c3 respec-
tively.
The substitute profiles provide the same guarantees against loss of 1
or 2 devices, and in some respect can be an improvement. Recovering
from one missing device will only need to access the remaining 1st or
2nd copy, that in general may be stored on some other devices due to
the way RAID1 works on btrfs, unlike on a striped profile (similar to
raid0) that would need all devices all the time.
The space allocation pattern and consumption is different (e.g. on N
devices): for raid5 as an example, a 1GiB chunk is reserved on each de-
vice, while with raid1 there's each 1GiB chunk stored on 2 devices. The
consumption of each 1GiB of used metadata is then N * 1GiB for vs 2 *
1GiB. Using raid1 is also more convenient for balancing/converting to
other profile due to lower requirement on the available chunk space.
Missing/incomplete support
When RAID56 is on the same filesystem with different raid profiles, the
space reporting is inaccurate, e.g. df, btrfs filesystem df or btrfs
filesystem usage. When there's only a one profile per block group type
(e.g. RAID5 for data) the reporting is accurate.
When scrub is started on a RAID56 filesystem, it's started on all de-
vices that degrade the performance. The workaround is to start it on
each device separately. Due to that the device stats may not match the
actual state and some errors might get reported multiple times.
The write hole problem. An unclean shutdown could leave a partially
written stripe in a state where the some stripe ranges and the parity
are from the old writes and some are new. The information which is
which is not tracked. Write journal is not implemented. Alternatively a
full read-modify-write would make sure that a full stripe is always
written, avoiding the write hole completely, but performance in that
case turned out to be too bad for use.
The striping happens on all available devices (at the time the chunks
were allocated), so in case a new device is added it may not be uti-
lized immediately and would require a rebalance. A fixed configured
stripe width is not implemented.
STORAGE MODEL, HARDWARE CONSIDERATIONS
Storage model
A storage model is a model that captures key physical aspects of data
structure in a data store. A filesystem is the logical structure orga-
nizing data on top of the storage device.
The filesystem assumes several features or limitations of the storage
device and utilizes them or applies measures to guarantee reliability.
BTRFS in particular is based on a COW (copy on write) mode of writing,
i.e. not updating data in place but rather writing a new copy to a dif-
ferent location and then atomically switching the pointers.
In an ideal world, the device does what it promises. The filesystem as-
sumes that this may not be true so additional mechanisms are applied to
either detect misbehaving hardware or get valid data by other means.
The devices may (and do) apply their own detection and repair mecha-
nisms but we won't assume any.
The following assumptions about storage devices are considered (sorted
by importance, numbers are for further reference):
1. atomicity of reads and writes of blocks/sectors (the smallest unit
of data the device presents to the upper layers)
2. there's a flush command that instructs the device to forcibly order
writes before and after the command; alternatively there's a barrier
command that facilitates the ordering but may not flush the data
3. data sent to write to a given device offset will be written without
further changes to the data and to the offset
4. writes can be reordered by the device, unless explicitly serialized
by the flush command
5. reads and writes can be freely reordered and interleaved
The consistency model of BTRFS builds on these assumptions. The logical
data updates are grouped, into a generation, written on the device, se-
rialized by the flush command and then the super block is written end-
ing the generation. All logical links among metadata comprising a con-
sistent view of the data may not cross the generation boundary.
When things go wrong
No or partial atomicity of block reads/writes (1)
• Problem: a partial block contents is written (torn write), e.g. due
to a power glitch or other electronics failure during the read/write
• Detection: checksum mismatch on read
• Repair: use another copy or rebuild from multiple blocks using some
encoding scheme
The flush command does not flush (2)
This is perhaps the most serious problem and impossible to mitigate by
filesystem without limitations and design restrictions. What could hap-
pen in the worst case is that writes from one generation bleed to an-
other one, while still letting the filesystem consider the generations
isolated. Crash at any point would leave data on the device in an in-
consistent state without any hint what exactly got written, what is
missing and leading to stale metadata link information.
Devices usually honor the flush command, but for performance reasons
may do internal caching, where the flushed data are not yet persis-
tently stored. A power failure could lead to a similar scenario as
above, although it's less likely that later writes would be written be-
fore the cached ones. This is beyond what a filesystem can take into
account. Devices or controllers are usually equipped with batteries or
capacitors to write the cache contents even after power is cut. (Bat-
tery backed write cache)
Data get silently changed on write (3)
Such thing should not happen frequently, but still can happen spuri-
ously due the complex internal workings of devices or physical effects
of the storage media itself.
• Problem: while the data are written atomically, the contents get
changed
• Detection: checksum mismatch on read
• Repair: use another copy or rebuild from multiple blocks using some
encoding scheme
Data get silently written to another offset (3)
This would be another serious problem as the filesystem has no informa-
tion when it happens. For that reason the measures have to be done
ahead of time. This problem is also commonly called ghost write.
The metadata blocks have the checksum embedded in the blocks, so a cor-
rect atomic write would not corrupt the checksum. It's likely that af-
ter reading such block the data inside would not be consistent with the
rest. To rule that out there's embedded block number in the metadata
block. It's the logical block number because this is what the logical
structure expects and verifies.
The following is based on information publicly available, user feed-
back, community discussions or bug report analyses. It's not complete
and further research is encouraged when in doubt.
Main memory
The data structures and raw data blocks are temporarily stored in com-
puter memory before they get written to the device. It is critical that
memory is reliable because even simple bit flips can have vast conse-
quences and lead to damaged structures, not only in the filesystem but
in the whole operating system.
Based on experience in the community, memory bit flips are more common
than one would think. When it happens, it's reported by the
tree-checker or by a checksum mismatch after reading blocks. There are
some very obvious instances of bit flips that happen, e.g. in an or-
dered sequence of keys in metadata blocks. We can easily infer from the
other data what values get damaged and how. However, fixing that is not
straightforward and would require cross-referencing data from the en-
tire filesystem to see the scope.
If available, ECC memory should lower the chances of bit flips, but
this type of memory is not available in all cases. A memory test should
be performed in case there's a visible bit flip pattern, though this
may not detect a faulty memory module because the actual load of the
system could be the factor making the problems appear. In recent years
attacks on how the memory modules operate have been demonstrated
(rowhammer) achieving specific bits to be flipped. While these were
targeted, this shows that a series of reads or writes can affect unre-
lated parts of memory.
Further reading:
• https://en.wikipedia.org/wiki/Row_hammer
What to do:
• run memtest, note that sometimes memory errors happen only when the
system is under heavy load that the default memtest cannot trigger
• memory errors may appear as filesystem going read-only due to "pre
write" check, that verify meta data before they get written but fail
some basic consistency checks
Direct memory access (DMA)
Another class of errors is related to DMA (direct memory access) per-
formed by device drivers. While this could be considered a software er-
ror, the data transfers that happen without CPU assistance may acciden-
tally corrupt other pages. Storage devices utilize DMA for performance
reasons, the filesystem structures and data pages are passed back and
forth, making errors possible in case page life time is not properly
tracked.
There are lots of quirks (device-specific workarounds) in Linux kernel
drivers (regarding not only DMA) that are added when found. The quirks
may avoid specific errors or disable some features to avoid worse prob-
lems.
What to do:
• use up-to-date kernel (recent releases or maintained long term sup-
port versions)
• as this may be caused by faulty drivers, keep the systems up-to-date
Rotational disks (HDD)
Rotational HDDs typically fail at the level of individual sectors or
small clusters. Read failures are caught on the levels below the
filesystem and are returned to the user as EIO - Input/output error.
Reading the blocks repeatedly may return the data eventually, but this
is better done by specialized tools and filesystem takes the result of
the lower layers. Rewriting the sectors may trigger internal remapping
but this inevitably leads to data loss.
Disk firmware is technically software but from the filesystem perspec-
tive is part of the hardware. IO requests are processed, and caching or
various other optimizations are performed, which may lead to bugs under
high load or unexpected physical conditions or unsupported use cases.
Disks are connected by cables with two ends, both of which can cause
problems when not attached properly. Data transfers are protected by
checksums and the lower layers try hard to transfer the data correctly
or not at all. The errors from badly-connecting cables may manifest as
large amount of failed read or write requests, or as short error bursts
depending on physical conditions.
What to do:
• check smartctl for potential issues
Solid state drives (SSD)
The mechanism of information storage is different from HDDs and this
affects the failure mode as well. The data are stored in cells grouped
in large blocks with limited number of resets and other write con-
straints. The firmware tries to avoid unnecessary resets and performs
optimizations to maximize the storage media lifetime. The known tech-
niques are deduplication (blocks with same fingerprint/hash are mapped
to same physical block), compression or internal remapping and garbage
collection of used memory cells. Due to the additional processing there
are measures to verity the data e.g. by ECC codes.
The observations of failing SSDs show that the whole electronic fails
at once or affects a lot of data (e.g. stored on one chip). Recovering
such data may need specialized equipment and reading data repeatedly
does not help as it's possible with HDDs.
There are several technologies of the memory cells with different char-
acteristics and price. The lifetime is directly affected by the type
and frequency of data written. Writing "too much" distinct data (e.g.
encrypted) may render the internal deduplication ineffective and lead
to a lot of rewrites and increased wear of the memory cells.
There are several technologies and manufacturers so it's hard to de-
scribe them but there are some that exhibit similar behaviour:
• expensive SSD will use more durable memory cells and is optimized for
reliability and high load
• cheap SSD is projected for a lower load ("desktop user") and is opti-
mized for cost, it may employ the optimizations and/or extended error
reporting partially or not at all
It's not possible to reliably determine the expected lifetime of an SSD
due to lack of information about how it works or due to lack of reli-
able stats provided by the device.
Metadata writes tend to be the biggest component of lifetime writes to
a SSD, so there is some value in reducing them. Depending on the device
class (high end/low end) the features like DUP block group profiles may
affect the reliability in both ways:
• high end are typically more reliable and using single for data and
metadata could be suitable to reduce device wear
• low end could lack ability to identify errors so an additional redun-
dancy at the filesystem level (checksums, DUP) could help
Only users who consume 50 to 100% of the SSD's actual lifetime writes
need to be concerned by the write amplification of btrfs DUP metadata.
Most users will be far below 50% of the actual lifetime, or will write
the drive to death and discover how many writes 100% of the actual
lifetime was. SSD firmware often adds its own write multipliers that
can be arbitrary and unpredictable and dependent on application behav-
ior, and these will typically have far greater effect on SSD lifespan
than DUP metadata. It's more or less impossible to predict when a SSD
will run out of lifetime writes to within a factor of two, so it's hard
to justify wear reduction as a benefit.
Further reading:
• https://www.snia.org/educational-library/ssd-and-deduplication-end-spinning-disk-2012
• https://www.snia.org/educational-library/realities-solid-state-storage-2013-2013
• https://www.snia.org/educational-library/ssd-performance-primer-2013
• https://www.snia.org/educational-library/how-controllers-maximize-ssd-life-2013
What to do:
• run smartctl or self-tests to look for potential issues
• keep the firmware up-to-date
NVM express, non-volatile memory (NVMe)
NVMe is a type of persistent memory usually connected over a system bus
(PCIe) or similar interface and the speeds are an order of magnitude
faster than SSD. It is also a non-rotating type of storage, and is not
typically connected by a cable. It's not a SCSI type device either but
rather a complete specification for logical device interface.
In a way the errors could be compared to a combination of SSD class and
regular memory. Errors may exhibit as random bit flips or IO failures.
There are tools to access the internal log (nvme log and nvme-cli) for
a more detailed analysis.
There are separate error detection and correction steps performed e.g.
on the bus level and in most cases never making in to the filesystem
level. Once this happens it could mean there's some systematic error
like overheating or bad physical connection of the device. You may want
to run self-tests (using smartctl).
• https://en.wikipedia.org/wiki/NVM_Express
• https://www.smartmontools.org/wiki/NVMe_Support
Drive firmware
Firmware is technically still software but embedded into the hardware.
As all software has bugs, so does firmware. Storage devices can update
the firmware and fix known bugs. In some cases the it's possible to
avoid certain bugs by quirks (device-specific workarounds) in Linux
kernel.
A faulty firmware can cause wide range of corruptions from small and
localized to large affecting lots of data. Self-repair capabilities may
not be sufficient.
What to do:
• check for firmware updates in case there are known problems, note
that updating firmware can be risky on itself
• use up-to-date kernel (recent releases or maintained long term sup-
port versions)
SD flash cards
There are a lot of devices with low power consumption and thus using
storage media based on low power consumption too, typically flash mem-
ory stored on a chip enclosed in a detachable card package. An improp-
erly inserted card may be damaged by electrical spikes when the device
is turned on or off. The chips storing data in turn may be damaged per-
manently. All types of flash memory have a limited number of rewrites,
so the data are internally translated by FTL (flash translation layer).
This is implemented in firmware (technically a software) and prone to
bugs that manifest as hardware errors.
Adding redundancy like using DUP profiles for both data and metadata
can help in some cases but a full backup might be the best option once
problems appear and replacing the card could be required as well.
Hardware as the main source of filesystem corruptions
If you use unreliable hardware and don't know about that, don't blame
the filesystem when it tells you.
SEE ALSO
acl(5), btrfs(8), chattr(1), fstrim(8), ioctl(2), mkfs.btrfs(8),
mount(8), swapon(8)
6.6.3 Mar 31, 2024 BTRFS(5)
Generated by dwww version 1.16 on Tue Dec 16 06:27:59 CET 2025.