Event counters are used in bees to collect simple branch-coverage statistics. Every time bees makes a decision, it increments an event counter, so there are many event counters.
Events are grouped by prefix in their event names, e.g. block
is block
I/O, dedup
is deduplication requests, tmp
is temporary files, etc.
Events with the suffix _ms
count total milliseconds spent performing
the operation. These are counted separately for each thread, so there
can be more than 1000 ms per second.
There is considerable overlap between some events, e.g. example_try
denotes an event that is counted when an action is attempted,
example_hit
is counted when the attempt succeeds and has a desired
outcome, and example_miss
is counted when the attempt succeeds but
the desired outcome is not achieved. In most cases example_try = example_hit + example_miss + (
example failed and threw an exception)
,
but some event groups defy such simplistic equations.
The addr
event group consists of operations related to translating (root, inode, offset)
tuples (i.e. logical position within a file) into btrfs
virtual block addresses (i.e. physical position on disk).
addr_block
: The address of a block was computed.addr_compressed
: Obsolete implementation ofaddr_compressed_offset
.addr_compressed_offset
: The address of a compressed block was computed.addr_delalloc
: The address of a block could not be computed due to delayed allocation. Only possible when using obsoleteFIEMAP
code.addr_eof_e
: The address of a block at EOF that was not block-aligned was computed.addr_from_fd
: The address of a block was computed using afd
(open to the file in question) andoffset
pair.addr_from_root_fd
: The address of a block was computed using the filesystem rootfd
instead of the open filefd
for theTREE_SEARCH_V2
ioctl. This is obsolete and should probably be removed at some point.addr_hole
: The address of a block in a hole was computed.addr_magic
: The address of a block cannot be determined in a way that bees can use (unrecognized flags or flags known to be incompatible with bees).addr_uncompressed
: The address of an uncompressed block was computed.addr_unrecognized
: The address of a block with unrecognized flags (i.e. kernel version newer than bees) was computed.addr_unusable
: The address of a block with unusable flags (i.e. flags that are known to be incompatible with bees) was computed.
The adjust
event group consists of operations related to translating stored virtual block addresses (i.e. physical position on disk) to (root, inode, offset)
tuples (i.e. logical positions within files). BeesResolver::adjust_offset
determines if a single candidate reference from the LOGICAL_INO
ioctl corresponds to the requested btrfs virtual block address.
adjust_compressed_offset_correct
: A block address corresponding to a compressed block was retrieved from the hash table and resolved to a physical block containing data that matches another block bees has already read.adjust_compressed_offset_wrong
: A block address corresponding to a compressed block was retrieved from the hash table and resolved to a physical block containing data that matches the hash but not the data from another block bees has already read (i.e. there was a hash collision).adjust_eof_fail
: A block address corresponding to a block at EOF that was not aligned to a block boundary matched another block bees already read, but the length of the unaligned data in both blocks was not equal. This is usually caused by stale entries in the hash table pointing to blocks that have been overwritten since the hash table entries were created. It can also be caused by hash collisions, but hashes are not yet computed at this point in the code, so this event does not correlate to thehash_collision
counter.adjust_eof_haystack
: A block address from the hash table corresponding to a block at EOF that was not aligned to a block boundary was processed.adjust_eof_hit
: A block address corresponding to a block at EOF that was not aligned to a block boundary matched a similarly unaligned block that bees already read.adjust_eof_miss
: A block address from the hash table corresponding to a block at EOF that was not aligned to a block boundary did not match a similarly unaligned block that bees already read.adjust_eof_needle
: A block address from scanning the disk corresponding to a block at EOF that was not aligned to a block boundary was processed.adjust_exact
: A block address from the hash table corresponding to an uncompressed data block was processed to find its(root, inode, offset)
references.adjust_exact_correct
: A block address corresponding to an uncompressed block was retrieved from the hash table and resolved to a physical block containing data that matches another block bees has already read.adjust_exact_wrong
: A block address corresponding to an uncompressed block was retrieved from the hash table and resolved to a physical block containing data that matches the hash but not the data from another block bees has already read (i.e. there was a hash collision).adjust_hit
: A block address was retrieved from the hash table and resolved to a physical block in an uncompressed extent containing data that matches the data from another block bees has already read (i.e. a duplicate match was found).adjust_miss
: A block address was retrieved from the hash table and resolved to a physical block containing a hash that does not match the hash from another block bees has already read (i.e. the hash table contained a stale entry and the data it referred to has since been overwritten in the filesystem).adjust_needle_too_long
: A block address was retrieved from the hash table, but when the corresponding extent item was retrieved, its offset or length were out of range to be a match (i.e. the hash table contained a stale entry and the data it referred to has since been overwritten in the filesystem).adjust_no_match
: A hash collision occurred (i.e. a block on disk was located with the same hash as the hash table entry but different data) . Effectively an alias forhash_collision
as it is not possible to have one event without the other.adjust_offset_high
: TheLOGICAL_INO
ioctl gave an extent item that does not overlap with the desired block because the extent item ends before the desired block in the extent data.adjust_offset_hit
: A block address was retrieved from the hash table and resolved to a physical block in a compressed extent containing data that matches the data from another block bees has already read (i.e. a duplicate match was found).adjust_offset_low
: TheLOGICAL_INO
ioctl gave an extent item that does not overlap with the desired block because the extent item begins after the desired block in the extent data.adjust_try
: A block address and extent item candidate were passed toBeesResolver::adjust_offset
for processing.
The block
event group consists of operations related to reading data blocks from the filesystem.
block_bytes
: Number of data bytes read.block_hash
: Number of block hashes computed.block_ms
: Total time reading data blocks.block_read
: Number of data blocks read.block_zero
: Number of data blocks read with zero contents (i.e. candidates for replacement with a hole).
The bug
event group consists of known bugs in bees.
bug_bad_max_transid
: A badmax_transid
was found and removed inbeescrawl.dat
.bug_bad_min_transid
: A badmin_transid
was found and removed inbeescrawl.dat
.bug_dedup_same_physical
:BeesContext::dedup
detected that the physical extent was the same forsrc
anddst
. This has no effect on space usage so it is a waste of time, and also carries the risk of creating a toxic extent.bug_grow_pair_overlaps
: Two identical blocks were found, and while searching matching adjacent extents, the potentialsrc
grew to overlap the potentialdst
. This would create a cycle where bees keeps trying to eliminate blocks but instead just moves them around.bug_hash_duplicate_cell
: Two entries in the hash table were identical. This only happens due to data corruption or a bug.bug_hash_magic_addr
: An entry in the hash table contains an address with magic. Magic addresses cannot be deduplicated so they should not be stored in the hash table.
The chase
event group consists of operations connecting btrfs virtual block addresses with (root, inode, offset)
tuples. resolve
is the top level, adjust
is the bottom level, and chase
is the middle level. BeesResolver::chase_extent_ref
iterates over (root, inode, offset)
tuples from LOGICAL_INO
and attempts to find a single matching block in the filesystem given a candidate block from an earlier scan
operation.
chase_corrected
: A matching block was resolved to a(root, inode, offset)
tuple, but the offset of a block matching data did not match the offset given byLOGICAL_INO
.chase_hit
: A block address was successfully and correctly translated to a(root, inode, offset)
tuple.chase_no_data
: A block address was not successfully translated to a(root, inode, offset)
tuple.chase_no_fd
: A(root, inode)
tuple could not be opened (i.e. the file was deleted on the filesystem).chase_try
: A block address translation attempt started.chase_uncorrected
: A matching block was resolved to a(root, inode, offset)
tuple, and the offset of a block matching data did match the offset given byLOGICAL_INO
.chase_wrong_addr
: The btrfs virtual address (i.e. physical block address) found at a candidate(root, inode, offset)
tuple did not match the expected btrfs virtual address (i.e. the filesystem was modified during the resolve operation).chase_wrong_magic
: The extent item at a candidate(root, inode, offset)
tuple has magic bits and cannot match any btrfs virtual address in the hash table (i.e. the filesystem was modified during the resolve operation).
The crawl
event group consists of operations related to scanning btrfs trees to find new extent refs to scan for dedupe.
crawl_again
: An inode crawl was restarted because the extent was already locked by another running crawl.crawl_blacklisted
: An extent was not scanned because it belongs to a blacklisted file.crawl_create
: A new subvol crawler was created.crawl_done
: One pass over all subvols on the filesystem was completed.crawl_empty
: ATREE_SEARCH_V2
ioctl call failed or returned an empty set (usually because all data in the subvol was scanned).crawl_fail
: ATREE_SEARCH_V2
ioctl call failed.crawl_gen_high
: An extent item in the search results refers to an extent that is newer than the current crawl'smax_transid
allows.crawl_gen_low
: An extent item in the search results refers to an extent that is older than the current crawl'smin_transid
allows.crawl_hole
: An extent item in the search results refers to a hole.crawl_inline
: An extent item in the search results contains an inline extent.crawl_items
: An item in theTREE_SEARCH_V2
data was processed.crawl_ms
: Time spent running theTREE_SEARCH_V2
ioctl.crawl_no_empty
: Attempted to delete the last crawler. Should never happen.crawl_nondata
: An item in the search results is not data.crawl_prealloc
: An extent item in the search results refers to aPREALLOC
extent.crawl_push
: An extent item in the search results is suitable for scanning and deduplication.crawl_scan
: An extent item in the search results is submitted toBeesContext::scan_forward
for scanning and deduplication.crawl_search
: ATREE_SEARCH_V2
ioctl call was successful.crawl_unknown
: An extent item in the search results has an unrecognized type.
The dedup
(sic) event group consists of operations that deduplicate data.
dedup_bytes
: Total bytes in extent references deduplicated.dedup_copy
: Total bytes copied to eliminate unique data in extents containing a mix of unique and duplicate data.dedup_hit
: Total number of pairs of identical extent references.dedup_miss
: Total number of pairs of non-identical extent references.dedup_ms
: Total time spent running theFILE_EXTENT_SAME
(akaFI_DEDUPERANGE
ordedupe_file_range
) ioctl.dedup_prealloc_bytes
: Total bytes in eliminatedPREALLOC
extent references.dedup_prealloc_hit
: Total number of successfully eliminatedPREALLOC
extent references.dedup_prealloc_hit
: Total number of unsuccessfully eliminatedPREALLOC
extent references (i.e. filesystem data changed between scan and dedupe).dedup_try
: Total number of pairs of extent references submitted for deduplication.dedup_workaround_btrfs_send
: Total number of extent reference pairs submitted for deduplication that were discarded to workaroundbtrfs send
bugs.
The exception
event group consists of C++ exceptions. C++ exceptions are thrown due to IO errors and internal constraint check failures.
exception_caught
: Total number of C++ exceptions thrown and caught by a generic exception handler.exception_caught_silent
: Total number of "silent" C++ exceptions thrown and caught by a generic exception handler. These are exceptions which are part of the correct and normal operation of bees. The exceptions are logged at a lower log level.
The hash
event group consists of operations related to the bees hash table.
hash_already
: A(hash, address)
pair was already present in the hash table during aBeesHashTable::push_random_hash_addr
operation.hash_bump
: An existing(hash, address)
pair was moved forward in the hash table by aBeesHashTable::push_random_hash_addr
operation.hash_collision
: A pair of data blocks was found with identical hashes but different data.hash_erase
: A(hash, address)
pair in the hash table was removed because a matching data block could not be found in the filesystem (i.e. the hash table entry is out of date).hash_erase_miss
: A(hash, address)
pair was reported missing from the filesystem but no such entry was found in the hash table (i.e. race between scanning threads or pair already evicted).hash_evict
: A(hash, address)
pair was evicted from the hash table to accommodate a new hash table entry.hash_extent_in
: A hash table extent was read.hash_extent_out
: A hash table extent was written.hash_front
: A(hash, address)
pair was pushed to the front of the list because it matched a duplicate block.hash_front_already
: A(hash, address)
pair was pushed to the front of the list because it matched a duplicate block, but the pair was already at the front of the list so no change occurred.hash_insert
: A(hash, address)
pair was inserted byBeesHashTable::push_random_hash_addr
.hash_lookup
: The hash table was searched for(hash, address)
pairs matching a givenhash
.
The inserted
event group consists of operations related to storing hash and address data in the hash table (i.e. the hash table client).
inserted_block
: Total number of data block references scanned and inserted into the hash table.inserted_clobbered
: Total number of data block references scanned and eliminated from the filesystem.
The matched
event group consists of events related to matching incoming data blocks against existing hash table entries.
matched_0
: A data block was scanned, hash table entries found, but no matching data blocks on the filesytem located.matched_1_or_more
: A data block was scanned, hash table entries found, and one or more matching data blocks on the filesystem located.matched_2_or_more
: A data block was scanned, hash table entries found, and two or more matching data blocks on the filesystem located.matched_3_or_more
: A data block was scanned, hash table entries found, and three or more matching data blocks on the filesystem located.
The open
event group consists of operations related to translating (root, inode)
tuples into open file descriptors (i.e. open_by_handle
emulation for btrfs).
open_clear
: The open FD cache was cleared to avoid keeping file descriptors open too long.open_fail_enoent
: A file could not be opened because it no longer exists (i.e. it was deleted or renamed during the lookup/resolve operations).open_fail_error
: A file could not be opened for other reasons (e.g. IO error, permission denied, out of resources).open_file
: A file was successfully opened. This counts only theopen()
system call, not other reasons why the opened FD might not be usable.open_hit
: A file was successfully opened and the FD was acceptable.open_ino_ms
: Total time spent executing theopen()
system call.open_lookup_empty
: No paths were found for the inode in theINO_PATHS
ioctl.open_lookup_enoent
: TheINO_PATHS
ioctl returned ENOENT.open_lookup_error
: TheINO_PATHS
ioctl returned a different error.open_lookup_ok
: TheINO_PATHS
ioctl successfully returned a list of one or more filenames.open_no_path
: All attempts to open a file by(root, inode)
pair failed.open_no_root
: An attempt to open a file by(root, inode)
pair failed because theroot
could not be opened.open_root_ms
: Total time spent opening subvol root FDs.open_wrong_dev
: A FD returned byopen()
did not match the device belonging to the filesystem subvol.open_wrong_flags
: A FD returned byopen()
had incompatible flags (NODATASUM
/NODATACOW
).open_wrong_ino
: A FD returned byopen()
did not match the expected inode (i.e. the file was renamed or replaced during the lookup/resolve operations).open_wrong_root
: A FD returned byopen()
did not match the expected subvol ID (i.e.root
).
The pairbackward
event group consists of events related to extending matching block ranges backward starting from the initial block match found using the hash table.
pairbackward_bof_first
: A matching pair of block ranges could not be extended backward because the beginning of the first (src) file was reached.pairbackward_bof_second
: A matching pair of block ranges could not be extended backward because the beginning of the second (dst) file was reached.pairbackward_hit
: A pair of matching block ranges was extended backward by one block.pairbackward_miss
: A pair of matching block ranges could not be extended backward by one block because the pair of blocks before the first block in the range did not contain identical data.pairbackward_ms
: Total time spent extending matching block ranges backward from the first matching block found by hash table lookup.pairbackward_overlap
: A pair of matching block ranges could not be extended backward by one block because this would cause the two block ranges to overlap.pairbackward_same
: A pair of matching block ranges could not be extended backward by one block because this would cause the two block ranges to refer to the same btrfs data extent.pairbackward_stop
: Stopped extending a pair of matching block ranges backward for any of the reasons listed here.pairbackward_toxic_addr
: A pair of matching block ranges was abandoned because the extended range would include a data block with a toxic address.pairbackward_toxic_hash
: A pair of matching block ranges was abandoned because the extended range would include a data block with a toxic hash.pairbackward_try
: Started extending a pair of matching block ranges backward.pairbackward_zero
: A pair of matching block ranges could not be extended backward by one block because the src block contained all zeros and was not compressed.
The pairforward
event group consists of events related to extending matching block ranges forward starting from the initial block match found using the hash table.
pairforward_eof_first
: A matching pair of block ranges could not be extended forward because the end of the first (src) file was reached.pairforward_eof_malign
: A matching pair of block ranges could not be extended forward because the end of the second (dst) file was not aligned to a 4K boundary nor the end of the first (src) file.pairforward_eof_second
: A matching pair of block ranges could not be extended forward because the end of the second (dst) file was reached.pairforward_hit
: A pair of matching block ranges was extended forward by one block.pairforward_hole
: A pair of matching block ranges was extended forward by one block, and the block was a hole in the second (dst) file.pairforward_miss
: A pair of matching block ranges could not be extended forward by one block because the pair of blocks after the last block in the range did not contain identical data.pairforward_ms
: Total time spent extending matching block ranges forward from the first matching block found by hash table lookup.pairforward_overlap
: A pair of matching block ranges could not be extended forward by one block because this would cause the two block ranges to overlap.pairforward_same
: A pair of matching block ranges could not be extended forward by one block because this would cause the two block ranges to refer to the same btrfs data extent.pairforward_stop
: Stopped extending a pair of matching block ranges forward for any of the reasons listed here.pairforward_toxic_addr
: A pair of matching block ranges was abandoned because the extended range would include a data block with a toxic address.pairforward_toxic_hash
: A pair of matching block ranges was abandoned because the extended range would include a data block with a toxic hash.pairforward_try
: Started extending a pair of matching block ranges forward.pairforward_zero
: A pair of matching block ranges could not be extended backward by one block because the src block contained all zeros and was not compressed.
The readahead
event group consists of events related to calls to posix_fadvise
.
readahead_ms
: Total time spent runningposix_fadvise(..., POSIX_FADV_WILLNEED)
akareadahead()
.readahead_unread_ms
: Total time spent runningposix_fadvise(..., POSIX_FADV_DONTNEED)
.
The replacedst
event group consists of events related to replacing a single reference to a dst extent using any suitable src extent (i.e. eliminating a single duplicate extent ref during a crawl).
replacedst_dedup_hit
: A duplicate extent reference was identified and removed.replacedst_dedup_miss
: A duplicate extent reference was identified, but src and dst extents did not match (i.e. the filesystem changed in the meantime).replacedst_grown
: A duplicate block was identified, and adjacent blocks were duplicate as well.replacedst_overlaps
: A pair of duplicate block ranges was identified, but the pair was not usable for dedupe because the two ranges overlap.replacedst_same
: A pair of duplicate block ranges was identified, but the pair was not usable for dedupe because the physical block ranges were the same.replacedst_try
: A duplicate block was identified and an attempt was made to remove it (i.e. this is the total number of replacedst calls).
The replacesrc
event group consists of events related to replacing every reference to a src extent using a temporary copy of the extent's data (i.e. eliminating leftover unique data in a partially duplicate extent during a crawl).
replacesrc_dedup_hit
: A duplicate extent reference was identified and removed.replacesrc_dedup_miss
: A duplicate extent reference was identified, but src and dst extents did not match (i.e. the filesystem changed in the meantime).replacesrc_grown
: A duplicate block was identified, and adjacent blocks were duplicate as well.replacesrc_overlaps
: A pair of duplicate block ranges was identified, but the pair was not usable for dedupe because the two ranges overlap.replacesrc_try
: A duplicate block was identified and an attempt was made to remove it (i.e. this is the total number of replacedst calls).
The resolve
event group consists of operations related to translating a btrfs virtual block address (i.e. physical block address) to a (root, inode, offset)
tuple (i.e. locating and opening the file containing a matching block). resolve
is the top level, chase
and adjust
are the lower two levels.
resolve_empty
: TheLOGICAL_INO
ioctl returned successfully with an empty reference list (0 items).resolve_fail
: TheLOGICAL_INO
ioctl returned an error.resolve_large
: TheLOGICAL_INO
ioctl returned more than 2730 results (the limit of the v1 ioctl).resolve_ms
: Total time spent in theLOGICAL_INO
ioctl (i.e. wallclock time, not kernel CPU time).resolve_ok
: TheLOGICAL_INO
ioctl returned success.resolve_overflow
: TheLOGICAL_INO
ioctl returned more than 655050 extents (the limit of the v2 ioctl).resolve_toxic
: TheLOGICAL_INO
ioctl took more than 0.1 seconds of kernel CPU time.
The root
event group consists of operations related to translating a btrfs root ID (i.e. subvol ID) into an open file descriptor by navigating the btrfs root tree.
root_clear
: The root FD cache was cleared.root_found
: A root FD was successfully opened.root_notfound
: A root FD could not be opened because all candidate paths could not be opened, or there were no paths available.root_ok
: A root FD was opened and its correctness verified.root_open_fail
: A root FDopen()
attempt returned an error.root_parent_open_fail
: A recursive call to open the parent of a subvol failed.root_parent_open_ok
: A recursive call to open the parent of a subvol succeeded.root_parent_open_try
: A recursive call to open the parent of a subvol was attempted.root_parent_path_empty
: No path could be found to connect a parent root FD to its child.root_parent_path_fail
: TheINO_PATH
ioctl failed to find a name for a child subvol relative to its parent.root_parent_path_open_fail
: Theopen()
call in a recursive call to open the parent of a subvol returned an error.root_workaround_btrfs_send
: A subvol was determined to be read-only and disabled to implement the btrfs send workaround.
The scan
event group consists of operations related to scanning incoming data. This is where bees finds duplicate data and populates the hash table.
scan_blacklisted
: A blacklisted extent was passed toscan_forward
and dropped.scan_block
: A block of data was scanned.scan_bump
: After deduping a block range, the scan pointer had to be moved past the end of the deduped byte range.scan_dup_block
: Number of duplicate blocks deduped.scan_dup_hit
: A pair of duplicate block ranges was found and removed.scan_dup_miss
: A pair of duplicate blocks was found in the hash table but not in the filesystem.scan_eof
: Scan past EOF was attempted.scan_erase_redundant
: Blocks in the hash table were removed because they were removed from the filesystem by dedupe.scan_extent
: An extent was scanned (scan_one_extent
).scan_extent_tiny
: An extent below 128K that was not the beginning or end of a file was scanned. No action is currently taken for these--they are merely counted.scan_forward
: A logical byte range was scanned (scan_forward
).scan_found
: An entry was found in the hash table matching a scanned block from the filesystem.scan_hash_hit
: A block was found on the filesystem corresponding to a block found in the hash table.scan_hash_miss
: A block was not found on the filesystem corresponding to a block found in the hash table.scan_hash_preinsert
: A block was prepared for insertion into the hash table.scan_hole
: A hole extent was found during scan and ignored.scan_interesting
: An extent had flags that were not recognized by bees and was ignored.scan_lookup
: A hash was looked up in the hash table.scan_malign
: A block being scanned matched a hash at EOF in the hash table, but the EOF was not aligned to a block boundary and the two blocks did not have the same length.scan_no_fd
: References to a block from the hash table were found, but a FD could not be opened.scan_no_rewrite
: All blocks in an extent were removed by dedupe (i.e. no copies).scan_push_front
: An entry in the hash table matched a duplicate block, so the entry was moved to the head of its LRU list.scan_reinsert
: A copied block's hash and block address was inserted into the hash table.scan_resolve_hit
: A block address in the hash table was successfully resolved to an open FD and offset pair.scan_resolve_zero
: A block address in the hash table was not resolved to any subvol/inode pair, so the corresponding hash table entry was removed.scan_rewrite
: A range of bytes in a file was copied, then the copy deduped over the original data.scan_toxic_hash
: A scanned block has the same hash as a hash table entry that is marked toxic.scan_toxic_match
: A hash table entry points to a block that is discovered to be toxic.scan_twice
: Two references to the same block have been found in the hash table.scan_zero_compressed
: An extent that was compressed and contained only zero bytes was found.scan_zero_uncompressed
: A block that contained only zero bytes was found in an uncompressed extent.
The scanf
event group consists of operations related to BeesContext::scan_forward
. This is the entry point where crawl
schedules new data for scanning.
scanf_deferred_extent
: Two tasks attempted to scan the same extent at the same time, so one was deferred.scanf_deferred_inode
: Two tasks attempted to scan the same inode at the same time, so one was deferred.scanf_extent
: A btrfs extent item was scanned.scanf_extent_ms
: Total thread-seconds spent scanning btrfs extent items.scanf_total
: A logical byte range of a file was scanned.scanf_total_ms
: Total thread-seconds spent scanning logical byte ranges.
Note that in current versions of bees, scan_forward
is passed extents
that correspond exactly to btrfs extent items, so the scanf_extent
and
scanf_total
numbers can only be different if the filesystem changes
between crawl time and scan time.
The sync
event group consists of operations related to the fsync
workarounds in bees.
sync_count
:fsync()
was called on a temporary file.sync_ms
: Total time spent executingfsync()
.
The sync
event group consists of operations related temporary files and the data within them.
tmp_aligned
: A temporary extent was allocated on a block boundary.tmp_block
: Total number of temporary blocks copied.tmp_block_zero
: Total number of temporary hole blocks copied.tmp_bytes
: Total number of temporary bytes copied.tmp_copy
: Total number of extents copied.tmp_copy_ms
: Total time spent copying extents.tmp_create
: Total number of temporary files created.tmp_create_ms
: Total time spent creating temporary files.tmp_hole
: Total number of hole extents created.tmp_realign
: A temporary extent was not aligned to a block boundary.tmp_resize
: A temporary file was resized withftruncate()
tmp_resize_ms
: Total time spent inftruncate()
tmp_trunc
: The temporary file size limit was exceeded, triggering a new temporary file creation.