mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-10-31 08:44:41 +00:00 
			
		
		
		
	 ccb8c3bd6d
			
		
	
	
		ccb8c3bd6d
		
	
	
	
	
		
			
			The cache_readaheads stat counter is not used anymore and should be removed. Signed-off-by: Andrea Tomassetti <andrea.tomassetti-opensource@devo.com> Signed-off-by: Coly Li <colyli@suse.de> Link: https://lore.kernel.org/r/20230615121223.22502-4-colyli@suse.de Signed-off-by: Jens Axboe <axboe@kernel.dk>
		
			
				
	
	
		
			653 lines
		
	
	
	
		
			24 KiB
		
	
	
	
		
			ReStructuredText
		
	
	
	
	
	
			
		
		
	
	
			653 lines
		
	
	
	
		
			24 KiB
		
	
	
	
		
			ReStructuredText
		
	
	
	
	
	
| ============================
 | |
| A block layer cache (bcache)
 | |
| ============================
 | |
| 
 | |
| Say you've got a big slow raid 6, and an ssd or three. Wouldn't it be
 | |
| nice if you could use them as cache... Hence bcache.
 | |
| 
 | |
| The bcache wiki can be found at:
 | |
|   https://bcache.evilpiepirate.org
 | |
| 
 | |
| This is the git repository of bcache-tools:
 | |
|   https://git.kernel.org/pub/scm/linux/kernel/git/colyli/bcache-tools.git/
 | |
| 
 | |
| The latest bcache kernel code can be found from mainline Linux kernel:
 | |
|   https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/
 | |
| 
 | |
| It's designed around the performance characteristics of SSDs - it only allocates
 | |
| in erase block sized buckets, and it uses a hybrid btree/log to track cached
 | |
| extents (which can be anywhere from a single sector to the bucket size). It's
 | |
| designed to avoid random writes at all costs; it fills up an erase block
 | |
| sequentially, then issues a discard before reusing it.
 | |
| 
 | |
| Both writethrough and writeback caching are supported. Writeback defaults to
 | |
| off, but can be switched on and off arbitrarily at runtime. Bcache goes to
 | |
| great lengths to protect your data - it reliably handles unclean shutdown. (It
 | |
| doesn't even have a notion of a clean shutdown; bcache simply doesn't return
 | |
| writes as completed until they're on stable storage).
 | |
| 
 | |
| Writeback caching can use most of the cache for buffering writes - writing
 | |
| dirty data to the backing device is always done sequentially, scanning from the
 | |
| start to the end of the index.
 | |
| 
 | |
| Since random IO is what SSDs excel at, there generally won't be much benefit
 | |
| to caching large sequential IO. Bcache detects sequential IO and skips it;
 | |
| it also keeps a rolling average of the IO sizes per task, and as long as the
 | |
| average is above the cutoff it will skip all IO from that task - instead of
 | |
| caching the first 512k after every seek. Backups and large file copies should
 | |
| thus entirely bypass the cache.
 | |
| 
 | |
| In the event of a data IO error on the flash it will try to recover by reading
 | |
| from disk or invalidating cache entries.  For unrecoverable errors (meta data
 | |
| or dirty data), caching is automatically disabled; if dirty data was present
 | |
| in the cache it first disables writeback caching and waits for all dirty data
 | |
| to be flushed.
 | |
| 
 | |
| Getting started:
 | |
| You'll need bcache util from the bcache-tools repository. Both the cache device
 | |
| and backing device must be formatted before use::
 | |
| 
 | |
|   bcache make -B /dev/sdb
 | |
|   bcache make -C /dev/sdc
 | |
| 
 | |
| `bcache make` has the ability to format multiple devices at the same time - if
 | |
| you format your backing devices and cache device at the same time, you won't
 | |
| have to manually attach::
 | |
| 
 | |
|   bcache make -B /dev/sda /dev/sdb -C /dev/sdc
 | |
| 
 | |
| If your bcache-tools is not updated to latest version and does not have the
 | |
| unified `bcache` utility, you may use the legacy `make-bcache` utility to format
 | |
| bcache device with same -B and -C parameters.
 | |
| 
 | |
| bcache-tools now ships udev rules, and bcache devices are known to the kernel
 | |
| immediately.  Without udev, you can manually register devices like this::
 | |
| 
 | |
|   echo /dev/sdb > /sys/fs/bcache/register
 | |
|   echo /dev/sdc > /sys/fs/bcache/register
 | |
| 
 | |
| Registering the backing device makes the bcache device show up in /dev; you can
 | |
| now format it and use it as normal. But the first time using a new bcache
 | |
| device, it'll be running in passthrough mode until you attach it to a cache.
 | |
| If you are thinking about using bcache later, it is recommended to setup all your
 | |
| slow devices as bcache backing devices without a cache, and you can choose to add
 | |
| a caching device later.
 | |
| See 'ATTACHING' section below.
 | |
| 
 | |
| The devices show up as::
 | |
| 
 | |
|   /dev/bcache<N>
 | |
| 
 | |
| As well as (with udev)::
 | |
| 
 | |
|   /dev/bcache/by-uuid/<uuid>
 | |
|   /dev/bcache/by-label/<label>
 | |
| 
 | |
| To get started::
 | |
| 
 | |
|   mkfs.ext4 /dev/bcache0
 | |
|   mount /dev/bcache0 /mnt
 | |
| 
 | |
| You can control bcache devices through sysfs at /sys/block/bcache<N>/bcache .
 | |
| You can also control them through /sys/fs//bcache/<cset-uuid>/ .
 | |
| 
 | |
| Cache devices are managed as sets; multiple caches per set isn't supported yet
 | |
| but will allow for mirroring of metadata and dirty data in the future. Your new
 | |
| cache set shows up as /sys/fs/bcache/<UUID>
 | |
| 
 | |
| Attaching
 | |
| ---------
 | |
| 
 | |
| After your cache device and backing device are registered, the backing device
 | |
| must be attached to your cache set to enable caching. Attaching a backing
 | |
| device to a cache set is done thusly, with the UUID of the cache set in
 | |
| /sys/fs/bcache::
 | |
| 
 | |
|   echo <CSET-UUID> > /sys/block/bcache0/bcache/attach
 | |
| 
 | |
| This only has to be done once. The next time you reboot, just reregister all
 | |
| your bcache devices. If a backing device has data in a cache somewhere, the
 | |
| /dev/bcache<N> device won't be created until the cache shows up - particularly
 | |
| important if you have writeback caching turned on.
 | |
| 
 | |
| If you're booting up and your cache device is gone and never coming back, you
 | |
| can force run the backing device::
 | |
| 
 | |
|   echo 1 > /sys/block/sdb/bcache/running
 | |
| 
 | |
| (You need to use /sys/block/sdb (or whatever your backing device is called), not
 | |
| /sys/block/bcache0, because bcache0 doesn't exist yet. If you're using a
 | |
| partition, the bcache directory would be at /sys/block/sdb/sdb2/bcache)
 | |
| 
 | |
| The backing device will still use that cache set if it shows up in the future,
 | |
| but all the cached data will be invalidated. If there was dirty data in the
 | |
| cache, don't expect the filesystem to be recoverable - you will have massive
 | |
| filesystem corruption, though ext4's fsck does work miracles.
 | |
| 
 | |
| Error Handling
 | |
| --------------
 | |
| 
 | |
| Bcache tries to transparently handle IO errors to/from the cache device without
 | |
| affecting normal operation; if it sees too many errors (the threshold is
 | |
| configurable, and defaults to 0) it shuts down the cache device and switches all
 | |
| the backing devices to passthrough mode.
 | |
| 
 | |
|  - For reads from the cache, if they error we just retry the read from the
 | |
|    backing device.
 | |
| 
 | |
|  - For writethrough writes, if the write to the cache errors we just switch to
 | |
|    invalidating the data at that lba in the cache (i.e. the same thing we do for
 | |
|    a write that bypasses the cache)
 | |
| 
 | |
|  - For writeback writes, we currently pass that error back up to the
 | |
|    filesystem/userspace. This could be improved - we could retry it as a write
 | |
|    that skips the cache so we don't have to error the write.
 | |
| 
 | |
|  - When we detach, we first try to flush any dirty data (if we were running in
 | |
|    writeback mode). It currently doesn't do anything intelligent if it fails to
 | |
|    read some of the dirty data, though.
 | |
| 
 | |
| 
 | |
| Howto/cookbook
 | |
| --------------
 | |
| 
 | |
| A) Starting a bcache with a missing caching device
 | |
| 
 | |
| If registering the backing device doesn't help, it's already there, you just need
 | |
| to force it to run without the cache::
 | |
| 
 | |
| 	host:~# echo /dev/sdb1 > /sys/fs/bcache/register
 | |
| 	[  119.844831] bcache: register_bcache() error opening /dev/sdb1: device already registered
 | |
| 
 | |
| Next, you try to register your caching device if it's present. However
 | |
| if it's absent, or registration fails for some reason, you can still
 | |
| start your bcache without its cache, like so::
 | |
| 
 | |
| 	host:/sys/block/sdb/sdb1/bcache# echo 1 > running
 | |
| 
 | |
| Note that this may cause data loss if you were running in writeback mode.
 | |
| 
 | |
| 
 | |
| B) Bcache does not find its cache::
 | |
| 
 | |
| 	host:/sys/block/md5/bcache# echo 0226553a-37cf-41d5-b3ce-8b1e944543a8 > attach
 | |
| 	[ 1933.455082] bcache: bch_cached_dev_attach() Couldn't find uuid for md5 in set
 | |
| 	[ 1933.478179] bcache: __cached_dev_store() Can't attach 0226553a-37cf-41d5-b3ce-8b1e944543a8
 | |
| 	[ 1933.478179] : cache set not found
 | |
| 
 | |
| In this case, the caching device was simply not registered at boot
 | |
| or disappeared and came back, and needs to be (re-)registered::
 | |
| 
 | |
| 	host:/sys/block/md5/bcache# echo /dev/sdh2 > /sys/fs/bcache/register
 | |
| 
 | |
| 
 | |
| C) Corrupt bcache crashes the kernel at device registration time:
 | |
| 
 | |
| This should never happen.  If it does happen, then you have found a bug!
 | |
| Please report it to the bcache development list: linux-bcache@vger.kernel.org
 | |
| 
 | |
| Be sure to provide as much information that you can including kernel dmesg
 | |
| output if available so that we may assist.
 | |
| 
 | |
| 
 | |
| D) Recovering data without bcache:
 | |
| 
 | |
| If bcache is not available in the kernel, a filesystem on the backing
 | |
| device is still available at an 8KiB offset. So either via a loopdev
 | |
| of the backing device created with --offset 8K, or any value defined by
 | |
| --data-offset when you originally formatted bcache with `bcache make`.
 | |
| 
 | |
| For example::
 | |
| 
 | |
| 	losetup -o 8192 /dev/loop0 /dev/your_bcache_backing_dev
 | |
| 
 | |
| This should present your unmodified backing device data in /dev/loop0
 | |
| 
 | |
| If your cache is in writethrough mode, then you can safely discard the
 | |
| cache device without losing data.
 | |
| 
 | |
| 
 | |
| E) Wiping a cache device
 | |
| 
 | |
| ::
 | |
| 
 | |
| 	host:~# wipefs -a /dev/sdh2
 | |
| 	16 bytes were erased at offset 0x1018 (bcache)
 | |
| 	they were: c6 85 73 f6 4e 1a 45 ca 82 65 f5 7f 48 ba 6d 81
 | |
| 
 | |
| After you boot back with bcache enabled, you recreate the cache and attach it::
 | |
| 
 | |
| 	host:~# bcache make -C /dev/sdh2
 | |
| 	UUID:                   7be7e175-8f4c-4f99-94b2-9c904d227045
 | |
| 	Set UUID:               5bc072a8-ab17-446d-9744-e247949913c1
 | |
| 	version:                0
 | |
| 	nbuckets:               106874
 | |
| 	block_size:             1
 | |
| 	bucket_size:            1024
 | |
| 	nr_in_set:              1
 | |
| 	nr_this_dev:            0
 | |
| 	first_bucket:           1
 | |
| 	[  650.511912] bcache: run_cache_set() invalidating existing data
 | |
| 	[  650.549228] bcache: register_cache() registered cache device sdh2
 | |
| 
 | |
| start backing device with missing cache::
 | |
| 
 | |
| 	host:/sys/block/md5/bcache# echo 1 > running
 | |
| 
 | |
| attach new cache::
 | |
| 
 | |
| 	host:/sys/block/md5/bcache# echo 5bc072a8-ab17-446d-9744-e247949913c1 > attach
 | |
| 	[  865.276616] bcache: bch_cached_dev_attach() Caching md5 as bcache0 on set 5bc072a8-ab17-446d-9744-e247949913c1
 | |
| 
 | |
| 
 | |
| F) Remove or replace a caching device::
 | |
| 
 | |
| 	host:/sys/block/sda/sda7/bcache# echo 1 > detach
 | |
| 	[  695.872542] bcache: cached_dev_detach_finish() Caching disabled for sda7
 | |
| 
 | |
| 	host:~# wipefs -a /dev/nvme0n1p4
 | |
| 	wipefs: error: /dev/nvme0n1p4: probing initialization failed: Device or resource busy
 | |
| 	Ooops, it's disabled, but not unregistered, so it's still protected
 | |
| 
 | |
| We need to go and unregister it::
 | |
| 
 | |
| 	host:/sys/fs/bcache/b7ba27a1-2398-4649-8ae3-0959f57ba128# ls -l cache0
 | |
| 	lrwxrwxrwx 1 root root 0 Feb 25 18:33 cache0 -> ../../../devices/pci0000:00/0000:00:1d.0/0000:70:00.0/nvme/nvme0/nvme0n1/nvme0n1p4/bcache/
 | |
| 	host:/sys/fs/bcache/b7ba27a1-2398-4649-8ae3-0959f57ba128# echo 1 > stop
 | |
| 	kernel: [  917.041908] bcache: cache_set_free() Cache set b7ba27a1-2398-4649-8ae3-0959f57ba128 unregistered
 | |
| 
 | |
| Now we can wipe it::
 | |
| 
 | |
| 	host:~# wipefs -a /dev/nvme0n1p4
 | |
| 	/dev/nvme0n1p4: 16 bytes were erased at offset 0x00001018 (bcache): c6 85 73 f6 4e 1a 45 ca 82 65 f5 7f 48 ba 6d 81
 | |
| 
 | |
| 
 | |
| G) dm-crypt and bcache
 | |
| 
 | |
| First setup bcache unencrypted and then install dmcrypt on top of
 | |
| /dev/bcache<N> This will work faster than if you dmcrypt both the backing
 | |
| and caching devices and then install bcache on top. [benchmarks?]
 | |
| 
 | |
| 
 | |
| H) Stop/free a registered bcache to wipe and/or recreate it
 | |
| 
 | |
| Suppose that you need to free up all bcache references so that you can
 | |
| fdisk run and re-register a changed partition table, which won't work
 | |
| if there are any active backing or caching devices left on it:
 | |
| 
 | |
| 1) Is it present in /dev/bcache* ? (there are times where it won't be)
 | |
| 
 | |
|    If so, it's easy::
 | |
| 
 | |
| 	host:/sys/block/bcache0/bcache# echo 1 > stop
 | |
| 
 | |
| 2) But if your backing device is gone, this won't work::
 | |
| 
 | |
| 	host:/sys/block/bcache0# cd bcache
 | |
| 	bash: cd: bcache: No such file or directory
 | |
| 
 | |
|    In this case, you may have to unregister the dmcrypt block device that
 | |
|    references this bcache to free it up::
 | |
| 
 | |
| 	host:~# dmsetup remove oldds1
 | |
| 	bcache: bcache_device_free() bcache0 stopped
 | |
| 	bcache: cache_set_free() Cache set 5bc072a8-ab17-446d-9744-e247949913c1 unregistered
 | |
| 
 | |
|    This causes the backing bcache to be removed from /sys/fs/bcache and
 | |
|    then it can be reused.  This would be true of any block device stacking
 | |
|    where bcache is a lower device.
 | |
| 
 | |
| 3) In other cases, you can also look in /sys/fs/bcache/::
 | |
| 
 | |
| 	host:/sys/fs/bcache# ls -l */{cache?,bdev?}
 | |
| 	lrwxrwxrwx 1 root root 0 Mar  5 09:39 0226553a-37cf-41d5-b3ce-8b1e944543a8/bdev1 -> ../../../devices/virtual/block/dm-1/bcache/
 | |
| 	lrwxrwxrwx 1 root root 0 Mar  5 09:39 0226553a-37cf-41d5-b3ce-8b1e944543a8/cache0 -> ../../../devices/virtual/block/dm-4/bcache/
 | |
| 	lrwxrwxrwx 1 root root 0 Mar  5 09:39 5bc072a8-ab17-446d-9744-e247949913c1/cache0 -> ../../../devices/pci0000:00/0000:00:01.0/0000:01:00.0/ata10/host9/target9:0:0/9:0:0:0/block/sdl/sdl2/bcache/
 | |
| 
 | |
|    The device names will show which UUID is relevant, cd in that directory
 | |
|    and stop the cache::
 | |
| 
 | |
| 	host:/sys/fs/bcache/5bc072a8-ab17-446d-9744-e247949913c1# echo 1 > stop
 | |
| 
 | |
|    This will free up bcache references and let you reuse the partition for
 | |
|    other purposes.
 | |
| 
 | |
| 
 | |
| 
 | |
| Troubleshooting performance
 | |
| ---------------------------
 | |
| 
 | |
| Bcache has a bunch of config options and tunables. The defaults are intended to
 | |
| be reasonable for typical desktop and server workloads, but they're not what you
 | |
| want for getting the best possible numbers when benchmarking.
 | |
| 
 | |
|  - Backing device alignment
 | |
| 
 | |
|    The default metadata size in bcache is 8k.  If your backing device is
 | |
|    RAID based, then be sure to align this by a multiple of your stride
 | |
|    width using `bcache make --data-offset`. If you intend to expand your
 | |
|    disk array in the future, then multiply a series of primes by your
 | |
|    raid stripe size to get the disk multiples that you would like.
 | |
| 
 | |
|    For example:  If you have a 64k stripe size, then the following offset
 | |
|    would provide alignment for many common RAID5 data spindle counts::
 | |
| 
 | |
| 	64k * 2*2*2*3*3*5*7 bytes = 161280k
 | |
| 
 | |
|    That space is wasted, but for only 157.5MB you can grow your RAID 5
 | |
|    volume to the following data-spindle counts without re-aligning::
 | |
| 
 | |
| 	3,4,5,6,7,8,9,10,12,14,15,18,20,21 ...
 | |
| 
 | |
|  - Bad write performance
 | |
| 
 | |
|    If write performance is not what you expected, you probably wanted to be
 | |
|    running in writeback mode, which isn't the default (not due to a lack of
 | |
|    maturity, but simply because in writeback mode you'll lose data if something
 | |
|    happens to your SSD)::
 | |
| 
 | |
| 	# echo writeback > /sys/block/bcache0/bcache/cache_mode
 | |
| 
 | |
|  - Bad performance, or traffic not going to the SSD that you'd expect
 | |
| 
 | |
|    By default, bcache doesn't cache everything. It tries to skip sequential IO -
 | |
|    because you really want to be caching the random IO, and if you copy a 10
 | |
|    gigabyte file you probably don't want that pushing 10 gigabytes of randomly
 | |
|    accessed data out of your cache.
 | |
| 
 | |
|    But if you want to benchmark reads from cache, and you start out with fio
 | |
|    writing an 8 gigabyte test file - so you want to disable that::
 | |
| 
 | |
| 	# echo 0 > /sys/block/bcache0/bcache/sequential_cutoff
 | |
| 
 | |
|    To set it back to the default (4 mb), do::
 | |
| 
 | |
| 	# echo 4M > /sys/block/bcache0/bcache/sequential_cutoff
 | |
| 
 | |
|  - Traffic's still going to the spindle/still getting cache misses
 | |
| 
 | |
|    In the real world, SSDs don't always keep up with disks - particularly with
 | |
|    slower SSDs, many disks being cached by one SSD, or mostly sequential IO. So
 | |
|    you want to avoid being bottlenecked by the SSD and having it slow everything
 | |
|    down.
 | |
| 
 | |
|    To avoid that bcache tracks latency to the cache device, and gradually
 | |
|    throttles traffic if the latency exceeds a threshold (it does this by
 | |
|    cranking down the sequential bypass).
 | |
| 
 | |
|    You can disable this if you need to by setting the thresholds to 0::
 | |
| 
 | |
| 	# echo 0 > /sys/fs/bcache/<cache set>/congested_read_threshold_us
 | |
| 	# echo 0 > /sys/fs/bcache/<cache set>/congested_write_threshold_us
 | |
| 
 | |
|    The default is 2000 us (2 milliseconds) for reads, and 20000 for writes.
 | |
| 
 | |
|  - Still getting cache misses, of the same data
 | |
| 
 | |
|    One last issue that sometimes trips people up is actually an old bug, due to
 | |
|    the way cache coherency is handled for cache misses. If a btree node is full,
 | |
|    a cache miss won't be able to insert a key for the new data and the data
 | |
|    won't be written to the cache.
 | |
| 
 | |
|    In practice this isn't an issue because as soon as a write comes along it'll
 | |
|    cause the btree node to be split, and you need almost no write traffic for
 | |
|    this to not show up enough to be noticeable (especially since bcache's btree
 | |
|    nodes are huge and index large regions of the device). But when you're
 | |
|    benchmarking, if you're trying to warm the cache by reading a bunch of data
 | |
|    and there's no other traffic - that can be a problem.
 | |
| 
 | |
|    Solution: warm the cache by doing writes, or use the testing branch (there's
 | |
|    a fix for the issue there).
 | |
| 
 | |
| 
 | |
| Sysfs - backing device
 | |
| ----------------------
 | |
| 
 | |
| Available at /sys/block/<bdev>/bcache, /sys/block/bcache*/bcache and
 | |
| (if attached) /sys/fs/bcache/<cset-uuid>/bdev*
 | |
| 
 | |
| attach
 | |
|   Echo the UUID of a cache set to this file to enable caching.
 | |
| 
 | |
| cache_mode
 | |
|   Can be one of either writethrough, writeback, writearound or none.
 | |
| 
 | |
| clear_stats
 | |
|   Writing to this file resets the running total stats (not the day/hour/5 minute
 | |
|   decaying versions).
 | |
| 
 | |
| detach
 | |
|   Write to this file to detach from a cache set. If there is dirty data in the
 | |
|   cache, it will be flushed first.
 | |
| 
 | |
| dirty_data
 | |
|   Amount of dirty data for this backing device in the cache. Continuously
 | |
|   updated unlike the cache set's version, but may be slightly off.
 | |
| 
 | |
| label
 | |
|   Name of underlying device.
 | |
| 
 | |
| readahead
 | |
|   Size of readahead that should be performed.  Defaults to 0.  If set to e.g.
 | |
|   1M, it will round cache miss reads up to that size, but without overlapping
 | |
|   existing cache entries.
 | |
| 
 | |
| running
 | |
|   1 if bcache is running (i.e. whether the /dev/bcache device exists, whether
 | |
|   it's in passthrough mode or caching).
 | |
| 
 | |
| sequential_cutoff
 | |
|   A sequential IO will bypass the cache once it passes this threshold; the
 | |
|   most recent 128 IOs are tracked so sequential IO can be detected even when
 | |
|   it isn't all done at once.
 | |
| 
 | |
| sequential_merge
 | |
|   If non zero, bcache keeps a list of the last 128 requests submitted to compare
 | |
|   against all new requests to determine which new requests are sequential
 | |
|   continuations of previous requests for the purpose of determining sequential
 | |
|   cutoff. This is necessary if the sequential cutoff value is greater than the
 | |
|   maximum acceptable sequential size for any single request.
 | |
| 
 | |
| state
 | |
|   The backing device can be in one of four different states:
 | |
| 
 | |
|   no cache: Has never been attached to a cache set.
 | |
| 
 | |
|   clean: Part of a cache set, and there is no cached dirty data.
 | |
| 
 | |
|   dirty: Part of a cache set, and there is cached dirty data.
 | |
| 
 | |
|   inconsistent: The backing device was forcibly run by the user when there was
 | |
|   dirty data cached but the cache set was unavailable; whatever data was on the
 | |
|   backing device has likely been corrupted.
 | |
| 
 | |
| stop
 | |
|   Write to this file to shut down the bcache device and close the backing
 | |
|   device.
 | |
| 
 | |
| writeback_delay
 | |
|   When dirty data is written to the cache and it previously did not contain
 | |
|   any, waits some number of seconds before initiating writeback. Defaults to
 | |
|   30.
 | |
| 
 | |
| writeback_percent
 | |
|   If nonzero, bcache tries to keep around this percentage of the cache dirty by
 | |
|   throttling background writeback and using a PD controller to smoothly adjust
 | |
|   the rate.
 | |
| 
 | |
| writeback_rate
 | |
|   Rate in sectors per second - if writeback_percent is nonzero, background
 | |
|   writeback is throttled to this rate. Continuously adjusted by bcache but may
 | |
|   also be set by the user.
 | |
| 
 | |
| writeback_running
 | |
|   If off, writeback of dirty data will not take place at all. Dirty data will
 | |
|   still be added to the cache until it is mostly full; only meant for
 | |
|   benchmarking. Defaults to on.
 | |
| 
 | |
| Sysfs - backing device stats
 | |
| ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 | |
| 
 | |
| There are directories with these numbers for a running total, as well as
 | |
| versions that decay over the past day, hour and 5 minutes; they're also
 | |
| aggregated in the cache set directory as well.
 | |
| 
 | |
| bypassed
 | |
|   Amount of IO (both reads and writes) that has bypassed the cache
 | |
| 
 | |
| cache_hits, cache_misses, cache_hit_ratio
 | |
|   Hits and misses are counted per individual IO as bcache sees them; a
 | |
|   partial hit is counted as a miss.
 | |
| 
 | |
| cache_bypass_hits, cache_bypass_misses
 | |
|   Hits and misses for IO that is intended to skip the cache are still counted,
 | |
|   but broken out here.
 | |
| 
 | |
| cache_miss_collisions
 | |
|   Counts instances where data was going to be inserted into the cache from a
 | |
|   cache miss, but raced with a write and data was already present (usually 0
 | |
|   since the synchronization for cache misses was rewritten)
 | |
| 
 | |
| Sysfs - cache set
 | |
| ~~~~~~~~~~~~~~~~~
 | |
| 
 | |
| Available at /sys/fs/bcache/<cset-uuid>
 | |
| 
 | |
| average_key_size
 | |
|   Average data per key in the btree.
 | |
| 
 | |
| bdev<0..n>
 | |
|   Symlink to each of the attached backing devices.
 | |
| 
 | |
| block_size
 | |
|   Block size of the cache devices.
 | |
| 
 | |
| btree_cache_size
 | |
|   Amount of memory currently used by the btree cache
 | |
| 
 | |
| bucket_size
 | |
|   Size of buckets
 | |
| 
 | |
| cache<0..n>
 | |
|   Symlink to each of the cache devices comprising this cache set.
 | |
| 
 | |
| cache_available_percent
 | |
|   Percentage of cache device which doesn't contain dirty data, and could
 | |
|   potentially be used for writeback.  This doesn't mean this space isn't used
 | |
|   for clean cached data; the unused statistic (in priority_stats) is typically
 | |
|   much lower.
 | |
| 
 | |
| clear_stats
 | |
|   Clears the statistics associated with this cache
 | |
| 
 | |
| dirty_data
 | |
|   Amount of dirty data is in the cache (updated when garbage collection runs).
 | |
| 
 | |
| flash_vol_create
 | |
|   Echoing a size to this file (in human readable units, k/M/G) creates a thinly
 | |
|   provisioned volume backed by the cache set.
 | |
| 
 | |
| io_error_halflife, io_error_limit
 | |
|   These determines how many errors we accept before disabling the cache.
 | |
|   Each error is decayed by the half life (in # ios).  If the decaying count
 | |
|   reaches io_error_limit dirty data is written out and the cache is disabled.
 | |
| 
 | |
| journal_delay_ms
 | |
|   Journal writes will delay for up to this many milliseconds, unless a cache
 | |
|   flush happens sooner. Defaults to 100.
 | |
| 
 | |
| root_usage_percent
 | |
|   Percentage of the root btree node in use.  If this gets too high the node
 | |
|   will split, increasing the tree depth.
 | |
| 
 | |
| stop
 | |
|   Write to this file to shut down the cache set - waits until all attached
 | |
|   backing devices have been shut down.
 | |
| 
 | |
| tree_depth
 | |
|   Depth of the btree (A single node btree has depth 0).
 | |
| 
 | |
| unregister
 | |
|   Detaches all backing devices and closes the cache devices; if dirty data is
 | |
|   present it will disable writeback caching and wait for it to be flushed.
 | |
| 
 | |
| Sysfs - cache set internal
 | |
| ~~~~~~~~~~~~~~~~~~~~~~~~~~
 | |
| 
 | |
| This directory also exposes timings for a number of internal operations, with
 | |
| separate files for average duration, average frequency, last occurrence and max
 | |
| duration: garbage collection, btree read, btree node sorts and btree splits.
 | |
| 
 | |
| active_journal_entries
 | |
|   Number of journal entries that are newer than the index.
 | |
| 
 | |
| btree_nodes
 | |
|   Total nodes in the btree.
 | |
| 
 | |
| btree_used_percent
 | |
|   Average fraction of btree in use.
 | |
| 
 | |
| bset_tree_stats
 | |
|   Statistics about the auxiliary search trees
 | |
| 
 | |
| btree_cache_max_chain
 | |
|   Longest chain in the btree node cache's hash table
 | |
| 
 | |
| cache_read_races
 | |
|   Counts instances where while data was being read from the cache, the bucket
 | |
|   was reused and invalidated - i.e. where the pointer was stale after the read
 | |
|   completed. When this occurs the data is reread from the backing device.
 | |
| 
 | |
| trigger_gc
 | |
|   Writing to this file forces garbage collection to run.
 | |
| 
 | |
| Sysfs - Cache device
 | |
| ~~~~~~~~~~~~~~~~~~~~
 | |
| 
 | |
| Available at /sys/block/<cdev>/bcache
 | |
| 
 | |
| block_size
 | |
|   Minimum granularity of writes - should match hardware sector size.
 | |
| 
 | |
| btree_written
 | |
|   Sum of all btree writes, in (kilo/mega/giga) bytes
 | |
| 
 | |
| bucket_size
 | |
|   Size of buckets
 | |
| 
 | |
| cache_replacement_policy
 | |
|   One of either lru, fifo or random.
 | |
| 
 | |
| discard
 | |
|   Boolean; if on a discard/TRIM will be issued to each bucket before it is
 | |
|   reused. Defaults to off, since SATA TRIM is an unqueued command (and thus
 | |
|   slow).
 | |
| 
 | |
| freelist_percent
 | |
|   Size of the freelist as a percentage of nbuckets. Can be written to to
 | |
|   increase the number of buckets kept on the freelist, which lets you
 | |
|   artificially reduce the size of the cache at runtime. Mostly for testing
 | |
|   purposes (i.e. testing how different size caches affect your hit rate), but
 | |
|   since buckets are discarded when they move on to the freelist will also make
 | |
|   the SSD's garbage collection easier by effectively giving it more reserved
 | |
|   space.
 | |
| 
 | |
| io_errors
 | |
|   Number of errors that have occurred, decayed by io_error_halflife.
 | |
| 
 | |
| metadata_written
 | |
|   Sum of all non data writes (btree writes and all other metadata).
 | |
| 
 | |
| nbuckets
 | |
|   Total buckets in this cache
 | |
| 
 | |
| priority_stats
 | |
|   Statistics about how recently data in the cache has been accessed.
 | |
|   This can reveal your working set size.  Unused is the percentage of
 | |
|   the cache that doesn't contain any data.  Metadata is bcache's
 | |
|   metadata overhead.  Average is the average priority of cache buckets.
 | |
|   Next is a list of quantiles with the priority threshold of each.
 | |
| 
 | |
| written
 | |
|   Sum of all data that has been written to the cache; comparison with
 | |
|   btree_written gives the amount of write inflation in bcache.
 |