Provided by: lvm2_2.03.22-1ubuntu2_amd64 bug

NAME

       lvmcache — LVM caching

DESCRIPTION

       lvm(8)  includes  two  kinds  of  caching that can be used to improve the performance of a Logical Volume
       (LV). When caching, varying subsets of an LV's data are temporarily stored on a  smaller,  faster  device
       (e.g. an SSD) to improve the performance of the LV.

       To  do  this  with  lvm,  a new special LV is first created from the faster device. This LV will hold the
       cache. Then, the new fast LV is attached to the main LV by way of an lvconvert command. lvconvert inserts
       one of the device mapper caching targets into the main LV's i/o path. The device mapper  target  combines
       the  main  LV  and  fast LV into a hybrid device that looks like the main LV, but has better performance.
       While the main LV is being used, portions of its data will be temporarily and transparently stored on the
       special fast LV.

       The two kinds of caching are:

       • A read and write hot-spot cache, using the dm-cache kernel module.  This cache tracks  access  patterns
         and  adjusts its content deliberately so that commonly used parts of the main LV are likely to be found
         on the fast storage. LVM refers to this using the LV type cache.

       • A write cache, using the dm-writecache kernel module.  This cache can be used with SSD or PMEM  devices
         to  speed  up  all  writes  to the main LV. Data read from the main LV is not stored in the cache, only
         newly written data.  LVM refers to this using the LV type writecache.

USAGE

   1. Identify main LV that needs caching
       The main LV may already exist, and is located on larger, slower devices.  A main LV would be created with
       a command like:

       # lvcreate -n main -L Size vg /dev/slow_hhd

   2. Identify fast LV to use as the cache
       A fast LV is created using one or more fast devices, like an SSD.  This special LV will be used  to  hold
       the cache:

       # lvcreate -n fast -L Size vg /dev/fast_ssd

       # lvs -a
         LV   Attr       Type   Devices
         fast -wi------- linear /dev/fast_ssd
         main -wi------- linear /dev/slow_hhd

   3. Start caching the main LV
       To start caching the main LV, convert the main LV to the desired caching type, and specify the fast LV to
       use as the cache:

       using dm-cache (with cachepool):

       # lvconvert --type cache --cachepool fast vg/main

       using dm-cache (with cachevol):

       # lvconvert --type cache --cachevol fast vg/main

       using dm-writecache (with cachevol):

       # lvconvert --type writecache --cachevol fast vg/main

       For more alternatives see:
       dm-cache command shortcut
       dm-cache with separate data and metadata LVs

   4. Display LVs
       Once  the  fast  LV  has  been  attached  to the main LV, lvm reports the main LV type as either cache or
       writecache depending on the type used.  While attached, the fast LV is hidden, and renamed with  a  _cvol
       or  _cpool  suffix.   It  is  displayed  by  lvs -a.  The _corig or _wcorig LV represents the original LV
       without the cache.

       using dm-cache (with cachepool):

       # lvs -ao+devices
         LV                 Pool         Type       Devices
         main               [fast_cpool] cache      main_corig(0)
         [fast_cpool]                    cache-pool fast_pool_cdata(0)
         [fast_cpool_cdata]              linear     /dev/fast_ssd
         [fast_cpool_cmeta]              linear     /dev/fast_ssd
         [main_corig]                    linear     /dev/slow_hhd

       using dm-cache (with cachevol):

       # lvs -ao+devices

         LV           Pool        Type   Devices
         main         [fast_cvol] cache  main_corig(0)
         [fast_cvol]              linear /dev/fast_ssd
         [main_corig]             linear /dev/slow_hhd

       using dm-writecache (with cachevol):

       # lvs -ao+devices

         LV            Pool        Type       Devices
         main          [fast_cvol] writecache main_wcorig(0)
         [fast_cvol]               linear     /dev/fast_ssd
         [main_wcorig]             linear     /dev/slow_hhd

   5. Use the main LV
       Use the LV until the cache is no longer wanted, or needs to be changed.

   6. Stop caching
       To stop caching the main LV and also remove unneeded cache pool, use the --uncache:

       # lvconvert --uncache vg/main

       # lvs -a
         LV   VG Attr       Type   Devices
         main vg -wi------- linear /dev/slow_hhd

       To stop caching the main LV, separate the fast LV from the main LV.  This changes the type of the main LV
       back to what it was before the cache was attached.

       # lvconvert --splitcache vg/main

       # lvs -a
         LV   VG Attr       Type   Devices
         fast vg -wi------- linear /dev/fast_ssd
         main vg -wi------- linear /dev/slow_hhd

   7. Create a new LV with caching
       A new LV can be created with caching attached at the time of creation using the following command:

       # lvcreate --type cache|writecache -n Name -L Size
            --cachedevice /dev/fast_ssd vg /dev/slow_hhd

       The main LV is created with the specified Name and Size from the slow_hhd.  A hidden fast LV  is  created
       on the fast_ssd and is then attached to the new main LV.  If the fast_ssd is unused, the entire disk will
       be  used  as  the  cache  unless  the  --cachesize option is used to specify a size for the fast LV.  The
       --cachedevice option can be repeated to use multiple disks for the fast LV.

OPTIONS

   option args
       --cachepool CachePoolLV|LV

       Pass this option a cachepool LV or a standard LV.  When using a cache pool, lvm  places  cache  data  and
       cache  metadata  on  different LVs.  The two LVs together are called a cache pool.  This has a bit better
       performance for dm-cache and permits specific placement and segment type selection for data and  metadata
       volumes.   A  cache  pool  is  represented  as  a  special type of LV that cannot be used directly.  If a
       standard LV is passed with this option, lvm will first convert it to a cache pool by  combining  it  with
       another LV to use for metadata.  This option can be used with dm-cache.

       --cachevol LV

       Pass  this  option  a  fast  LV  that  should be used to hold the cache.  With a cachevol, cache data and
       metadata are stored in different parts of the same fast LV.  This option can be used  with  dm-writecache
       or dm-cache.

       --cachedevice PV

       This  option  can  be  used in place of --cachevol, in which case a cachevol LV will be created using the
       specified device.  This option can be repeated to create a cachevol using multiple devices, or a tag name
       can be specified in which case the cachevol will be created using any of the devices with the given  tag.
       If  a  named  cache device is unused, the entire device will be used to create the cachevol.  To create a
       cachevol of a specific size from the cache devices, include the --cachesize option.

   dm-cache block size
       A cache pool will have a logical block size of 4096 bytes if it is created on a  device  with  a  logical
       block size of 4096 bytes.

       If  a  main  LV  has  logical  block size 512 (with an existing xfs file system using that size), then it
       cannot use a cache pool with a 4096 logical block size.  If the cache pool is attached, the main LV  will
       likely fail to mount.

       To  avoid this problem, use a mkfs option to specify a 4096 block size for the file system, or attach the
       cache pool before running mkfs.

   dm-writecache block size
       The dm-writecache block size can be 4096 bytes (the default), or 512 bytes.  The default 4096 has  better
       performance  and  should be used except when 512 is necessary for compatibility.  The dm-writecache block
       size is specified with --cachesettings block_size=4096|512 when caching is started.

       When a file system like xfs already exists on the main LV prior to caching, and the file system is  using
       a  block  size of 512, then the writecache block size should be set to 512.  (The file system will likely
       fail to mount if writecache block size of 4096 is used in this case.)

       Check the xfs sector size while the fs is mounted:

       # xfs_info /dev/vg/main
       Look for sectsz=512 or sectsz=4096

       The writecache block size should be chosen to match the xfs sectsz value.

       It is also possible to specify a sector size of 4096 to mkfs.xfs when creating the file system.  In  this
       case the writecache block size of 4096 can be used.

       The writecache block size is displayed by the command:
       lvs -o writecacheblocksize VG/LV

   dm-writecache memory usage
       The  amount  of  main  system  memory used by dm-writecache can be a factor when selecting the writecache
       cachevol size and the writecache block size.

       • writecache block size 4096: each 100 GiB of writecache cachevol uses  slightly  over  2 GiB  of  system
         memory.

       • writecache  block  size  512:  each  100 GiB of writecache cachevol uses a little over 16 GiB of system
         memory.

   dm-writecache settings
       To specify dm-writecache tunable settings on the command line, use:
       --cachesettings 'option=N' or
       --cachesettings 'option1=N option2=N ...'

       For example, --cachesettings 'high_watermark=90 writeback_jobs=4'.

       To include settings when caching is started, run:

       # lvconvert --type writecache --cachevol fast \
            --cachesettings 'option=N' vg/main

       To change settings for an existing writecache, run:

       # lvchange --cachesettings 'option=N' vg/main

       To clear all settings that have been applied, run:

       # lvchange --cachesettings '' vg/main

       To view the settings that are applied to a writecache LV, run:

       # lvs -o cachesettings vg/main

       Tunable settings are:

       high_watermark = <percent>
              Start writeback when the writecache usage reaches this percent (0-100).

       low_watermark = <percent>
              Stop writeback when the writecache usage reaches this percent (0-100).

       writeback_jobs = <count>
              Limit the number of blocks that are in  flight  during  writeback.   Setting  this  value  reduces
              writeback throughput, but it may improve latency of read requests.

       autocommit_blocks = <count>
              When  the  application  writes this amount of blocks without issuing the FLUSH request, the blocks
              are automatically committed.

       autocommit_time = <milliseconds>
              The data is automatically committed if this time passes and no FLUSH request is received.

       fua = 0|1
              Use the FUA flag when  writing  data  from  persistent  memory  back  to  the  underlying  device.
              Applicable only to persistent memory.

       nofua = 0|1
              Don't  use  the  FUA  flag  when  writing  back  data and send the FLUSH request afterwards.  Some
              underlying devices perform better with fua, some with nofua.  Testing is  necessary  to  determine
              which.  Applicable only to persistent memory.

       cleaner = 0|1
              Setting  cleaner=1 enables the writecache cleaner mode in which data is gradually flushed from the
              cache.  If this is done prior to detaching the writecache, then the splitcache command  will  have
              little  or  no  flushing  to  perform.  If not done beforehand, the splitcache command enables the
              cleaner mode and waits for flushing to complete before detaching the writecache.  Adding cleaner=0
              to the splitcache command will skip the cleaner mode, and any required flushing  is  performed  in
              device suspend.

       max_age = <milliseconds>
              Specifies  the  maximum  age of a block in milliseconds. If a block is stored in the cache for too
              long, it will be written to the underlying device and cleaned up.

       metadata_only = 0|1
              Only metadata is promoted to the cache. This option  improves  performance  for  heavier  REQ_META
              workloads.

       pause_writeback = <milliseconds>
              Pause  writeback if there was some write I/O redirected to the origin volume in the last number of
              milliseconds.

   dm-writecache using metadata profiles
       In addition to specifying writecache settings on the command line, they can also be set in  lvm.conf,  or
       in a profile file, using the allocation/cache_settings/writecache config structure shown below.

       It's  possible  to  prepare  a  number  of  different profile files in the /etc/lvm/profile directory and
       specify the file name of the profile when starting writecache.

       Example
       # cat <<EOF > /etc/lvm/profile/cache_writecache.profile
       allocation {
              cache_settings {
                     writecache {
                            high_watermark=60
                            writeback_jobs=1024
                     }
              }
       }
       EOF

       # lvcreate -an -L10G --name fast vg /dev/fast_ssd
       # lvcreate --type writecache -L10G --name main  --cachevol fast \
          --metadataprofile cache_writecache vg /dev/slow_hdd

   dm-cache with separate data and metadata LVs
       Preferred way of using dm-cache is to place the cache metadata and cache data on  separate  LVs.   To  do
       this,  a "cache pool" is created, which is a special LV that references two sub LVs, one for data and one
       for metadata.

       To create a cache pool of given data size and let lvm2 calculate appropriate metadata size:

       # lvcreate --type cache-pool -L DataSize -n fast vg /dev/fast_ssd1

       To create a cache pool from separate LV and let lvm2 calculate appropriate cache metadata size:

       # lvcreate -n fast -L DataSize vg /dev/fast_ssd1
       # lvconvert --type cache-pool vg/fast /dev/fast_ssd1

       To create a cache pool from two separate LVs:

       # lvcreate -n fast -L DataSize vg /dev/fast_ssd1
       # lvcreate -n fastmeta -L MetadataSize vg /dev/fast_ssd2
       # lvconvert --type cache-pool --poolmetadata fastmeta vg/fast

       Then use the cache pool LV to start caching the main LV:

       # lvconvert --type cache --cachepool fast vg/main

       A variation of the same procedure automatically creates a cache pool when  caching  is  started.   To  do
       this,  use  a  standard LV as the --cachepool (this will hold cache data), and use another standard LV as
       the --poolmetadata (this will hold cache metadata).  LVM will  create  a  cache  pool  LV  from  the  two
       specified LVs, and use the cache pool to start caching the main LV.

       # lvcreate -n fast -L DataSize vg /dev/fast_ssd1
       # lvcreate -n fastmeta -L MetadataSize vg /dev/fast_ssd2
       # lvconvert --type cache --cachepool fast \
               --poolmetadata fastmeta vg/main

   dm-cache cache modes
       The  default  dm-cache  cache mode is "writethrough".  Writethrough ensures that any data written will be
       stored both in the cache and on the origin LV.  The loss of a device associated with the  cache  in  this
       case would not mean the loss of any data.

       A  second  cache  mode  is  "writeback".  Writeback delays writing data blocks from the cache back to the
       origin LV.  This mode will increase performance, but the loss of a cache device can result in lost data.

       With the --cachemode option, the cache mode can be set when caching is started, or changed on an LV  that
       is already cached.  The current cache mode can be displayed with the cache_mode reporting option:

       lvs -o+cache_mode VG/LV

       lvm.conf(5) allocation/cache_mode
       defines the default cache mode.

       # lvconvert --type cache --cachemode writethrough \
               --cachepool fast vg/main

       # lvconvert --type cache --cachemode writethrough \
               --cachevol fast  vg/main

   dm-cache chunk size
       The  size of data blocks managed by dm-cache can be specified with the --chunksize option when caching is
       started.  The default unit is KiB.  The value must be a multiple of  32 KiB  between  32 KiB  and  1 GiB.
       Cache chunks bigger then 512KiB shall be only used when necessary.

       Using  a  chunk  size that is too large can result in wasteful use of the cache, in which small reads and
       writes cause large sections of an LV to be stored in the cache. It can also require increasing  migration
       threshold  which  defaults to 2048 sectors (1 MiB). Lvm2 ensures migration threshold is at least 8 chunks
       in size. This may in some cases result in very high bandwidth load of transferring data between the cache
       LV and its cache origin LV. However, choosing a chunk size that is too small can result in more  overhead
       trying  to  manage  the  numerous  chunks  that  become mapped into the cache.  Overhead can include both
       excessive CPU time searching for chunks, and excessive memory tracking chunks.

       Command to display the chunk size:

       lvs -o+chunksize VG/LV

       lvm.conf(5) allocation/cache_pool_chunk_size

       controls the default chunk size.

       The default value is shown by:

       lvmconfig --type default allocation/cache_pool_chunk_size

       Checking migration threshold (in sectors) of running cached LV:
       lvs -o+kernel_cache_settings VG/LV

   dm-cache cache settings
       To set dm-cache cache setting use:

       --cachesettings 'option1=N option2=N ...'

       To unset/drop cache setting and restore its default kernel value use special keyword 'default' as  option
       parameter:

       --cachesettings 'option1=default option2=default ...'

   dm-cache migration threshold cache setting
       Migrating  data  between  the origin and cache LV uses bandwidth.  The user can set a throttle to prevent
       more than a certain amount of migration occurring at any one time.  Currently dm-cache is not taking  any
       account of normal io traffic going to the devices.

       User can set migration threshold via cache policy settings as "migration_threshold=<#sectors>" to set the
       maximum  number  of  sectors  being  migrated,  the  default being 2048 sectors (1 MiB) or 8 cache chunks
       whichever of those two values is larger.

       Command to set migration threshold to 2 MiB (4096 sectors):

       lvcreate --cachesettings 'migration_threshold=4096' VG/LV

       Command to display the migration threshold:

       lvs -o+kernel_cache_settings,cache_settings VG/LV
       lvs -o+chunksize VG/LV

       Command to restore/revert to default value:

       lvchange --cachesettings 'migration_threshold=default' VG/LV

   dm-cache cache policy
       The dm-cache subsystem has additional per-LV parameters: the cache policy to use,  and  possibly  tunable
       parameters  for  the  cache policy.  Three policies are currently available: "smq" is the default policy,
       "mq" is an older implementation, and "cleaner" is used to force the  cache  to  write  back  (flush)  all
       cached writes to the origin LV.

       The  older "mq" policy has a number of tunable parameters. The defaults are chosen to be suitable for the
       majority of systems, but in special circumstances, changing the settings can improve performance.   Newer
       kernels  however  alias  this  policy  with  "smq"  policy.  Cache settings used to configure "mq" policy
       [random_threshold,     sequential_threshold,     discard_promote_adjustment,     read_promote_adjustment,
       write_promote_adjustment] are thus silently ignored also performance matches "smq".

       With the --cachepolicy and --cachesettings options, the cache policy and settings can be set when caching
       is  started,  or changed on an existing cached LV (both options can be used together).  The current cache
       policy and settings can be displayed with the cache_policy and cache_settings reporting options:

       lvs -o+cache_policy,cache_settings VG/LV

       Change the cache policy and settings of an existing LV.
       # lvchange --cachepolicy mq --cachesettings \
            'migration_threshold=2048 random_threshold=4' vg/main

       lvm.conf(5) allocation/cache_policy
       defines the default cache policy.

       lvm.conf(5) allocation/cache_settings
       defines the default cache settings.

   dm-cache using metadata profiles
       Cache pools allows to set a variety of options. Lots of these settings can be specified  in  lvm.conf  or
       profile  settings.  You  can prepare a number of different profiles in the /etc/lvm/profile directory and
       just specify the metadata profile file name when caching LV or creating cache-pool.  Check the output  of
       lvmconfig --type default --withcomments for a detailed description of all individual cache settings.

       Example
       # cat <<EOF > /etc/lvm/profile/cache_big_chunk.profile
       allocation {
              cache_pool_metadata_require_separate_pvs=0
              cache_pool_chunk_size=512
              cache_metadata_format=2
              cache_mode="writethrough"
              cache_policy="smq"
              cache_settings {
                     smq {
                            migration_threshold=8192
                     }
              }
       }
       EOF

       # lvcreate --cache -L10G --metadataprofile cache_big_chunk vg/main \
               /dev/fast_ssd
       # lvcreate --cache -L10G vg/main --config \
               'allocation/cache_pool_chunk_size=512' /dev/fast_ssd

   dm-cache spare metadata LV
       See  lvmthin(7)  for  a  description of the "pool metadata spare" LV.  The same concept is used for cache
       pools.

   dm-cache metadata formats
       There are  two  disk  formats  for  dm-cache  metadata.   The  metadata  format  can  be  specified  with
       --cachemetadataformat  when  caching is started, and cannot be changed.  Format 2 has better performance;
       it is more compact, and stores dirty bits in a separate btree, which improves the speed of shutting  down
       the cache.  With auto, lvm selects the best option provided by the current dm-cache kernel module.

   RAID1 cache device
       RAID1  can  be  used  to  create  the fast LV holding the cache so that it can tolerate a device failure.
       (When using dm-cache with separate data and metadata LVs, each of the sub-LVs can use RAID1.)

       # lvcreate -n main -L Size vg /dev/slow
       # lvcreate --type raid1 -m 1 -n fast -L Size vg /dev/ssd1 /dev/ssd2
       # lvconvert --type cache --cachevol fast vg/main

   dm-cache command shortcut
       A single command can be used to cache main LV with automatic creation of a cache-pool:

       # lvcreate --cache --size CacheDataSize VG/LV [FastPVs]

       or the longer variant

       # lvcreate --type cache --size CacheDataSize \
               --name NameCachePool VG/LV [FastPVs]

       In this command, the specified LV already exists, and is the main LV to be cached.  The command creates a
       new cache pool with size  and  given  name  or  the  name  is  automatically  selected  from  a  sequence
       lvolX_cpool,  using  the  optionally  specified  fast PV(s) (typically an ssd).  Then it attaches the new
       cache pool to the existing main LV to begin caching.

       (Note: ensure that the specified main LV is a standard LV.  If a cache pool LV is  mistakenly  specified,
       then the command does something different.)

       (Note:  the  type  option  is interpreted differently by this command than by normal lvcreate commands in
       which --type specifies the type of the newly created LV.  In this case, an LV  with  type  cache-pool  is
       being created, and the existing main LV is being converted to type cache.)

SEE ALSO

       lvm.conf(5), lvchange(8), lvcreate(8), lvdisplay(8), lvextend(8), lvremove(8), lvrename(8), lvresize(8),
       lvs(8),
       vgchange(8), vgmerge(8), vgreduce(8), vgsplit(8),

       cache_check(8), cache_dump(8), cache_repair(8)

Red Hat, Inc                            LVM TOOLS 2.03.22(2) (2023-08-02)                            LVMCACHE(7)