Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- application: ceph-osd
- application-config:
- trust:
- default: false
- description: Does this application have access to trusted credentials
- source: default
- type: bool
- value: false
- charm: ceph-osd
- settings:
- aa-profile-mode:
- default: disable
- description: |
- Enable apparmor profile. Valid settings: 'complain', 'enforce' or
- 'disable'.
- .
- NOTE: changing the value of this option is disruptive to a running Ceph
- cluster as all ceph-osd processes must be restarted as part of changing
- the apparmor profile enforcement mode. Always test in pre-production
- before enabling AppArmor on a live cluster.
- source: default
- type: string
- value: disable
- autotune:
- default: false
- description: |
- Enabling this option will attempt to tune your network card sysctls and
- hard drive settings. This changes hard drive read ahead settings and
- max_sectors_kb. For the network card this will detect the link speed
- and make appropriate sysctl changes.
- WARNING: This option is DEPRECATED and will be removed in the next release.
- Exercise caution when enabling this feature; examine and
- confirm sysctl values are appropriate for your environment. See
- http://pad.lv/1798794 for a full discussion.
- source: default
- type: boolean
- value: false
- availability_zone:
- description: |
- Custom availability zone to provide to Ceph for the OSD placement
- source: unset
- type: string
- bdev-enable-discard:
- default: auto
- description: |
- Enables async discard on devices. This option will enable/disable both
- bdev-enable-discard and bdev-async-discard options in ceph configuration
- at the same time. The default value "auto" will try to autodetect and
- should work in most cases. If you need to force a behaviour you can
- set it to "enable" or "disable". Only applies for Ceph Mimic or later.
- source: default
- type: string
- value: auto
- bluestore:
- default: true
- description: |
- Enable BlueStore storage backend for OSD devices.
- .
- Only supported with ceph >= 12.2.0.
- .
- Setting to 'False' will use FileStore as the storage format.
- source: default
- type: boolean
- value: true
- bluestore-block-db-size:
- default: 0
- description: |
- Size (in bytes) of a partition, file or LV to use for BlueStore
- metadata or RocksDB SSTs, provided on a per backend device basis.
- .
- Example: 128 GB device, 8 data devices provided in "osd-devices"
- gives 128 / 8 GB = 16 GB = 16000000000 bytes per device.
- .
- A default value is not set as it is calculated by ceph-disk (before Luminous)
- or the charm itself, when ceph-volume is used (Luminous and above).
- source: default
- type: int
- value: 0
- bluestore-block-wal-size:
- default: 0
- description: |
- Size (in bytes) of a partition, file or LV to use for
- BlueStore WAL (RocksDB WAL), provided on a per backend device basis.
- .
- Example: 128 GB device, 8 data devices provided in "osd-devices"
- gives 128 / 8 GB = 16 GB = 16000000000 bytes per device.
- .
- A default value is not set as it is calculated by ceph-disk (before Luminous)
- or the charm itself, when ceph-volume is used (Luminous and above).
- source: default
- type: int
- value: 0
- bluestore-compression-algorithm:
- default: lz4
- description: |
- The default compressor to use (if any) if the per-pool property
- compression_algorithm is not set.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: default
- type: string
- value: lz4
- bluestore-compression-max-blob-size:
- description: |
- Chunks larger than this are broken into smaller blobs sizing bluestore
- compression max blob size before being compressed. The per-pool property
- `compression_max_blob_size` overrides this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: int
- bluestore-compression-max-blob-size-hdd:
- description: |
- Default value of bluestore compression max blob size for rotational
- media. The per-pool property `compression-max-blob-size-hdd` overrides
- this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: int
- bluestore-compression-max-blob-size-ssd:
- description: |
- Default value of bluestore compression max blob size for solid state
- media. The per-pool property `compression-max-blob-size-ssd` overrides
- this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: int
- bluestore-compression-min-blob-size:
- description: |
- Chunks smaller than this are never compressed. The per-pool property
- `compression_min_blob_size` overrides this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: int
- bluestore-compression-min-blob-size-hdd:
- description: |
- Default value of bluestore compression min blob size for rotational
- media. The per-pool property `compression-min-blob-size-hdd` overrides
- this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: int
- bluestore-compression-min-blob-size-ssd:
- description: |
- Default value of bluestore compression min blob size for solid state
- media. The per-pool property `compression-min-blob-size-ssd` overrides
- this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: int
- bluestore-compression-mode:
- description: |
- The default policy for using compression if the per-pool property
- compression_mode is not set. 'none' means never use compression.
- 'passive' means use compression when clients hint that data is
- compressible. 'aggressive' means use compression unless clients hint that
- data is not compressible. 'force' means use compression under all
- circumstances even if the clients hint that the data is not compressible.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: string
- bluestore-compression-required-ratio:
- description: |
- The ratio of the size of the data chunk after compression relative to the
- original size must be at least this small in order to store the
- compressed version. The per-pool property `compression-required-ratio`
- overrides this setting.
- .
- NOTE: The recommended approach is to adjust this configuration option on
- the charm responsible for creating the specific pool you are interested
- in tuning. Changing the configuration option on the ceph-osd charm will
- affect ALL pools on the OSDs managed by the named application of the
- ceph-osd charm in the Juju model.
- source: unset
- type: float
- bluestore-db:
- description: |
- Path to a BlueStore WAL db block device or file. If you have a separate
- physical device faster than the block device this will store all of the
- filesystem metadata (RocksDB) there and also integrates the Write Ahead
- Log (WAL) unless a further separate bluestore-wal device is configured
- which is not needed unless it is faster again than the bluestore-db
- device. This block device is used as an LVM PV and then space is
- allocated for each block device as needed based on the
- bluestore-block-db-size setting.
- source: unset
- type: string
- bluestore-wal:
- description: |
- Path to a BlueStore WAL block device or file. Should only be set if using
- a separate physical device that is faster than the DB device (such as an
- NVDIMM or faster SSD). Otherwise BlueStore automatically maintains the
- WAL inside of the DB device. This block device is used as an LVM PV and
- then space is allocated for each block device as needed based on the
- bluestore-block-wal-size setting.
- source: unset
- type: string
- ceph-cluster-network:
- description: |
- The IP address and netmask of the cluster (back-side) network (e.g.,
- 192.168.0.0/24)
- .
- If multiple networks are to be used, a space-delimited list of a.b.c.d/x
- can be provided.
- source: unset
- type: string
- ceph-public-network:
- description: |
- The IP address and netmask of the public (front-side) network (e.g.,
- 192.168.0.0/24)
- .
- If multiple networks are to be used, a space-delimited list of a.b.c.d/x
- can be provided.
- source: unset
- type: string
- config-flags:
- description: |
- User provided Ceph configuration. Supports a string representation of
- a python dictionary where each top-level key represents a section in
- the ceph.conf template. You may only use sections supported in the
- template.
- .
- WARNING: this is not the recommended way to configure the underlying
- services that this charm installs and is used at the user's own risk.
- This option is mainly provided as a stop-gap for users that either
- want to test the effect of modifying some config or who have found
- a critical bug in the way the charm has configured their services
- and need it fixed immediately. We ask that whenever this is used,
- that the user consider opening a bug on this charm at
- http://bugs.launchpad.net/charms providing an explanation of why the
- config was needed so that we may consider it for inclusion as a
- natively supported config in the charm.
- source: unset
- type: string
- crush-initial-weight:
- description: |
- The initial crush weight for newly added osds into crushmap. Use this
- option only if you wish to set the weight for newly added OSDs in order
- to gradually increase the weight over time. Be very aware that setting
- this overrides the default setting, which can lead to imbalance in the
- cluster, especially if there are OSDs of different sizes in use. By
- default, the initial crush weight for the newly added osd is set to its
- volume size in TB. Leave this option unset to use the default provided
- by Ceph itself. This option only affects NEW OSDs, not existing ones.
- source: unset
- type: float
- customize-failure-domain:
- default: false
- description: |
- Setting this to true will tell Ceph to replicate across Juju's
- Availability Zone instead of specifically by host.
- source: default
- type: boolean
- value: false
- ephemeral-unmount:
- description: |
- Cloud instances provide ephemeral storage which is normally mounted
- on /mnt.
- .
- Setting this option to the path of the ephemeral mountpoint will force
- an unmount of the corresponding device so that it can be used as a OSD
- storage device. This is useful for testing purposes (cloud deployment
- is not a typical use case).
- source: unset
- type: string
- harden:
- description: |
- Apply system hardening. Supports a space-delimited list of modules
- to run. Supported modules currently include os, ssh, apache and mysql.
- source: unset
- type: string
- ignore-device-errors:
- default: false
- description: |
- By default, the charm will raise errors if a whitelisted device is found,
- but for some reason the charm is unable to initialize the device for use
- by Ceph.
- .
- Setting this option to 'True' will result in the charm classifying such
- problems as warnings only and will not result in a hook error.
- source: default
- type: boolean
- value: false
- key:
- description: "Key ID to import to the apt keyring to support use with arbitrary
- source\nconfiguration from outside of Launchpad archives or PPA's.\nThe accepted
- formats should be a GPG key in ASCII armor format, \nincluding BEGIN and END
- markers or a keyid.\n"
- source: unset
- type: string
- loglevel:
- default: 1
- description: OSD debug level. Max is 20.
- source: default
- type: int
- value: 1
- max-sectors-kb:
- default: 1.048576e+06
- description: |
- This parameter will adjust every block device in your server to allow
- greater IO operation sizes. If you have a RAID card with cache on it
- consider tuning this much higher than the 1MB default. 1MB is a safe
- default for spinning HDDs that don't have much cache.
- source: default
- type: int
- value: 1.048576e+06
- nagios_context:
- default: juju
- description: |
- Used by the nrpe-external-master subordinate charm.
- A string that will be prepended to instance name to set the hostname
- in nagios. So for instance the hostname would be something like:
- .
- juju-myservice-0
- .
- If you're running multiple environments with the same services in them
- this allows you to differentiate between them.
- source: default
- type: string
- value: juju
- nagios_servicegroups:
- default: ""
- description: |
- A comma-separated list of nagios servicegroups.
- If left empty, the nagios_context will be used as the servicegroup
- source: default
- type: string
- value: ""
- osd-devices:
- default: /dev/vdb
- description: |
- The devices to format and set up as OSD volumes.
- .
- These devices are the range of devices that will be checked for and
- used across all service units, in addition to any volumes attached
- via the --storage flag during deployment.
- .
- For ceph < 14.2.0 (Nautilus) these can also be directories instead of
- devices. If the value does not start with "/dev" then it will be
- interpreted as a directory.
- source: user
- type: string
- value: /dev/sdb
- osd-encrypt:
- default: false
- description: |
- By default, the charm will not encrypt Ceph OSD devices; however, by
- setting osd-encrypt to True, Ceph's dmcrypt support will be used to
- encrypt OSD devices.
- .
- Specifying this option on a running Ceph OSD node will have no effect
- until new disks are added, at which point new disks will be encrypted.
- source: default
- type: boolean
- value: false
- osd-encrypt-keymanager:
- default: ceph
- description: |
- Keymanager to use for storage of dm-crypt keys used for OSD devices;
- by default 'ceph' itself will be used for storage of keys, making use
- of the key/value storage provided by the ceph-mon cluster.
- .
- Alternatively 'vault' may be used for storage of dm-crypt keys. Both
- approaches ensure that keys are never written to the local filesystem.
- This also requires a relation to the vault charm.
- source: default
- type: string
- value: ceph
- osd-format:
- default: xfs
- description: |
- Format of filesystem to use for OSD devices. Supported formats include:
- .
- xfs (Default with >= ceph 0.48.3)
- ext4 (Only option < ceph 0.48.3)
- btrfs (experimental and not recommended)
- .
- Only supported with >= ceph 0.48.3.
- .
- Used with FileStore storage backend.
- .
- Always applies prior to ceph 12.2.0. Otherwise, only applies when the
- "bluestore" option is False.
- source: default
- type: string
- value: xfs
- osd-journal:
- description: |
- The device to use as a shared journal drive for all OSDs on a node. By
- default a journal partition will be created on each OSD volume device for
- use by that OSD. The default behaviour is also the fallback for the case
- where the specified journal device does not exist on a node.
- .
- Only supported with ceph >= 0.48.3.
- source: unset
- type: string
- osd-journal-size:
- default: 1024
- description: |
- Ceph OSD journal size. The journal size should be at least twice the
- product of the expected drive speed multiplied by filestore max sync
- interval. However, the most common practice is to partition the journal
- drive (often an SSD), and mount it such that Ceph uses the entire
- partition for the journal.
- .
- Only supported with ceph >= 0.48.3.
- source: default
- type: int
- value: 1024
- osd-max-backfills:
- description: |
- The maximum number of backfills allowed to or from a single OSD.
- .
- Setting this option on a running Ceph OSD node will not affect running
- OSD devices, but will add the setting to ceph.conf for the next restart.
- source: unset
- type: int
- osd-recovery-max-active:
- description: |
- The number of active recovery requests per OSD at one time. More requests
- will accelerate recovery, but the requests places an increased load on the
- cluster.
- .
- Setting this option on a running Ceph OSD node will not affect running
- OSD devices, but will add the setting to ceph.conf for the next restart.
- source: unset
- type: int
- prefer-ipv6:
- default: false
- description: |
- If True enables IPv6 support. The charm will expect network interfaces
- to be configured with an IPv6 address. If set to False (default) IPv4
- is expected.
- .
- NOTE: these charms do not currently support IPv6 privacy extension. In
- order for this charm to function correctly, the privacy extension must be
- disabled and a non-temporary address must be configured/available on
- your network interface.
- source: default
- type: boolean
- value: false
- source:
- description: |
- Optional configuration to support use of additional sources such as:
- .
- - ppa:myteam/ppa
- - cloud:bionic-ussuri
- - cloud:xenial-proposed/queens
- - http://my.archive.com/ubuntu main
- .
- The last option should be used in conjunction with the key configuration
- option.
- source: user
- type: string
- value: cloud:focal-xena
- sysctl:
- default: '{ kernel.pid_max : 2097152, vm.max_map_count : 524288, kernel.threads-max:
- 2097152 }'
- description: |
- YAML-formatted associative array of sysctl key/value pairs to be set
- persistently. By default we set pid_max, max_map_count and
- threads-max to a high value to avoid problems with large numbers (>20)
- of OSDs recovering. very large clusters should set those values even
- higher (e.g. max for kernel.pid_max is 4194303).
- source: default
- type: string
- value: '{ kernel.pid_max : 2097152, vm.max_map_count : 524288, kernel.threads-max:
- 2097152 }'
- use-direct-io:
- default: true
- description: Configure use of direct IO for OSD journals.
- source: default
- type: boolean
- value: true
- use-syslog:
- default: false
- description: |
- If set to True, supporting services will log to syslog.
- source: default
- type: boolean
- value: false
Advertisement
Add Comment
Please, Sign In to add comment