X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fsrc%2Fsample.ceph.conf;fp=src%2Fceph%2Fsrc%2Fsample.ceph.conf;h=c8affbad704c3b3cbc819675e418740a8d7b791f;hb=812ff6ca9fcd3e629e49d4328905f33eee8ca3f5;hp=0000000000000000000000000000000000000000;hpb=15280273faafb77777eab341909a3f495cf248d9;p=stor4nfv.git diff --git a/src/ceph/src/sample.ceph.conf b/src/ceph/src/sample.ceph.conf new file mode 100644 index 0000000..c8affba --- /dev/null +++ b/src/ceph/src/sample.ceph.conf @@ -0,0 +1,536 @@ +## +# Sample ceph ceph.conf file. +## +# This file defines cluster membership, the various locations +# that Ceph stores data, and any other runtime options. + +# If a 'host' is defined for a daemon, the init.d start/stop script will +# verify that it matches the hostname (or else ignore it). If it is +# not defined, it is assumed that the daemon is intended to start on +# the current host (e.g., in a setup with a startup.conf on each +# node). + +## Metavariables +# $cluster ; Expands to the Ceph Storage Cluster name. Useful +# ; when running multiple Ceph Storage Clusters +# ; on the same hardware. +# ; Example: /etc/ceph/$cluster.keyring +# ; (Default: ceph) +# +# $type ; Expands to one of mds, osd, or mon, depending on +# ; the type of the instant daemon. +# ; Example: /var/lib/ceph/$type +# +# $id ; Expands to the daemon identifier. For osd.0, this +# ; would be 0; for mds.a, it would be a. +# ; Example: /var/lib/ceph/$type/$cluster-$id +# +# $host ; Expands to the host name of the instant daemon. +# +# $name ; Expands to $type.$id. +# ; Example: /var/run/ceph/$cluster-$name.asok + +[global] +### http://docs.ceph.com/docs/master/rados/configuration/general-config-ref/ + + ;fsid = {UUID} # use `uuidgen` to generate your own UUID + ;public network = 192.168.0.0/24 + ;cluster network = 192.168.0.0/24 + + # Each running Ceph daemon has a running process identifier (PID) file. + # The PID file is generated upon start-up. + # Type: String (optional) + # (Default: N/A). The default path is /var/run/$cluster/$name.pid. + pid file = /var/run/ceph/$name.pid + + # If set, when the Ceph Storage Cluster starts, Ceph sets the max open fds + # at the OS level (i.e., the max # of file descriptors). + # It helps prevents Ceph OSD Daemons from running out of file descriptors. + # Type: 64-bit Integer (optional) + # (Default: 0) + ;max open files = 131072 + + +### http://docs.ceph.com/docs/master/rados/operations/ +### http://docs.ceph.com/docs/master/rados/configuration/auth-config-ref/ + + # If enabled, the Ceph Storage Cluster daemons (i.e., ceph-mon, ceph-osd, + # and ceph-mds) must authenticate with each other. + # Type: String (optional); Valid settings are "cephx" or "none". + # (Default: cephx) + auth cluster required = cephx + + # If enabled, the Ceph Storage Cluster daemons require Ceph Clients to + # authenticate with the Ceph Storage Cluster in order to access Ceph + # services. + # Type: String (optional); Valid settings are "cephx" or "none". + # (Default: cephx) + auth service required = cephx + + # If enabled, the Ceph Client requires the Ceph Storage Cluster to + # authenticate with the Ceph Client. + # Type: String (optional); Valid settings are "cephx" or "none". + # (Default: cephx) + auth client required = cephx + + # If set to true, Ceph requires signatures on all message traffic between + # the Ceph Client and the Ceph Storage Cluster, and between daemons + # comprising the Ceph Storage Cluster. + # Type: Boolean (optional) + # (Default: false) + ;cephx require signatures = true + + # kernel RBD client do not support authentication yet: + cephx cluster require signatures = true + cephx service require signatures = false + + # The path to the keyring file. + # Type: String (optional) + # Default: /etc/ceph/$cluster.$name.keyring,/etc/ceph/$cluster.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin + ;keyring = /etc/ceph/$cluster.$name.keyring + + +### http://docs.ceph.com/docs/master/rados/configuration/pool-pg-config-ref/ + + + ## Replication level, number of data copies. + # Type: 32-bit Integer + # (Default: 3) + ;osd pool default size = 3 + + ## Replication level in degraded state, less than 'osd pool default size' value. + # Sets the minimum number of written replicas for objects in the + # pool in order to acknowledge a write operation to the client. If + # minimum is not met, Ceph will not acknowledge the write to the + # client. This setting ensures a minimum number of replicas when + # operating in degraded mode. + # Type: 32-bit Integer + # (Default: 0), which means no particular minimum. If 0, minimum is size - (size / 2). + ;osd pool default min size = 2 + + ## Ensure you have a realistic number of placement groups. We recommend + ## approximately 100 per OSD. E.g., total number of OSDs multiplied by 100 + ## divided by the number of replicas (i.e., osd pool default size). So for + ## 10 OSDs and osd pool default size = 3, we'd recommend approximately + ## (100 * 10) / 3 = 333 + + # Description: The default number of placement groups for a pool. The + # default value is the same as pg_num with mkpool. + # Type: 32-bit Integer + # (Default: 8) + ;osd pool default pg num = 128 + + # Description: The default number of placement groups for placement for a + # pool. The default value is the same as pgp_num with mkpool. + # PG and PGP should be equal (for now). + # Type: 32-bit Integer + # (Default: 8) + ;osd pool default pgp num = 128 + + # The default CRUSH ruleset to use when creating a pool + # Type: 32-bit Integer + # (Default: 0) + ;osd pool default crush rule = 0 + + # The bucket type to use for chooseleaf in a CRUSH rule. + # Uses ordinal rank rather than name. + # Type: 32-bit Integer + # (Default: 1) Typically a host containing one or more Ceph OSD Daemons. + ;osd crush chooseleaf type = 1 + + +### http://docs.ceph.com/docs/master/rados/troubleshooting/log-and-debug/ + + # The location of the logging file for your cluster. + # Type: String + # Required: No + # Default: /var/log/ceph/$cluster-$name.log + ;log file = /var/log/ceph/$cluster-$name.log + + # Determines if logging messages should appear in syslog. + # Type: Boolean + # Required: No + # (Default: false) + ;log to syslog = true + + +### http://docs.ceph.com/docs/master/rados/configuration/ms-ref/ + + # Enable if you want your daemons to bind to IPv6 address instead of + # IPv4 ones. (Not required if you specify a daemon or cluster IP.) + # Type: Boolean + # (Default: false) + ;ms bind ipv6 = true + +################## +## Monitors +## You need at least one. You need at least three if you want to +## tolerate any node failures. Always create an odd number. +[mon] +### http://docs.ceph.com/docs/master/rados/configuration/mon-config-ref/ +### http://docs.ceph.com/docs/master/rados/configuration/mon-osd-interaction/ + + # The IDs of initial monitors in a cluster during startup. + # If specified, Ceph requires an odd number of monitors to form an + # initial quorum (e.g., 3). + # Type: String + # (Default: None) + ;mon initial members = mycephhost + + ;mon host = cephhost01,cephhost02 + ;mon addr = 192.168.0.101,192.168.0.102 + + # The monitor's data location + # Default: /var/lib/ceph/mon/$cluster-$id + ;mon data = /var/lib/ceph/mon/$name + + # The clock drift in seconds allowed between monitors. + # Type: Float + # (Default: .050) + ;mon clock drift allowed = .15 + + # Exponential backoff for clock drift warnings + # Type: Float + # (Default: 5) + ;mon clock drift warn backoff = 30 # Tell the monitor to backoff from this warning for 30 seconds + + # The percentage of disk space used before an OSD is considered full. + # Type: Float + # (Default: .95) + ;mon osd full ratio = .95 + + # The percentage of disk space used before an OSD is considered nearfull. + # Type: Float + # (Default: .85) + ;mon osd nearfull ratio = .85 + + # The number of seconds Ceph waits before marking a Ceph OSD + # Daemon "down" and "out" if it doesn't respond. + # Type: 32-bit Integer + # (Default: 600) + ;mon osd down out interval = 600 + + # The grace period in seconds before declaring unresponsive Ceph OSD + # Daemons "down". + # Type: 32-bit Integer + # (Default: 900) + ;mon osd report timeout = 300 + +### http://docs.ceph.com/docs/master/rados/troubleshooting/log-and-debug/ + + # logging, for debugging monitor crashes, in order of + # their likelihood of being helpful :) + ;debug ms = 1 + ;debug mon = 20 + ;debug paxos = 20 + ;debug auth = 20 + + +;[mon.alpha] +; host = alpha +; mon addr = 192.168.0.10:6789 + +;[mon.beta] +; host = beta +; mon addr = 192.168.0.11:6789 + +;[mon.gamma] +; host = gamma +; mon addr = 192.168.0.12:6789 + + +################## +## Metadata servers +# You must deploy at least one metadata server to use CephFS. There is +# experimental support for running multiple metadata servers. Do not run +# multiple metadata servers in production. +[mds] +### http://docs.ceph.com/docs/master/cephfs/mds-config-ref/ + + # where the mds keeps it's secret encryption keys + ;keyring = /var/lib/ceph/mds/$name/keyring + + # Determines whether a 'ceph-mds' daemon should poll and + # replay the log of an active MDS (hot standby). + # Type: Boolean + # (Default: false) + ;mds standby replay = true + + ; mds logging to debug issues. + ;debug ms = 1 + ;debug mds = 20 + ;debug journaler = 20 + + # The number of inodes to cache. + # Type: 32-bit Integer + # (Default: 100000) + ;mds cache size = 250000 + +;[mds.alpha] +; host = alpha + +;[mds.beta] +; host = beta + +################## +## osd +# You need at least one. Two or more if you want data to be replicated. +# Define as many as you like. +[osd] +### http://docs.ceph.com/docs/master/rados/configuration/osd-config-ref/ + + # The path to the OSDs data. + # You must create the directory when deploying Ceph. + # You should mount a drive for OSD data at this mount point. + # We do not recommend changing the default. + # Type: String + # Default: /var/lib/ceph/osd/$cluster-$id + ;osd data = /var/lib/ceph/osd/$name + + ## You can change the number of recovery operations to speed up recovery + ## or slow it down if your machines can't handle it + + # The number of active recovery requests per OSD at one time. + # More requests will accelerate recovery, but the requests + # places an increased load on the cluster. + # Type: 32-bit Integer + # (Default: 5) + ;osd recovery max active = 3 + + # The maximum number of backfills allowed to or from a single OSD. + # Type: 64-bit Integer + # (Default: 10) + ;osd max backfills = 5 + + # The maximum number of simultaneous scrub operations for a Ceph OSD Daemon. + # Type: 32-bit Int + # (Default: 1) + ;osd max scrubs = 2 + + # You may add settings for ceph-deploy so that it will create and mount + # the correct type of file system. Remove the comment `#` character for + # the following settings and replace the values in parenthesis + # with appropriate values, or leave the following settings commented + # out to accept the default values. + + #osd mkfs type = {fs-type} + #osd mkfs options {fs-type} = {mkfs options} # default for xfs is "-f" + #osd mount options {fs-type} = {mount options} # default mount option is "rw, noatime" + ;osd mkfs type = btrfs + ;osd mount options btrfs = noatime,nodiratime + + ## Ideally, make this a separate disk or partition. A few + ## hundred MB should be enough; more if you have fast or many + ## disks. You can use a file under the osd data dir if need be + ## (e.g. /data/$name/journal), but it will be slower than a + ## separate disk or partition. + # The path to the OSD's journal. This may be a path to a file or a block + # device (such as a partition of an SSD). If it is a file, you must + # create the directory to contain it. + # We recommend using a drive separate from the osd data drive. + # Type: String + # Default: /var/lib/ceph/osd/$cluster-$id/journal + ;osd journal = /var/lib/ceph/osd/$name/journal + + # Check log files for corruption. Can be computationally expensive. + # Type: Boolean + # (Default: false) + ;osd check for log corruption = true + +### http://docs.ceph.com/docs/master/rados/configuration/journal-ref/ + + # The size of the journal in megabytes. If this is 0, + # and the journal is a block device, the entire block device is used. + # Since v0.54, this is ignored if the journal is a block device, + # and the entire block device is used. + # Type: 32-bit Integer + # (Default: 5120) + # Recommended: Begin with 1GB. Should be at least twice the product + # of the expected speed multiplied by "filestore max sync interval". + ;osd journal size = 2048 ; journal size, in megabytes + + ## If you want to run the journal on a tmpfs, disable DirectIO + # Enables direct i/o to the journal. + # Requires "journal block align" set to "true". + # Type: Boolean + # Required: Yes when using aio. + # (Default: true) + ;journal dio = false + + # osd logging to debug osd issues, in order of likelihood of being helpful + ;debug ms = 1 + ;debug osd = 20 + ;debug filestore = 20 + ;debug journal = 20 + +### http://docs.ceph.com/docs/master/rados/configuration/filestore-config-ref/ + + # The maximum interval in seconds for synchronizing the filestore. + # Type: Double (optional) + # (Default: 5) + ;filestore max sync interval = 5 + + # Enable snapshots for a btrfs filestore. + # Type: Boolean + # Required: No. Only used for btrfs. + # (Default: true) + ;filestore btrfs snap = false + + # Enables the filestore flusher. + # Type: Boolean + # Required: No + # (Default: false) + ;filestore flusher = true + + # Defines the maximum number of in progress operations the file store + # accepts before blocking on queuing new operations. + # Type: Integer + # Required: No. Minimal impact on performance. + # (Default: 500) + ;filestore queue max ops = 500 + + ## Filestore and OSD settings can be tweak to achieve better performance + +### http://docs.ceph.com/docs/master/rados/configuration/filestore-config-ref/#misc + + # Min number of files in a subdir before merging into parent NOTE: A negative value means to disable subdir merging + # Type: Integer + # Required: No + # Default: 10 + ;filestore merge threshold = 10 + + # filestore_split_multiple * abs(filestore_merge_threshold) * 16 is the maximum number of files in a subdirectory before splitting into child directories. + # Type: Integer + # Required: No + # Default: 2 + ;filestore split multiple = 2 + + # The number of filesystem operation threads that execute in parallel. + # Type: Integer + # Required: No + # Default: 2 + ;filestore op threads = 4 + + # The number of threads to service Ceph OSD Daemon operations. Set to 0 to disable it. Increasing the number may increase the request processing rate. + # Type: 32-bit Integer + # Default: 2 + ;osd op threads = 2 + + ## CRUSH + + # By default OSDs update their details (location, weight and root) on the CRUSH map during startup + # Type: Boolean + # Required: No; + # (Default: true) + ;osd crush update on start = false + +;[osd.0] +; host = delta + +;[osd.1] +; host = epsilon + +;[osd.2] +; host = zeta + +;[osd.3] +; host = eta + + +################## +## client settings +[client] + +### http://docs.ceph.com/docs/master/rbd/rbd-config-ref/ + + # Enable caching for RADOS Block Device (RBD). + # Type: Boolean + # Required: No + # (Default: true) + rbd cache = true + + # The RBD cache size in bytes. + # Type: 64-bit Integer + # Required: No + # (Default: 32 MiB) + ;rbd cache size = 33554432 + + # The dirty limit in bytes at which the cache triggers write-back. + # If 0, uses write-through caching. + # Type: 64-bit Integer + # Required: No + # Constraint: Must be less than rbd cache size. + # (Default: 24 MiB) + ;rbd cache max dirty = 25165824 + + # The dirty target before the cache begins writing data to the data storage. + # Does not block writes to the cache. + # Type: 64-bit Integer + # Required: No + # Constraint: Must be less than rbd cache max dirty. + # (Default: 16 MiB) + ;rbd cache target dirty = 16777216 + + # The number of seconds dirty data is in the cache before writeback starts. + # Type: Float + # Required: No + # (Default: 1.0) + ;rbd cache max dirty age = 1.0 + + # Start out in write-through mode, and switch to write-back after the + # first flush request is received. Enabling this is a conservative but + # safe setting in case VMs running on rbd are too old to send flushes, + # like the virtio driver in Linux before 2.6.32. + # Type: Boolean + # Required: No + # (Default: true) + ;rbd cache writethrough until flush = true + + # The Ceph admin socket allows you to query a daemon via a socket interface + # From a client perspective this can be a virtual machine using librbd + # Type: String + # Required: No + ;admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok + + +################## +## radosgw client settings +[client.radosgw.gateway] + +### http://docs.ceph.com/docs/master/radosgw/config-ref/ + + # Sets the location of the data files for Ceph Object Gateway. + # You must create the directory when deploying Ceph. + # We do not recommend changing the default. + # Type: String + # Default: /var/lib/ceph/radosgw/$cluster-$id + ;rgw data = /var/lib/ceph/radosgw/$name + + # Client's hostname + ;host = ceph-radosgw + + # where the radosgw keeps it's secret encryption keys + ;keyring = /etc/ceph/ceph.client.radosgw.keyring + + # FastCgiExternalServer uses this socket. + # If you do not specify a socket path, Ceph Object Gateway will not run as an external server. + # The path you specify here must be the same as the path specified in the rgw.conf file. + # Type: String + # Default: None + ;rgw socket path = /var/run/ceph/ceph.radosgw.gateway.fastcgi.sock + + # The location of the logging file for your radosgw. + # Type: String + # Required: No + # Default: /var/log/ceph/$cluster-$name.log + ;log file = /var/log/ceph/client.radosgw.gateway.log + + # Enable 100-continue if it is operational. + # Type: Boolean + # Default: true + ;rgw print continue = false + + # The DNS name of the served domain. + # Type: String + # Default: None + ;rgw dns name = radosgw.ceph.internal