X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fsrc%2Fsample.ceph.conf;fp=src%2Fceph%2Fsrc%2Fsample.ceph.conf;h=0000000000000000000000000000000000000000;hb=7da45d65be36d36b880cc55c5036e96c24b53f00;hp=c8affbad704c3b3cbc819675e418740a8d7b791f;hpb=691462d09d0987b47e112d6ee8740375df3c51b2;p=stor4nfv.git diff --git a/src/ceph/src/sample.ceph.conf b/src/ceph/src/sample.ceph.conf deleted file mode 100644 index c8affba..0000000 --- a/src/ceph/src/sample.ceph.conf +++ /dev/null @@ -1,536 +0,0 @@ -## -# Sample ceph ceph.conf file. -## -# This file defines cluster membership, the various locations -# that Ceph stores data, and any other runtime options. - -# If a 'host' is defined for a daemon, the init.d start/stop script will -# verify that it matches the hostname (or else ignore it). If it is -# not defined, it is assumed that the daemon is intended to start on -# the current host (e.g., in a setup with a startup.conf on each -# node). - -## Metavariables -# $cluster ; Expands to the Ceph Storage Cluster name. Useful -# ; when running multiple Ceph Storage Clusters -# ; on the same hardware. -# ; Example: /etc/ceph/$cluster.keyring -# ; (Default: ceph) -# -# $type ; Expands to one of mds, osd, or mon, depending on -# ; the type of the instant daemon. -# ; Example: /var/lib/ceph/$type -# -# $id ; Expands to the daemon identifier. For osd.0, this -# ; would be 0; for mds.a, it would be a. -# ; Example: /var/lib/ceph/$type/$cluster-$id -# -# $host ; Expands to the host name of the instant daemon. -# -# $name ; Expands to $type.$id. -# ; Example: /var/run/ceph/$cluster-$name.asok - -[global] -### http://docs.ceph.com/docs/master/rados/configuration/general-config-ref/ - - ;fsid = {UUID} # use `uuidgen` to generate your own UUID - ;public network = 192.168.0.0/24 - ;cluster network = 192.168.0.0/24 - - # Each running Ceph daemon has a running process identifier (PID) file. - # The PID file is generated upon start-up. - # Type: String (optional) - # (Default: N/A). The default path is /var/run/$cluster/$name.pid. - pid file = /var/run/ceph/$name.pid - - # If set, when the Ceph Storage Cluster starts, Ceph sets the max open fds - # at the OS level (i.e., the max # of file descriptors). - # It helps prevents Ceph OSD Daemons from running out of file descriptors. - # Type: 64-bit Integer (optional) - # (Default: 0) - ;max open files = 131072 - - -### http://docs.ceph.com/docs/master/rados/operations/ -### http://docs.ceph.com/docs/master/rados/configuration/auth-config-ref/ - - # If enabled, the Ceph Storage Cluster daemons (i.e., ceph-mon, ceph-osd, - # and ceph-mds) must authenticate with each other. - # Type: String (optional); Valid settings are "cephx" or "none". - # (Default: cephx) - auth cluster required = cephx - - # If enabled, the Ceph Storage Cluster daemons require Ceph Clients to - # authenticate with the Ceph Storage Cluster in order to access Ceph - # services. - # Type: String (optional); Valid settings are "cephx" or "none". - # (Default: cephx) - auth service required = cephx - - # If enabled, the Ceph Client requires the Ceph Storage Cluster to - # authenticate with the Ceph Client. - # Type: String (optional); Valid settings are "cephx" or "none". - # (Default: cephx) - auth client required = cephx - - # If set to true, Ceph requires signatures on all message traffic between - # the Ceph Client and the Ceph Storage Cluster, and between daemons - # comprising the Ceph Storage Cluster. - # Type: Boolean (optional) - # (Default: false) - ;cephx require signatures = true - - # kernel RBD client do not support authentication yet: - cephx cluster require signatures = true - cephx service require signatures = false - - # The path to the keyring file. - # Type: String (optional) - # Default: /etc/ceph/$cluster.$name.keyring,/etc/ceph/$cluster.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin - ;keyring = /etc/ceph/$cluster.$name.keyring - - -### http://docs.ceph.com/docs/master/rados/configuration/pool-pg-config-ref/ - - - ## Replication level, number of data copies. - # Type: 32-bit Integer - # (Default: 3) - ;osd pool default size = 3 - - ## Replication level in degraded state, less than 'osd pool default size' value. - # Sets the minimum number of written replicas for objects in the - # pool in order to acknowledge a write operation to the client. If - # minimum is not met, Ceph will not acknowledge the write to the - # client. This setting ensures a minimum number of replicas when - # operating in degraded mode. - # Type: 32-bit Integer - # (Default: 0), which means no particular minimum. If 0, minimum is size - (size / 2). - ;osd pool default min size = 2 - - ## Ensure you have a realistic number of placement groups. We recommend - ## approximately 100 per OSD. E.g., total number of OSDs multiplied by 100 - ## divided by the number of replicas (i.e., osd pool default size). So for - ## 10 OSDs and osd pool default size = 3, we'd recommend approximately - ## (100 * 10) / 3 = 333 - - # Description: The default number of placement groups for a pool. The - # default value is the same as pg_num with mkpool. - # Type: 32-bit Integer - # (Default: 8) - ;osd pool default pg num = 128 - - # Description: The default number of placement groups for placement for a - # pool. The default value is the same as pgp_num with mkpool. - # PG and PGP should be equal (for now). - # Type: 32-bit Integer - # (Default: 8) - ;osd pool default pgp num = 128 - - # The default CRUSH ruleset to use when creating a pool - # Type: 32-bit Integer - # (Default: 0) - ;osd pool default crush rule = 0 - - # The bucket type to use for chooseleaf in a CRUSH rule. - # Uses ordinal rank rather than name. - # Type: 32-bit Integer - # (Default: 1) Typically a host containing one or more Ceph OSD Daemons. - ;osd crush chooseleaf type = 1 - - -### http://docs.ceph.com/docs/master/rados/troubleshooting/log-and-debug/ - - # The location of the logging file for your cluster. - # Type: String - # Required: No - # Default: /var/log/ceph/$cluster-$name.log - ;log file = /var/log/ceph/$cluster-$name.log - - # Determines if logging messages should appear in syslog. - # Type: Boolean - # Required: No - # (Default: false) - ;log to syslog = true - - -### http://docs.ceph.com/docs/master/rados/configuration/ms-ref/ - - # Enable if you want your daemons to bind to IPv6 address instead of - # IPv4 ones. (Not required if you specify a daemon or cluster IP.) - # Type: Boolean - # (Default: false) - ;ms bind ipv6 = true - -################## -## Monitors -## You need at least one. You need at least three if you want to -## tolerate any node failures. Always create an odd number. -[mon] -### http://docs.ceph.com/docs/master/rados/configuration/mon-config-ref/ -### http://docs.ceph.com/docs/master/rados/configuration/mon-osd-interaction/ - - # The IDs of initial monitors in a cluster during startup. - # If specified, Ceph requires an odd number of monitors to form an - # initial quorum (e.g., 3). - # Type: String - # (Default: None) - ;mon initial members = mycephhost - - ;mon host = cephhost01,cephhost02 - ;mon addr = 192.168.0.101,192.168.0.102 - - # The monitor's data location - # Default: /var/lib/ceph/mon/$cluster-$id - ;mon data = /var/lib/ceph/mon/$name - - # The clock drift in seconds allowed between monitors. - # Type: Float - # (Default: .050) - ;mon clock drift allowed = .15 - - # Exponential backoff for clock drift warnings - # Type: Float - # (Default: 5) - ;mon clock drift warn backoff = 30 # Tell the monitor to backoff from this warning for 30 seconds - - # The percentage of disk space used before an OSD is considered full. - # Type: Float - # (Default: .95) - ;mon osd full ratio = .95 - - # The percentage of disk space used before an OSD is considered nearfull. - # Type: Float - # (Default: .85) - ;mon osd nearfull ratio = .85 - - # The number of seconds Ceph waits before marking a Ceph OSD - # Daemon "down" and "out" if it doesn't respond. - # Type: 32-bit Integer - # (Default: 600) - ;mon osd down out interval = 600 - - # The grace period in seconds before declaring unresponsive Ceph OSD - # Daemons "down". - # Type: 32-bit Integer - # (Default: 900) - ;mon osd report timeout = 300 - -### http://docs.ceph.com/docs/master/rados/troubleshooting/log-and-debug/ - - # logging, for debugging monitor crashes, in order of - # their likelihood of being helpful :) - ;debug ms = 1 - ;debug mon = 20 - ;debug paxos = 20 - ;debug auth = 20 - - -;[mon.alpha] -; host = alpha -; mon addr = 192.168.0.10:6789 - -;[mon.beta] -; host = beta -; mon addr = 192.168.0.11:6789 - -;[mon.gamma] -; host = gamma -; mon addr = 192.168.0.12:6789 - - -################## -## Metadata servers -# You must deploy at least one metadata server to use CephFS. There is -# experimental support for running multiple metadata servers. Do not run -# multiple metadata servers in production. -[mds] -### http://docs.ceph.com/docs/master/cephfs/mds-config-ref/ - - # where the mds keeps it's secret encryption keys - ;keyring = /var/lib/ceph/mds/$name/keyring - - # Determines whether a 'ceph-mds' daemon should poll and - # replay the log of an active MDS (hot standby). - # Type: Boolean - # (Default: false) - ;mds standby replay = true - - ; mds logging to debug issues. - ;debug ms = 1 - ;debug mds = 20 - ;debug journaler = 20 - - # The number of inodes to cache. - # Type: 32-bit Integer - # (Default: 100000) - ;mds cache size = 250000 - -;[mds.alpha] -; host = alpha - -;[mds.beta] -; host = beta - -################## -## osd -# You need at least one. Two or more if you want data to be replicated. -# Define as many as you like. -[osd] -### http://docs.ceph.com/docs/master/rados/configuration/osd-config-ref/ - - # The path to the OSDs data. - # You must create the directory when deploying Ceph. - # You should mount a drive for OSD data at this mount point. - # We do not recommend changing the default. - # Type: String - # Default: /var/lib/ceph/osd/$cluster-$id - ;osd data = /var/lib/ceph/osd/$name - - ## You can change the number of recovery operations to speed up recovery - ## or slow it down if your machines can't handle it - - # The number of active recovery requests per OSD at one time. - # More requests will accelerate recovery, but the requests - # places an increased load on the cluster. - # Type: 32-bit Integer - # (Default: 5) - ;osd recovery max active = 3 - - # The maximum number of backfills allowed to or from a single OSD. - # Type: 64-bit Integer - # (Default: 10) - ;osd max backfills = 5 - - # The maximum number of simultaneous scrub operations for a Ceph OSD Daemon. - # Type: 32-bit Int - # (Default: 1) - ;osd max scrubs = 2 - - # You may add settings for ceph-deploy so that it will create and mount - # the correct type of file system. Remove the comment `#` character for - # the following settings and replace the values in parenthesis - # with appropriate values, or leave the following settings commented - # out to accept the default values. - - #osd mkfs type = {fs-type} - #osd mkfs options {fs-type} = {mkfs options} # default for xfs is "-f" - #osd mount options {fs-type} = {mount options} # default mount option is "rw, noatime" - ;osd mkfs type = btrfs - ;osd mount options btrfs = noatime,nodiratime - - ## Ideally, make this a separate disk or partition. A few - ## hundred MB should be enough; more if you have fast or many - ## disks. You can use a file under the osd data dir if need be - ## (e.g. /data/$name/journal), but it will be slower than a - ## separate disk or partition. - # The path to the OSD's journal. This may be a path to a file or a block - # device (such as a partition of an SSD). If it is a file, you must - # create the directory to contain it. - # We recommend using a drive separate from the osd data drive. - # Type: String - # Default: /var/lib/ceph/osd/$cluster-$id/journal - ;osd journal = /var/lib/ceph/osd/$name/journal - - # Check log files for corruption. Can be computationally expensive. - # Type: Boolean - # (Default: false) - ;osd check for log corruption = true - -### http://docs.ceph.com/docs/master/rados/configuration/journal-ref/ - - # The size of the journal in megabytes. If this is 0, - # and the journal is a block device, the entire block device is used. - # Since v0.54, this is ignored if the journal is a block device, - # and the entire block device is used. - # Type: 32-bit Integer - # (Default: 5120) - # Recommended: Begin with 1GB. Should be at least twice the product - # of the expected speed multiplied by "filestore max sync interval". - ;osd journal size = 2048 ; journal size, in megabytes - - ## If you want to run the journal on a tmpfs, disable DirectIO - # Enables direct i/o to the journal. - # Requires "journal block align" set to "true". - # Type: Boolean - # Required: Yes when using aio. - # (Default: true) - ;journal dio = false - - # osd logging to debug osd issues, in order of likelihood of being helpful - ;debug ms = 1 - ;debug osd = 20 - ;debug filestore = 20 - ;debug journal = 20 - -### http://docs.ceph.com/docs/master/rados/configuration/filestore-config-ref/ - - # The maximum interval in seconds for synchronizing the filestore. - # Type: Double (optional) - # (Default: 5) - ;filestore max sync interval = 5 - - # Enable snapshots for a btrfs filestore. - # Type: Boolean - # Required: No. Only used for btrfs. - # (Default: true) - ;filestore btrfs snap = false - - # Enables the filestore flusher. - # Type: Boolean - # Required: No - # (Default: false) - ;filestore flusher = true - - # Defines the maximum number of in progress operations the file store - # accepts before blocking on queuing new operations. - # Type: Integer - # Required: No. Minimal impact on performance. - # (Default: 500) - ;filestore queue max ops = 500 - - ## Filestore and OSD settings can be tweak to achieve better performance - -### http://docs.ceph.com/docs/master/rados/configuration/filestore-config-ref/#misc - - # Min number of files in a subdir before merging into parent NOTE: A negative value means to disable subdir merging - # Type: Integer - # Required: No - # Default: 10 - ;filestore merge threshold = 10 - - # filestore_split_multiple * abs(filestore_merge_threshold) * 16 is the maximum number of files in a subdirectory before splitting into child directories. - # Type: Integer - # Required: No - # Default: 2 - ;filestore split multiple = 2 - - # The number of filesystem operation threads that execute in parallel. - # Type: Integer - # Required: No - # Default: 2 - ;filestore op threads = 4 - - # The number of threads to service Ceph OSD Daemon operations. Set to 0 to disable it. Increasing the number may increase the request processing rate. - # Type: 32-bit Integer - # Default: 2 - ;osd op threads = 2 - - ## CRUSH - - # By default OSDs update their details (location, weight and root) on the CRUSH map during startup - # Type: Boolean - # Required: No; - # (Default: true) - ;osd crush update on start = false - -;[osd.0] -; host = delta - -;[osd.1] -; host = epsilon - -;[osd.2] -; host = zeta - -;[osd.3] -; host = eta - - -################## -## client settings -[client] - -### http://docs.ceph.com/docs/master/rbd/rbd-config-ref/ - - # Enable caching for RADOS Block Device (RBD). - # Type: Boolean - # Required: No - # (Default: true) - rbd cache = true - - # The RBD cache size in bytes. - # Type: 64-bit Integer - # Required: No - # (Default: 32 MiB) - ;rbd cache size = 33554432 - - # The dirty limit in bytes at which the cache triggers write-back. - # If 0, uses write-through caching. - # Type: 64-bit Integer - # Required: No - # Constraint: Must be less than rbd cache size. - # (Default: 24 MiB) - ;rbd cache max dirty = 25165824 - - # The dirty target before the cache begins writing data to the data storage. - # Does not block writes to the cache. - # Type: 64-bit Integer - # Required: No - # Constraint: Must be less than rbd cache max dirty. - # (Default: 16 MiB) - ;rbd cache target dirty = 16777216 - - # The number of seconds dirty data is in the cache before writeback starts. - # Type: Float - # Required: No - # (Default: 1.0) - ;rbd cache max dirty age = 1.0 - - # Start out in write-through mode, and switch to write-back after the - # first flush request is received. Enabling this is a conservative but - # safe setting in case VMs running on rbd are too old to send flushes, - # like the virtio driver in Linux before 2.6.32. - # Type: Boolean - # Required: No - # (Default: true) - ;rbd cache writethrough until flush = true - - # The Ceph admin socket allows you to query a daemon via a socket interface - # From a client perspective this can be a virtual machine using librbd - # Type: String - # Required: No - ;admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok - - -################## -## radosgw client settings -[client.radosgw.gateway] - -### http://docs.ceph.com/docs/master/radosgw/config-ref/ - - # Sets the location of the data files for Ceph Object Gateway. - # You must create the directory when deploying Ceph. - # We do not recommend changing the default. - # Type: String - # Default: /var/lib/ceph/radosgw/$cluster-$id - ;rgw data = /var/lib/ceph/radosgw/$name - - # Client's hostname - ;host = ceph-radosgw - - # where the radosgw keeps it's secret encryption keys - ;keyring = /etc/ceph/ceph.client.radosgw.keyring - - # FastCgiExternalServer uses this socket. - # If you do not specify a socket path, Ceph Object Gateway will not run as an external server. - # The path you specify here must be the same as the path specified in the rgw.conf file. - # Type: String - # Default: None - ;rgw socket path = /var/run/ceph/ceph.radosgw.gateway.fastcgi.sock - - # The location of the logging file for your radosgw. - # Type: String - # Required: No - # Default: /var/log/ceph/$cluster-$name.log - ;log file = /var/log/ceph/client.radosgw.gateway.log - - # Enable 100-continue if it is operational. - # Type: Boolean - # Default: true - ;rgw print continue = false - - # The DNS name of the served domain. - # Type: String - # Default: None - ;rgw dns name = radosgw.ceph.internal