X-Git-Url: https://gerrit.opnfv.org/gerrit/gitweb?a=blobdiff_plain;f=src%2Fceph%2Fdoc%2Fglossary.rst;fp=src%2Fceph%2Fdoc%2Fglossary.rst;h=82546408a6e251b6bf66bff66a68fc0b19accfea;hb=812ff6ca9fcd3e629e49d4328905f33eee8ca3f5;hp=0000000000000000000000000000000000000000;hpb=15280273faafb77777eab341909a3f495cf248d9;p=stor4nfv.git diff --git a/src/ceph/doc/glossary.rst b/src/ceph/doc/glossary.rst new file mode 100644 index 0000000..8254640 --- /dev/null +++ b/src/ceph/doc/glossary.rst @@ -0,0 +1,187 @@ +=============== + Ceph Glossary +=============== + +Ceph is growing rapidly. As firms deploy Ceph, the technical terms such as +"RADOS", "RBD," "RGW" and so forth require corresponding marketing terms +that explain what each component does. The terms in this glossary are +intended to complement the existing technical terminology. + +Sometimes more than one term applies to a definition. Generally, the first +term reflects a term consistent with Ceph's marketing, and secondary terms +reflect either technical terms or legacy ways of referring to Ceph systems. + + +.. glossary:: + + Ceph Project + The aggregate term for the people, software, mission and infrastructure + of Ceph. + + cephx + The Ceph authentication protocol. Cephx operates like Kerberos, but it + has no single point of failure. + + Ceph + Ceph Platform + All Ceph software, which includes any piece of code hosted at + `http://github.com/ceph`_. + + Ceph System + Ceph Stack + A collection of two or more components of Ceph. + + Ceph Node + Node + Host + Any single machine or server in a Ceph System. + + Ceph Storage Cluster + Ceph Object Store + RADOS + RADOS Cluster + Reliable Autonomic Distributed Object Store + The core set of storage software which stores the user's data (MON+OSD). + + Ceph Cluster Map + cluster map + The set of maps comprising the monitor map, OSD map, PG map, MDS map and + CRUSH map. See `Cluster Map`_ for details. + + Ceph Object Storage + The object storage "product", service or capabilities, which consists + essentially of a Ceph Storage Cluster and a Ceph Object Gateway. + + Ceph Object Gateway + RADOS Gateway + RGW + The S3/Swift gateway component of Ceph. + + Ceph Block Device + RBD + The block storage component of Ceph. + + Ceph Block Storage + The block storage "product," service or capabilities when used in + conjunction with ``librbd``, a hypervisor such as QEMU or Xen, and a + hypervisor abstraction layer such as ``libvirt``. + + Ceph Filesystem + CephFS + Ceph FS + The POSIX filesystem components of Ceph. + + Cloud Platforms + Cloud Stacks + Third party cloud provisioning platforms such as OpenStack, CloudStack, + OpenNebula, ProxMox, etc. + + Object Storage Device + OSD + A physical or logical storage unit (*e.g.*, LUN). + Sometimes, Ceph users use the + term "OSD" to refer to :term:`Ceph OSD Daemon`, though the + proper term is "Ceph OSD". + + Ceph OSD Daemon + Ceph OSD Daemons + Ceph OSD + The Ceph OSD software, which interacts with a logical + disk (:term:`OSD`). Sometimes, Ceph users use the + term "OSD" to refer to "Ceph OSD Daemon", though the + proper term is "Ceph OSD". + + OSD id + The integer that defines an OSD. It is generated by the monitors as part + of the creation of a new OSD. + + OSD fsid + This is a unique identifier used to further improve the uniqueness of an + OSD and it is found in the OSD path in a file called ``osd_fsid``. This + ``fsid`` term is used interchangeably with ``uuid`` + + OSD uuid + Just like the OSD fsid, this is the OSD unique identifer and is used + interchangeably with ``fsid`` + + bluestore + OSD BlueStore is a new back end for OSD daemons (kraken and newer + versions). Unlike :term:`filestore` it stores objects directly on the + Ceph block devices without any file system interface. + + filestore + A back end for OSD daemons, where a Journal is needed and files are + written to the filesystem. + + Ceph Monitor + MON + The Ceph monitor software. + + Ceph Manager + MGR + The Ceph manager software, which collects all the state from the whole + cluster in one place. + + Ceph Metadata Server + MDS + The Ceph metadata software. + + Ceph Clients + Ceph Client + The collection of Ceph components which can access a Ceph Storage + Cluster. These include the Ceph Object Gateway, the Ceph Block Device, + the Ceph Filesystem, and their corresponding libraries, kernel modules, + and FUSEs. + + Ceph Kernel Modules + The collection of kernel modules which can be used to interact with the + Ceph System (e.g,. ``ceph.ko``, ``rbd.ko``). + + Ceph Client Libraries + The collection of libraries that can be used to interact with components + of the Ceph System. + + Ceph Release + Any distinct numbered version of Ceph. + + Ceph Point Release + Any ad-hoc release that includes only bug or security fixes. + + Ceph Interim Release + Versions of Ceph that have not yet been put through quality assurance + testing, but may contain new features. + + Ceph Release Candidate + A major version of Ceph that has undergone initial quality assurance + testing and is ready for beta testers. + + Ceph Stable Release + A major version of Ceph where all features from the preceding interim + releases have been put through quality assurance testing successfully. + + Ceph Test Framework + Teuthology + The collection of software that performs scripted tests on Ceph. + + CRUSH + Controlled Replication Under Scalable Hashing. It is the algorithm + Ceph uses to compute object storage locations. + + ruleset + A set of CRUSH data placement rules that applies to a particular pool(s). + + Pool + Pools + Pools are logical partitions for storing objects. + + systemd oneshot + A systemd ``type`` where a command is defined in ``ExecStart`` which will + exit upon completion (it is not intended to daemonize) + + LVM tags + Extensible metadata for LVM volumes and groups. It is used to store + Ceph-specific information about devices and its relationship with + OSDs. + +.. _http://github.com/ceph: http://github.com/ceph +.. _Cluster Map: ../architecture#cluster-map