Fix some bugs when testing opensds ansible
[stor4nfv.git] / src / ceph / doc / rados / operations / index.rst
1 ====================
2  Cluster Operations
3 ====================
4
5 .. raw:: html
6
7         <table><colgroup><col width="50%"><col width="50%"></colgroup><tbody valign="top"><tr><td><h3>High-level Operations</h3>
8
9 High-level cluster operations consist primarily of starting, stopping, and
10 restarting a cluster with the ``ceph`` service;  checking the cluster's health;
11 and, monitoring an operating cluster.
12
13 .. toctree::
14         :maxdepth: 1 
15         
16         operating
17         health-checks
18         monitoring
19         monitoring-osd-pg
20         user-management
21
22 .. raw:: html 
23
24         </td><td><h3>Data Placement</h3>
25
26 Once you have your cluster up and running, you may begin working with data
27 placement. Ceph supports petabyte-scale data storage clusters, with storage
28 pools and placement groups that distribute data across the cluster using Ceph's
29 CRUSH algorithm.
30
31 .. toctree::
32         :maxdepth: 1
33
34         data-placement
35         pools
36         erasure-code
37         cache-tiering
38         placement-groups
39         upmap
40         crush-map
41         crush-map-edits
42
43
44
45 .. raw:: html
46
47         </td></tr><tr><td><h3>Low-level Operations</h3>
48
49 Low-level cluster operations consist of starting, stopping, and restarting a
50 particular daemon within a cluster; changing the settings of a particular
51 daemon or subsystem; and, adding a daemon to the cluster or removing a  daemon
52 from the cluster. The most common use cases for low-level operations include
53 growing or shrinking the Ceph cluster and replacing legacy or failed hardware
54 with new hardware.
55
56 .. toctree::
57         :maxdepth: 1
58
59         add-or-rm-osds
60         add-or-rm-mons
61         Command Reference <control>
62
63         
64
65 .. raw:: html
66
67         </td><td><h3>Troubleshooting</h3>
68
69 Ceph is still on the leading edge, so you may encounter situations that require
70 you to evaluate your Ceph configuration and modify your logging and debugging
71 settings to identify and remedy issues you are encountering with your cluster.
72
73 .. toctree::
74         :maxdepth: 1 
75
76         ../troubleshooting/community
77         ../troubleshooting/troubleshooting-mon
78         ../troubleshooting/troubleshooting-osd
79         ../troubleshooting/troubleshooting-pg
80         ../troubleshooting/log-and-debug
81         ../troubleshooting/cpu-profiling
82         ../troubleshooting/memory-profiling
83
84
85
86
87 .. raw:: html
88
89         </td></tr></tbody></table>
90