Skip to content

Commit

Permalink
Doc: deleted gratuitious PlanningImplementation.html,
Browse files Browse the repository at this point in the history
which was a copy of PlanningImplementation.txt
(and not html at all).

restored previous index.rst, which was overwritten with a copy
of PlanninImplementation.txt, but removed all of the recursively
included content from the document.

I will cherry-pick merge the new contents in a subsequent commit.

Signed-off-by: Mark Kampe <[email protected]>
  • Loading branch information
Mark Kampe committed Dec 2, 2011
1 parent fdf3f7b commit 3c7a82a
Show file tree
Hide file tree
Showing 2 changed files with 16 additions and 85 deletions.
43 changes: 0 additions & 43 deletions doc/dev/PlanningImplementation.html

This file was deleted.

58 changes: 16 additions & 42 deletions doc/dev/index.rst
Original file line number Diff line number Diff line change
@@ -1,45 +1,19 @@
About this Document
This document contains planning and implementation procedures for Ceph. The audience for this document includes technical support personnel, installation engineers, system administrators, and quality assurance.
Prerequisites
Users of this document must be familiar with Linux command line options. They must also be familiar with the overall Ceph product.
Before You Begin
Before implementing a new Ceph System, first answer the questions in the Ceph Getting Started Guide to determine your configuration needs. Once you have determined your hardware and configuration needs, the following decisions must be made:
• Determine what level of technical support you need. Pick from the Ceph Technical Support options in the next section.
• Determine how much and what level of training your organization needs.
Ceph Technical Support Options
The Ceph Technical support model provides 4 tiers of technical support options:
1st – This option is for brand new customers that need installation, configuration, and setup on their production environment.
2nd – This level of support requires a trouble ticket to be generated on a case by case basis as customer difficulties arise. Customers can choose between two maintenance options; they can either purchase a yearly maintenance contract, or pay for each trouble resolution as it occurs.
3rd – This option comes with our bundled packages for customers who have also purchased our hosting plans. In this case, the customer is a service provider. The Help Desk can generally provide this level of incident resolution. (NEED MORE INFO)
4th – This level of support requires a Service Level Agreement (SLA) between the customer and Dreamhost. This level is used for handling the most difficult or advanced problems.
Planning a Ceph Cluster Configuration
The following section contains guidelines for planning the deployment for a Ceph cluster configuration. A Ceph cluster consists of the following core components:
• Monitors – These must be an odd number, such as one, three, or five. Three is the preferred configuration.
• Object Storage Devices (OSD) – used as storage nodes
• Metadata Servers (MDS)
For redundancy, you should employ several of these components.
Monitors
The monitors handle central cluster management, configuration, and state.
Hardware Requirements:
• A few gigs of local disk space
• A fixed network address
Warning: Never configure 2 monitors per cluster. If you do, they will both have to be up all of the time, which will greatly degrade system performance.
Object Storage Devices
The OSDs store the actual data on the disks. A minimum of two is required.
Hardware Requirements:
• As many disks as possible for faster performance and scalability
• An SSD or NVRAM for a journal, or a RAID controller with a battery-backed NVRAM.
• Ample RAM for better file system caching
• Fast network
Metadata Servers
The metadata server daemon commands act as a distributed, coherent cache of file system metadata. They do not store data locally; all metadata is stored on disk via the storage nodes.
Metadata servers can be added into the cluster on an as-needed basis. The load is automatically balanced. The max_mds parameter controls how many cmds instances are active. Any additional running instances are put in standby mode and can be activated if one of the active daemons becomes unresponsive.
Hardware Requirements:
• Large amount of RAM
• Fast CPU
• Fast (low latency) network
• At least two servers for redundancy and load balancing
TIPS: If you have just a few nodes, put cmon, cmds, and cosd on the same node. For moderate node configurations, put cmon and cmds together, and cosd on the disk nodes. For large node configurations, put cmon, cmds, and cosd each on their own dedicated machine.
==================================
Internal developer documentation
==================================

.. note:: If you're looking for how to use Ceph as a library from your
own software, please see :doc:`/api/index`.

You can start a development mode Ceph cluster, after compiling the source, with::

cd src
install -d -m0755 out dev/osd0
./vstart.sh -n -x -l
# check that it's there
./ceph health

.. todo:: vstart is woefully undocumented and full of sharp sticks to poke yourself with.


.. toctree::
Expand Down

0 comments on commit 3c7a82a

Please sign in to comment.