mirror of
				https://git.proxmox.com/git/pve-docs
				synced 2025-10-31 09:00:11 +00:00 
			
		
		
		
	ceph: follow up rewords & fixes
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
This commit is contained in:
		
							parent
							
								
									76f6eca4ee
								
							
						
					
					
						commit
						2f19a6b0b6
					
				
							
								
								
									
										43
									
								
								pveceph.adoc
									
									
									
									
									
								
							
							
						
						
									
										43
									
								
								pveceph.adoc
									
									
									
									
									
								
							| @ -79,17 +79,16 @@ Check also the recommendations from | ||||
| http://docs.ceph.com/docs/luminous/start/hardware-recommendations/[Ceph's website]. | ||||
| 
 | ||||
| .CPU | ||||
| As higher the core frequency the better, this will reduce latency.  Among other | ||||
| things, this benefits the services of Ceph, as they can process data faster. | ||||
| To simplify planning, you should assign a CPU core (or thread) to each Ceph | ||||
| service to provide enough resources for stable and durable Ceph performance. | ||||
| Higher CPU core frequency reduce latency and should be preferred. As a simple | ||||
| rule of thumb, you should assign a CPU core (or thread) to each Ceph service to | ||||
| provide enough resources for stable and durable Ceph performance. | ||||
| 
 | ||||
| .Memory | ||||
| Especially in a hyper-converged setup, the memory consumption needs to be | ||||
| carefully monitored. In addition to the intended workload (VM / Container), | ||||
| Ceph needs enough memory to provide good and stable performance. As a rule of | ||||
| thumb, for roughly 1TiB of data, 1 GiB of memory will be used by an OSD. With | ||||
| additionally needed memory for OSD caching. | ||||
| carefully monitored. In addition to the intended workload from virtual machines | ||||
| and container, Ceph needs enough memory available to provide good and stable | ||||
| performance. As a rule of thumb, for roughly 1 TiB of data, 1 GiB of memory | ||||
| will be used by an OSD. OSD caching will use additional memory. | ||||
| 
 | ||||
| .Network | ||||
| We recommend a network bandwidth of at least 10 GbE or more, which is used | ||||
| @ -97,15 +96,14 @@ exclusively for Ceph. A meshed network setup | ||||
| footnote:[Full Mesh Network for Ceph {webwiki-url}Full_Mesh_Network_for_Ceph_Server] | ||||
| is also an option if there are no 10 GbE switches available. | ||||
| 
 | ||||
| To be explicit about the network, since Ceph is a distributed network storage, | ||||
| its traffic must be put on its own physical network. The volume of traffic | ||||
| especially during recovery will interfere with other services on the same | ||||
| network. | ||||
| The volume of traffic, especially during recovery, will interfere with other | ||||
| services on the same network and may even break the {pve} cluster stack. | ||||
| 
 | ||||
| Further, estimate your bandwidth needs. While one HDD might not saturate a 1 Gb | ||||
| link, a SSD or a NVMe SSD certainly can. Modern NVMe SSDs will even saturate 10 | ||||
| Gb of bandwidth. You also should consider higher bandwidths, as these tend to | ||||
| come with lower latency. | ||||
| link, multiple HDD OSDs per node can, and modern NVMe SSDs will even saturate | ||||
| 10 Gbps of bandwidth quickly. Deploying a network capable of even more bandwith | ||||
| will ensure that it isn't your bottleneck and won't be anytime soon, 25, 40 or | ||||
| even 100 GBps are possible. | ||||
| 
 | ||||
| .Disks | ||||
| When planning the size of your Ceph cluster, it is important to take the | ||||
| @ -114,16 +112,21 @@ might take long. It is recommended that you use SSDs instead of HDDs in small | ||||
| setups to reduce recovery time, minimizing the likelihood of a subsequent | ||||
| failure event during recovery. | ||||
| 
 | ||||
| In general SSDs will provide more IOPs then spinning disks. This fact and the | ||||
| In general SSDs will provide more IOPs than spinning disks. This fact and the | ||||
| higher cost may make a xref:pve_ceph_device_classes[class based] separation of | ||||
| pools appealing.  Another possibility to speedup OSDs is to use a faster disk | ||||
| pools appealing. Another possibility to speedup OSDs is to use a faster disk | ||||
| as journal or DB/WAL device, see xref:pve_ceph_osds[creating Ceph OSDs]. If a | ||||
| faster disk is used for multiple OSDs, a proper balance between OSD and WAL / | ||||
| DB (or journal) disk must be selected, otherwise the faster disk becomes the | ||||
| bottleneck for all linked OSDs. | ||||
| 
 | ||||
| Aside from the disk type, Ceph best performs with an even sized and distributed | ||||
| amount of disks per node. For example, 4x disks à 500 GB in each node. | ||||
| amount of disks per node. For example, 4 x 500 GB disks with in each node is | ||||
| better than a mixed setup with a single 1 TB and three 250 GB disk. | ||||
| 
 | ||||
| One also need to balance OSD count and single OSD capacity. More capacity | ||||
| allows to increase storage density, but it also means that a single OSD | ||||
| failure forces ceph to recover more data at once. | ||||
| 
 | ||||
| .Avoid RAID | ||||
| As Ceph handles data object redundancy and multiple parallel writes to disks | ||||
| @ -137,8 +140,8 @@ the ones from Ceph. | ||||
| WARNING: Avoid RAID controller, use host bus adapter (HBA) instead. | ||||
| 
 | ||||
| NOTE: Above recommendations should be seen as a rough guidance for choosing | ||||
| hardware. Therefore, it is still essential to test your setup and monitor | ||||
| health & performance. | ||||
| hardware. Therefore, it is still essential to adapt it to your specific needs, | ||||
| test your setup and monitor health and performance continuously. | ||||
| 
 | ||||
| 
 | ||||
| [[pve_ceph_install]] | ||||
|  | ||||
		Loading…
	
		Reference in New Issue
	
	Block a user
	 Thomas Lamprecht
						Thomas Lamprecht