ceph/ceph
Thomas Lamprecht f38dd50b34 import ceph reef 18.2.4
Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com>
2024-07-25 18:23:05 +02:00
..
.github update ceph source to reef 18.2.1 2023-12-19 09:13:36 +01:00
admin import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
bin
ceph-menv import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
cmake/modules import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
debian import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
doc import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
etc import 14.2.4 nautilus point release 2019-09-19 13:49:59 +02:00
examples update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
fusetrace import 15.2.4 2020-07-02 15:37:13 +02:00
keys
man import quincy beta 17.1.0 2022-03-31 12:54:16 +02:00
mirroring update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
monitoring import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
qa import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
selinux update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
share
src import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
sudoers.d import ceph 16.2.7 2021-12-08 13:07:09 +01:00
systemd import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
udev Import ceph 15.2.8 2020-12-17 12:55:37 +01:00
.gitattributes update source to Ceph Pacific 16.2.2 2021-05-06 16:12:51 +02:00
.githubmap update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
.gitleaks.toml update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
.gitmodule_mirrors
.gitmodules update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
.mailmap update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
.organizationmap update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
.peoplemap import quincy beta 17.1.0 2022-03-31 12:54:16 +02:00
.readthedocs.yml import ceph quincy 17.2.6 2023-04-11 09:43:51 +02:00
AUTHORS
ceph.spec import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
ceph.spec.in import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
changelog.upstream import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
CMakeLists.txt import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
CodingStyle import ceph quincy 17.2.6 2023-04-11 09:43:51 +02:00
CONTRIBUTING.rst update source to Ceph Pacific 16.2.2 2021-05-06 16:12:51 +02:00
COPYING import quincy beta 17.1.0 2022-03-31 12:54:16 +02:00
COPYING-GPL2
COPYING-LGPL2.1
COPYING-LGPL3 import 15.2.0 Octopus source 2020-03-24 14:49:51 +01:00
do_cmake.sh update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
do_freebsd.sh update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
doc_deps.deb.txt import ceph 16.2.7 2021-12-08 13:07:09 +01:00
install-deps.sh update ceph source to reef 18.2.1 2023-12-19 09:13:36 +01:00
make-debs.sh import ceph 16.2.6 2021-09-20 09:15:23 +02:00
make-dist import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
make-srpm.sh
mingw_conf.sh update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
PendingReleaseNotes import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00
pom.xml
README.aix import ceph 16.2.6 2021-09-20 09:15:23 +02:00
README.FreeBSD
README.md update ceph source to reef 18.2.1 2023-12-19 09:13:36 +01:00
README.solaris
README.windows.rst update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
run-make-check.sh update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
SECURITY.md import quincy beta 17.1.0 2022-03-31 12:54:16 +02:00
SubmittingPatches-backports.rst update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
SubmittingPatches-kernel.rst import 15.2.0 Octopus source 2020-03-24 14:49:51 +01:00
SubmittingPatches.rst update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
win32_build.sh update ceph source to reef 18.1.2 2023-07-25 17:18:44 +02:00
win32_deps_build.sh import ceph reef 18.2.4 2024-07-25 18:23:05 +02:00

Ceph - a scalable distributed storage system

See https://ceph.com/ for current information about Ceph.

Contributing Code

Most of Ceph is dual-licensed under the LGPL version 2.1 or 3.0. Some miscellaneous code is either public domain or licensed under a BSD-style license.

The Ceph documentation is licensed under Creative Commons Attribution Share Alike 3.0 (CC-BY-SA-3.0).

Some headers included in the ceph/ceph repository are licensed under the GPL. See the file COPYING for a full inventory of licenses by file.

All code contributions must include a valid "Signed-off-by" line. See the file SubmittingPatches.rst for details on this and instructions on how to generate and submit patches.

Assignment of copyright is not required to contribute code. Code is contributed under the terms of the applicable license.

Checking out the source

Clone the ceph/ceph repository from github by running the following command on a system that has git installed:

git clone git@github.com:ceph/ceph

Alternatively, if you are not a github user, you should run the following command on a system that has git installed:

git clone https://github.com/ceph/ceph.git

When the ceph/ceph repository has been cloned to your system, run the following commands to move into the cloned ceph/ceph repository and to check out the git submodules associated with it:

cd ceph
git submodule update --init --recursive

Build Prerequisites

section last updated 27 Jul 2023

Make sure that curl is installed. The Debian and Ubuntu apt command is provided here, but if you use a system with a different package manager, then you must use whatever command is the proper counterpart of this one:

apt install curl

Install Debian or RPM package dependencies by running the following command:

./install-deps.sh

Install the python3-routes package:

apt install python3-routes

Building Ceph

These instructions are meant for developers who are compiling the code for development and testing. To build binaries that are suitable for installation we recommend that you build .deb or .rpm packages, or refer to ceph.spec.in or debian/rules to see which configuration options are specified for production builds.

To build Ceph, make sure that you are in the top-level ceph directory that contains do_cmake.sh and CONTRIBUTING.rst and run the following commands:

./do_cmake.sh
cd build
ninja

do_cmake.sh by default creates a "debug build" of Ceph, which can be up to five times slower than a non-debug build. Pass -DCMAKE_BUILD_TYPE=RelWithDebInfo to do_cmake.sh to create a non-debug build.

Ninja is the buildsystem used by the Ceph project to build test builds. The number of jobs used by ninja is derived from the number of CPU cores of the building host if unspecified. Use the -j option to limit the job number if the build jobs are running out of memory. If you attempt to run ninja and receive a message that reads g++: fatal error: Killed signal terminated program cc1plus, then you have run out of memory. Using the -j option with an argument appropriate to the hardware on which the ninja command is run is expected to result in a successful build. For example, to limit the job number to 3, run the command ninja -j 3. On average, each ninja job run in parallel needs approximately 2.5 GiB of RAM.

This documentation assumes that your build directory is a subdirectory of the ceph.git checkout. If the build directory is located elsewhere, point CEPH_GIT_DIR to the correct path of the checkout. Additional CMake args can be specified by setting ARGS before invoking do_cmake.sh. See cmake options for more details. For example:

ARGS="-DCMAKE_C_COMPILER=gcc-7" ./do_cmake.sh

To build only certain targets, run a command of the following form:

ninja [target name]

To install:

ninja install

CMake Options

If you run the cmake command by hand, there are many options you can set with "-D". For example, the option to build the RADOS Gateway is defaulted to ON. To build without the RADOS Gateway:

cmake -DWITH_RADOSGW=OFF [path to top-level ceph directory]

Another example below is building with debugging and alternate locations for a couple of external dependencies:

cmake -DLEVELDB_PREFIX="/opt/hyperleveldb" \
-DCMAKE_INSTALL_PREFIX=/opt/ceph -DCMAKE_C_FLAGS="-Og -g3 -gdwarf-4" \
..

Ceph has several bundled dependencies such as Boost, RocksDB and Arrow. By default, cmake will build these bundled dependencies from source instead of using libraries that are already installed on the system. You can opt-in to using these system libraries, provided they meet the minimum version required by Ceph, with cmake options like WITH_SYSTEM_BOOST:

cmake -DWITH_SYSTEM_BOOST=ON [...]

To view an exhaustive list of -D options, you can invoke cmake with:

cmake -LH

If you often pipe ninja to less and would like to maintain the diagnostic colors for errors and warnings (and if your compiler supports it), you can invoke cmake with:

cmake -DDIAGNOSTICS_COLOR=always ...

Then you'll get the diagnostic colors when you execute:

ninja | less -R

Other available values for 'DIAGNOSTICS_COLOR' are 'auto' (default) and 'never'.

Building a source tarball

To build a complete source tarball with everything needed to build from source and/or build a (deb or rpm) package, run

./make-dist

This will create a tarball like ceph-$version.tar.bz2 from git. (Ensure that any changes you want to include in your working directory are committed to git.)

Running a test cluster

From the ceph/ directory, run the following commands to launch a test Ceph cluster:

cd build
ninja vstart        # builds just enough to run vstart
../src/vstart.sh --debug --new -x --localhost --bluestore
./bin/ceph -s

Most Ceph commands are available in the bin/ directory. For example:

./bin/rbd create foo --size 1000
./bin/rados -p foo bench 30 write

To shut down the test cluster, run the following command from the build/ directory:

../src/stop.sh

Use the sysvinit script to start or stop individual daemons:

./bin/init-ceph restart osd.0
./bin/init-ceph stop

Running unit tests

To build and run all tests (in parallel using all processors), use ctest:

cd build
ninja
ctest -j$(nproc)

(Note: Many targets built from src/test are not run using ctest. Targets starting with "unittest" are run in ninja check and thus can be run with ctest. Targets starting with "ceph_test" can not, and should be run by hand.)

When failures occur, look in build/Testing/Temporary for logs.

To build and run all tests and their dependencies without other unnecessary targets in Ceph:

cd build
ninja check -j$(nproc)

To run an individual test manually, run ctest with -R (regex matching):

ctest -R [regex matching test name(s)]

(Note: ctest does not build the test it's running or the dependencies needed to run it)

To run an individual test manually and see all the tests output, run ctest with the -V (verbose) flag:

ctest -V -R [regex matching test name(s)]

To run tests manually and run the jobs in parallel, run ctest with the -j flag:

ctest -j [number of jobs]

There are many other flags you can give ctest for better control over manual test execution. To view these options run:

man ctest

Building the Documentation

Prerequisites

The list of package dependencies for building the documentation can be found in doc_deps.deb.txt:

sudo apt-get install `cat doc_deps.deb.txt`

Building the Documentation

To build the documentation, ensure that you are in the top-level /ceph directory, and execute the build script. For example:

admin/build-doc

Reporting Issues

To report an issue and view existing issues, please visit https://tracker.ceph.com/projects/ceph.