Ceph入门到精通-ceph 源码编译

Ceph - a scalable distributed storage system

Please see https://ceph.com/ for current info.

Contributing Code

Most of Ceph is dual licensed under the LGPL version 2.1 or 3.0. Some miscellaneous code is under a BSD-style license or is public domain. The documentation is licensed under Creative Commons Attribution Share Alike 3.0 (CC-BY-SA-3.0). There are a handful of headers included here that are licensed under the GPL. Please see the file COPYING for a full inventory of licenses by file.

Code contributions must include a valid "Signed-off-by" acknowledging the license for the modified or contributed file. Please see the file SubmittingPatches.rst for details on what that means and on how to generate and submit patches.

We do not require assignment of copyright to contribute code; code is contributed under the terms of the applicable license.

Checking out the source

You can clone from github with

复制代码
git clone git@github.com:ceph/ceph

or, if you are not a github user,

复制代码
git clone https://github.com/ceph/ceph.git

Ceph contains many git submodules that need to be checked out with

复制代码
git submodule update --init --recursive

Build Prerequisites

The list of Debian or RPM packages dependencies can be installed with:

复制代码
./install-deps.sh

Building Ceph

Note that these instructions are meant for developers who are compiling the code for development and testing. To build binaries suitable for installation we recommend you build deb or rpm packages or refer to the ceph.spec.in or debian/rules to see which configuration options are specified for production builds.

Build instructions:

复制代码
./do_cmake.sh
cd build
ninja

(do_cmake.sh now defaults to creating a debug build of ceph that can be up to 5x slower with some workloads. Please pass "-DCMAKE_BUILD_TYPE=RelWithDebInfo" to do_cmake.sh to create a non-debug release.

The number of jobs used by ninja is derived from the number of CPU cores of the building host if unspecified. Use the -j option to limit the job number if the build jobs are running out of memory. On average, each job takes around 2.5GiB memory.)

This assumes you make your build dir a subdirectory of the ceph.git checkout. If you put it elsewhere, just point CEPH_GIT_DIR to the correct path to the checkout. Any additional CMake args can be specified by setting ARGS before invoking do_cmake. See cmake options for more details. Eg.

复制代码
ARGS="-DCMAKE_C_COMPILER=gcc-7" ./do_cmake.sh

To build only certain targets use:

复制代码
ninja [target name]

To install:

复制代码
ninja install

CMake Options

If you run the cmake command by hand, there are many options you can set with "-D". For example, the option to build the RADOS Gateway is defaulted to ON. To build without the RADOS Gateway:

复制代码
cmake -DWITH_RADOSGW=OFF [path to top-level ceph directory]

Another example below is building with debugging and alternate locations for a couple of external dependencies:

复制代码
cmake -DLEVELDB_PREFIX="/opt/hyperleveldb" \
-DCMAKE_INSTALL_PREFIX=/opt/ceph -DCMAKE_C_FLAGS="-Og -g3 -gdwarf-4" \
..

Ceph has several bundled dependencies such as Boost, RocksDB and Arrow. By default, cmake will build these bundled dependencies from source instead of using libraries that are already installed on the system. You can opt-in to using these system libraries, provided they meet the minimum version required by Ceph, with cmake options like WITH_SYSTEM_BOOST:

复制代码
cmake -DWITH_SYSTEM_BOOST=ON [...]

To view an exhaustive list of -D options, you can invoke cmake with:

复制代码
cmake -LH

If you often pipe ninja to less and would like to maintain the diagnostic colors for errors and warnings (and if your compiler supports it), you can invoke cmake with:

复制代码
cmake -DDIAGNOSTICS_COLOR=always ...

Then you'll get the diagnostic colors when you execute:

复制代码
ninja | less -R

Other available values for 'DIAGNOSTICS_COLOR' are 'auto' (default) and 'never'.

Building a source tarball

To build a complete source tarball with everything needed to build from source and/or build a (deb or rpm) package, run

复制代码
./make-dist

This will create a tarball like ceph-$version.tar.bz2 from git. (Ensure that any changes you want to include in your working directory are committed to git.)

Running a test cluster

To run a functional test cluster,

复制代码
cd build
ninja vstart        # builds just enough to run vstart
../src/vstart.sh --debug --new -x --localhost --bluestore
./bin/ceph -s

Almost all of the usual commands are available in the bin/ directory. For example,

复制代码
./bin/rados -p rbd bench 30 write
./bin/rbd create foo --size 1000

To shut down the test cluster,

复制代码
../src/stop.sh

To start or stop individual daemons, the sysvinit script can be used:

复制代码
./bin/init-ceph restart osd.0
./bin/init-ceph stop

Running unit tests

To build and run all tests (in parallel using all processors), use ctest:

复制代码
cd build
ninja
ctest -j$(nproc)

(Note: Many targets built from src/test are not run using ctest. Targets starting with "unittest" are run in ninja check and thus can be run with ctest. Targets starting with "ceph_test" can not, and should be run by hand.)

When failures occur, look in build/Testing/Temporary for logs.

To build and run all tests and their dependencies without other unnecessary targets in Ceph:

复制代码
cd build
ninja check -j$(nproc)

To run an individual test manually, run ctest with -R (regex matching):

复制代码
ctest -R [regex matching test name(s)]

(Note: ctest does not build the test it's running or the dependencies needed to run it)

To run an individual test manually and see all the tests output, run ctest with the -V (verbose) flag:

复制代码
ctest -V -R [regex matching test name(s)]

To run tests manually and run the jobs in parallel, run ctest with the -j flag:

复制代码
ctest -j [number of jobs]

There are many other flags you can give ctest for better control over manual test execution. To view these options run:

复制代码
man ctest

Building the Documentation

Prerequisites

The list of package dependencies for building the documentation can be found in doc_deps.deb.txt:

复制代码
sudo apt-get install `cat doc_deps.deb.txt`

Building the Documentation

To build the documentation, ensure that you are in the top-level /ceph directory, and execute the build script. For example:

复制代码
admin/build-doc

Reporting Issues

To report an issue and view existing issues, please visit Overview - Ceph - Ceph.

相关推荐
lisanmengmeng2 天前
cephfs rbd应用
linux·运维·服务器·ceph
oMcLin3 天前
如何在 Manjaro Linux 上实现高效的 Ceph 存储集群,提升大规模文件存储的冗余性与性能?
linux·运维·ceph
wniuniu_5 天前
ceph的osd
java·前端·ceph
mixboot6 天前
Ceph PG 不一致问题排查与修复 scrub errors
ceph·scrub
斯普信专业组6 天前
从 Deep Scrubbing 滞后到集群性能跃迁:一次“以小见大”的 Ceph 优化实录
ceph
oMcLin6 天前
如何在CentOS 7.9 服务器上配置并优化 Ceph 分布式存储集群,提升数据冗余与性能?
服务器·ceph·centos
mixboot7 天前
Ceph BlueFS 溢出修复
ceph·bluefs溢出
only火车头11 天前
升级 ceph (16.2 -> 18.2) ceph mon 启动失败
服务器·ceph
iconball13 天前
个人用云计算学习笔记 --35 Ceph 分布式存储
运维·笔记·ceph·学习·云计算
become__better14 天前
判断ceph osd 节点磁盘异常
linux·运维·ceph