Skip site navigation (1)Skip section navigation (2)

FreeBSD Manual Pages

  
 
  

home | help
CEPH(8)				     Ceph			       CEPH(8)

NAME
       ceph - ceph administration tool

SYNOPSIS
       ceph auth [ add | caps |	del | export | get | get-key | get-or-create | get-or-create-key | import | list | print-key | print_key ] ...

       ceph compact

       ceph config-key [ del | exists |	get | list | put ] ...

       ceph daemon _name_ | _path_ _command_ ...

       ceph daemonperf _name_ |	_path_ [ interval [ count ] ]

       ceph df {detail}

       ceph fs [ ls | new | reset | rm ] ...

       ceph fsid

       ceph health {detail}

       ceph heap [ dump	| start_profiler | stop_profiler | release | stats ] ...

       ceph injectargs _injectedargs_ [	_injectedargs_... ]

       ceph log	_logtext_ [ _logtext_... ]

       ceph mds	[ compat | deactivate |	fail | rm | rmfailed | set_state | stat	| tell ] ...

       ceph mon	[ add |	dump | getmap |	remove | stat ]	...

       ceph mon_status

       ceph osd	[ blacklist | blocked-by | create | deep-scrub | df | down | dump | erasure-code-profile | find	| getcrushmap |	getmap | getmaxosd | in	| lspools | map	| metadata | out | pause | perf	| pg-temp | primary-affinity | primary-temp | repair | reweight	| reweight-by-pg | rm |	scrub |	set | setcrushmap | setmaxosd  | stat |	thrash | tree |	unpause	| unset	] ...

       ceph osd	crush [	add | add-bucket | create-or-move | dump | get-tunable | link |	move | remove |	rename-bucket |	reweight | reweight-all	| reweight-subtree | rm	| rule | set | set-tunable | show-tunables | tunables |	unlink ] ...

       ceph osd	pool [ create |	delete | get | get-quota | ls |	mksnap | rename	| rmsnap | set | set-quota | stats ] ...

       ceph osd	tier [ add | add-cache | cache-mode | remove | remove-overlay |	set-overlay ] ...

       ceph pg [ debug | deep-scrub | dump | dump_json | dump_pools_json | dump_stuck |	force_create_pg	| getmap | ls |	ls-by-osd | ls-by-pool | ls-by-primary | map | repair |	scrub |	set_full_ratio | set_nearfull_ratio | stat ] ...

       ceph quorum [ enter | exit ]

       ceph quorum_status

       ceph report { _tags_ [ _tags_...	] }

       ceph scrub

       ceph status

       ceph sync force {--yes-i-really-mean-it}	{--i-know-what-i-am-doing}

       ceph tell _name (type.id)_ _args_ [_args_...]

       ceph version

DESCRIPTION
       ceph is a control utility which is used for manual deployment and main-
       tenance of a Ceph cluster. It provides a	diverse	set of	commands  that
       allows  deployment of monitors, OSDs, placement groups, MDS and overall
       maintenance, administration of the cluster.

COMMANDS
   auth
       Manage authentication keys. It is used for adding, removing,  exporting
       or  updating  of	authentication keys for	a particular  entity such as a
       monitor or OSD. It uses some additional subcommands.

       Subcommand add adds authentication info for a  particular  entity  from
       input  file,  or	random key if no input is given	and/or any caps	speci-
       fied in the command.

       Usage:

	  ceph auth add	<entity> {<caps> [<caps>...]}

       Subcommand caps updates caps for	name from caps specified in  the  com-
       mand.

       Usage:

	  ceph auth caps <entity> <caps> [<caps>...]

       Subcommand del deletes all caps for name.

       Usage:

	  ceph auth del	<entity>

       Subcommand  export  writes  keyring  for	 requested  entity,  or	master
       keyring if none given.

       Usage:

	  ceph auth export {<entity>}

       Subcommand get writes keyring file with requested key.

       Usage:

	  ceph auth get	<entity>

       Subcommand get-key displays requested key.

       Usage:

	  ceph auth get-key <entity>

       Subcommand get-or-create	adds authentication info for a particular  en-
       tity  from  input file, or random key if	no input given and/or any caps
       specified in the	command.

       Usage:

	  ceph auth get-or-create <entity> {<caps> [<caps>...]}

       Subcommand get-or-create-key gets or adds key for name from system/caps
       pairs  specified	in the command.	 If key	already	exists,	any given caps
       must match the existing caps for	that key.

       Usage:

	  ceph auth get-or-create-key <entity> {<caps> [<caps>...]}

       Subcommand import reads keyring from input file.

       Usage:

	  ceph auth import

       Subcommand list lists authentication state.

       Usage:

	  ceph auth list

       Subcommand print-key displays requested key.

       Usage:

	  ceph auth print-key <entity>

       Subcommand print_key displays requested key.

       Usage:

	  ceph auth print_key <entity>

   compact
       Causes compaction of monitor's leveldb storage.

       Usage:

	  ceph compact

   config-key
       Manage configuration key. It uses some additional subcommands.

       Subcommand del deletes configuration key.

       Usage:

	  ceph config-key del <key>

       Subcommand exists checks	for configuration keys existence.

       Usage:

	  ceph config-key exists <key>

       Subcommand get gets the configuration key.

       Usage:

	  ceph config-key get <key>

       Subcommand list lists configuration keys.

       Usage:

	  ceph config-key list

       Subcommand put puts configuration key and values.

       Usage:

	  ceph config-key put <key> {<val>}

   daemon
       Submit admin-socket commands.

       Usage:

	  ceph daemon {daemon_name|socket_path}	{command} ...

       Example:

	  ceph daemon osd.0 help

   daemonperf
       Watch performance counters from a Ceph daemon.

       Usage:

	  ceph daemonperf {daemon_name|socket_path} [{interval}	[{count}]]

   df
       Show cluster's free space status.

       Usage:

	  ceph df {detail}

   fs
       Manage cephfs filesystems. It uses some additional subcommands.

       Subcommand ls to	list filesystems

       Usage:

	  ceph fs ls

       Subcommand new to make a	new filesystem using  named  pools  <metadata>
       and <data>

       Usage:

	  ceph fs new <fs_name>	<metadata> <data>

       Subcommand  reset  is  used for disaster	recovery only: reset to	a sin-
       gle-MDS map

       Usage:

	  ceph fs reset	<fs_name> {--yes-i-really-mean-it}

       Subcommand rm to	disable	the named filesystem

       Usage:

	  ceph fs rm <fs_name> {--yes-i-really-mean-it}

   fsid
       Show cluster's FSID/UUID.

       Usage:

	  ceph fsid

   health
       Show cluster's health.

       Usage:

	  ceph health {detail}

   heap
       Show heap usage info (available only if compiled	with tcmalloc)

       Usage:

	  ceph heap dump|start_profiler|stop_profiler|release|stats

   injectargs
       Inject configuration arguments into monitor.

       Usage:

	  ceph injectargs <injected_args> [<injected_args>...]

   log
       Log supplied text to the	monitor	log.

       Usage:

	  ceph log <logtext> [<logtext>...]

   mds
       Manage metadata server configuration and	administration.	It  uses  some
       additional subcommands.

       Subcommand  compat manages compatible features. It uses some additional
       subcommands.

       Subcommand rm_compat removes compatible feature.

       Usage:

	  ceph mds compat rm_compat <int[0-]>

       Subcommand rm_incompat removes incompatible feature.

       Usage:

	  ceph mds compat rm_incompat <int[0-]>

       Subcommand show shows mds compatibility settings.

       Usage:

	  ceph mds compat show

       Subcommand deactivate stops mds.

       Usage:

	  ceph mds deactivate <who>

       Subcommand fail forces mds to status fail.

       Usage:

	  ceph mds fail	<who>

       Subcommand rm removes inactive mds.

       Usage:

	  ceph mds rm <int[0-]>	<name> (type.id)>

       Subcommand rmfailed removes failed mds.

       Usage:

	  ceph mds rmfailed <int[0-]>

       Subcommand set_state sets mds state of <gid> to <numeric-state>.

       Usage:

	  ceph mds set_state <int[0-]> <int[0-20]>

       Subcommand stat shows MDS status.

       Usage:

	  ceph mds stat

       Subcommand tell sends command to	particular mds.

       Usage:

	  ceph mds tell	<who> <args> [<args>...]

   mon
       Manage monitor configuration and	administration.	 It  uses  some	 addi-
       tional subcommands.

       Subcommand add adds new monitor named <name> at <addr>.

       Usage:

	  ceph mon add <name> <IPaddr[:port]>

       Subcommand dump dumps formatted monmap (optionally from epoch)

       Usage:

	  ceph mon dump	{<int[0-]>}

       Subcommand getmap gets monmap.

       Usage:

	  ceph mon getmap {<int[0-]>}

       Subcommand remove removes monitor named <name>.

       Usage:

	  ceph mon remove <name>

       Subcommand stat summarizes monitor status.

       Usage:

	  ceph mon stat

   mon_status
       Reports status of monitors.

       Usage:

	  ceph mon_status

   osd
       Manage  OSD  configuration  and administration. It uses some additional
       subcommands.

       Subcommand blacklist manage blacklisted clients.	 It  uses  some	 addi-
       tional subcommands.

       Subcommand  add add <addr> to blacklist (optionally until <expire> sec-
       onds from now)

       Usage:

	  ceph osd blacklist add <EntityAddr> {<float[0.0-]>}

       Subcommand ls show blacklisted clients

       Usage:

	  ceph osd blacklist ls

       Subcommand rm remove <addr> from	blacklist

       Usage:

	  ceph osd blacklist rm	<EntityAddr>

       Subcommand blocked-by prints a histogram	of  which  OSDs	 are  blocking
       their peers

       Usage:

	  ceph osd blocked-by

       Subcommand create creates new osd (with optional	UUID and ID).

       Usage:

	  ceph osd create {<uuid>} {<id>}

       Subcommand  crush is used for CRUSH management. It uses some additional
       subcommands.

       Subcommand add adds or updates crushmap position	and weight for	<name>
       with <weight> and location <args>.

       Usage:

	  ceph osd crush add <osdname (id|osd.id)> <float[0.0-]> <args>	[<args>...]

       Subcommand  add-bucket  adds  no-parent	(probably  root)  crush	bucket
       <name> of type <type>.

       Usage:

	  ceph osd crush add-bucket <name> <type>

       Subcommand create-or-move creates entry or  moves  existing  entry  for
       <name> <weight> at/to location <args>.

       Usage:

	  ceph osd crush create-or-move	<osdname (id|osd.id)> <float[0.0-]> <args>
	  [<args>...]

       Subcommand dump dumps crush map.

       Usage:

	  ceph osd crush dump

       Subcommand get-tunable get crush	tunable	straw_calc_version

       Usage:

	  ceph osd crush get-tunable straw_calc_version

       Subcommand link links existing entry for	<name> under location <args>.

       Usage:

	  ceph osd crush link <name> <args> [<args>...]

       Subcommand move moves existing entry for	<name> to location <args>.

       Usage:

	  ceph osd crush move <name> <args> [<args>...]

       Subcommand remove removes <name>	from crush map (everywhere, or just at
       <ancestor>).

       Usage:

	  ceph osd crush remove	<name> {<ancestor>}

       Subcommand rename-bucket	renames	buchket	<srcname> to <stname>

       Usage:

	  ceph osd crush rename-bucket <srcname> <dstname>

       Subcommand reweight change <name>'s weight to <weight> in crush map.

       Usage:

	  ceph osd crush reweight <name> <float[0.0-]>

       Subcommand reweight-all recalculate the weights for the tree to	ensure
       they sum	correctly

       Usage:

	  ceph osd crush reweight-all

       Subcommand  reweight-subtree  changes  all leaf items beneath <name> to
       <weight>	in crush map

       Usage:

	  ceph osd crush reweight-subtree <name> <weight>

       Subcommand rm removes <name> from crush map  (everywhere,  or  just  at
       <ancestor>).

       Usage:

	  ceph osd crush rm <name> {<ancestor>}

       Subcommand  rule	 is  used for creating crush rules. It uses some addi-
       tional subcommands.

       Subcommand create-erasure creates crush rule <name> for	erasure	 coded
       pool created with <profile> (default default).

       Usage:

	  ceph osd crush rule create-erasure <name> {<profile>}

       Subcommand  create-simple  creates  crush  rule	<name>	to  start from
       <root>, replicate across	buckets	of type	<type>,	using a	choose mode of
       <firstn|indep> (default firstn; indep best for erasure pools).

       Usage:

	  ceph osd crush rule create-simple <name> <root> <type> {firstn|indep}

       Subcommand dump dumps crush rule	<name> (default	all).

       Usage:

	  ceph osd crush rule dump {<name>}

       Subcommand list lists crush rules.

       Usage:

	  ceph osd crush rule list

       Subcommand ls lists crush rules.

       Usage:

	  ceph osd crush rule ls

       Subcommand rm removes crush rule	<name>.

       Usage:

	  ceph osd crush rule rm <name>

       Subcommand set used alone, sets crush map from input file.

       Usage:

	  ceph osd crush set

       Subcommand  set with osdname/osd.id update crushmap position and	weight
       for <name> to <weight> with location <args>.

       Usage:

	  ceph osd crush set <osdname (id|osd.id)> <float[0.0-]> <args>	[<args>...]

       Subcommand set-tunable set crush	tunable	 <tunable>  to	<value>.   The
       only tunable that can be	set is straw_calc_version.

       Usage:

	  ceph osd crush set-tunable straw_calc_version	<value>

       Subcommand show-tunables	shows current crush tunables.

       Usage:

	  ceph osd crush show-tunables

       Subcommand tree shows the crush buckets and items in a tree view.

       Usage:

	  ceph osd crush tree

       Subcommand tunables sets	crush tunables values to <profile>.

       Usage:

	  ceph osd crush tunables legacy|argonaut|bobtail|firefly|hammer|optimal|default

       Subcommand unlink unlinks <name>	from crush map (everywhere, or just at
       <ancestor>).

       Usage:

	  ceph osd crush unlink	<name> {<ancestor>}

       Subcommand df shows OSD utilization

       Usage:

	  ceph osd df {plain|tree}

       Subcommand deep-scrub initiates deep scrub on specified osd.

       Usage:

	  ceph osd deep-scrub <who>

       Subcommand down sets osd(s) <id>	[<id>...] down.

       Usage:

	  ceph osd down	<ids> [<ids>...]

       Subcommand dump prints summary of OSD map.

       Usage:

	  ceph osd dump	{<int[0-]>}

       Subcommand erasure-code-profile is used for managing the	 erasure  code
       profiles. It uses some additional subcommands.

       Subcommand get gets erasure code	profile	<name>.

       Usage:

	  ceph osd erasure-code-profile	get <name>

       Subcommand ls lists all erasure code profiles.

       Usage:

	  ceph osd erasure-code-profile	ls

       Subcommand rm removes erasure code profile <name>.

       Usage:

	  ceph osd erasure-code-profile	rm <name>

       Subcommand  set creates erasure code profile <name> with	[<key[=value]>
       ...]  pairs. Add	a --force at the end to	override an  existing  profile
       (IT IS RISKY).

       Usage:

	  ceph osd erasure-code-profile	set <name> {<profile> [<profile>...]}

       Subcommand find find osd	<id> in	the CRUSH map and shows	its location.

       Usage:

	  ceph osd find	<int[0-]>

       Subcommand getcrushmap gets CRUSH map.

       Usage:

	  ceph osd getcrushmap {<int[0-]>}

       Subcommand getmap gets OSD map.

       Usage:

	  ceph osd getmap {<int[0-]>}

       Subcommand getmaxosd shows largest OSD id.

       Usage:

	  ceph osd getmaxosd

       Subcommand in sets osd(s) <id> [<id>...]	in.

       Usage:

	  ceph osd in <ids> [<ids>...]

       Subcommand lost marks osd as permanently	lost. THIS DESTROYS DATA IF NO
       MORE REPLICAS EXIST, BE CAREFUL.

       Usage:

	  ceph osd lost	<int[0-]> {--yes-i-really-mean-it}

       Subcommand ls shows all OSD ids.

       Usage:

	  ceph osd ls {<int[0-]>}

       Subcommand lspools lists	pools.

       Usage:

	  ceph osd lspools {<int>}

       Subcommand map finds pg for <object> in <pool>.

       Usage:

	  ceph osd map <poolname> <objectname>

       Subcommand metadata fetches metadata for	osd <id>.

       Usage:

	  ceph osd metadata {int[0-]} (default all)

       Subcommand out sets osd(s) <id> [<id>...] out.

       Usage:

	  ceph osd out <ids> [<ids>...]

       Subcommand pause	pauses osd.

       Usage:

	  ceph osd pause

       Subcommand perf prints dump of OSD perf summary stats.

       Usage:

	  ceph osd perf

       Subcommand pg-temp set pg_temp mapping pgid:[<id> [<id>...]]  (develop-
       ers only).

       Usage:

	  ceph osd pg-temp <pgid> {<id>	[<id>...]}

       Subcommand  pool	 is  used  for managing	data pools. It uses some addi-
       tional subcommands.

       Subcommand create creates pool.

       Usage:

	  ceph osd pool	create <poolname> <int[0-]> {<int[0-]>}	{replicated|erasure}
	  {<erasure_code_profile>} {<ruleset>} {<int>}

       Subcommand delete deletes pool.

       Usage:

	  ceph osd pool	delete <poolname> {<poolname>} {--yes-i-really-really-mean-it}

       Subcommand get gets pool	parameter <var>.

       Usage:

	  ceph osd pool	get <poolname> size|min_size|crash_replay_interval|pg_num|
	  pgp_num|crush_ruleset|auid|write_fadvise_dontneed

       Only for	tiered pools:

	  ceph osd pool	get <poolname> hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|
	  target_max_objects|target_max_bytes|cache_target_dirty_ratio|cache_target_dirty_high_ratio|
	  cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|
	  min_read_recency_for_promote|hit_set_grade_decay_rate|hit_set_search_last_n

       Only for	erasure	coded pools:

	  ceph osd pool	get <poolname> erasure_code_profile

       Use all to get all pool parameters that apply to	the pool's type:

	  ceph osd pool	get <poolname> all

       Subcommand get-quota obtains object or byte limits for pool.

       Usage:

	  ceph osd pool	get-quota <poolname>

       Subcommand ls list pools

       Usage:

	  ceph osd pool	ls {detail}

       Subcommand mksnap makes snapshot	<snap> in <pool>.

       Usage:

	  ceph osd pool	mksnap <poolname> <snap>

       Subcommand rename renames <srcpool> to <destpool>.

       Usage:

	  ceph osd pool	rename <poolname> <poolname>

       Subcommand rmsnap removes snapshot <snap> from <pool>.

       Usage:

	  ceph osd pool	rmsnap <poolname> <snap>

       Subcommand set sets pool	parameter <var>	to <val>.

       Usage:

	  ceph osd pool	set <poolname> size|min_size|crash_replay_interval|pg_num|
	  pgp_num|crush_ruleset|hashpspool|nodelete|nopgchange|nosizechange|
	  hit_set_type|hit_set_period|hit_set_count|hit_set_fpp|debug_fake_ec_pool|
	  target_max_bytes|target_max_objects|cache_target_dirty_ratio|
	  cache_target_dirty_high_ratio|
	  cache_target_full_ratio|cache_min_flush_age|cache_min_evict_age|auid|
	  min_read_recency_for_promote|write_fadvise_dontneed|hit_set_grade_decay_rate|
	  hit_set_search_last_n
	  <val>	{--yes-i-really-mean-it}

       Subcommand set-quota sets object	or byte	limit on pool.

       Usage:

	  ceph osd pool	set-quota <poolname> max_objects|max_bytes <val>

       Subcommand stats	obtain stats from all pools, or	from specified pool.

       Usage:

	  ceph osd pool	stats {<name>}

       Subcommand  primary-affinity  adjust  osd  primary-affinity  from   0.0
       <=<weight> <= 1.0

       Usage:

	  ceph osd primary-affinity <osdname (id|osd.id)> <float[0.0-1.0]>

       Subcommand  primary-temp	sets primary_temp mapping pgid:<id>|-1 (devel-
       opers only).

       Usage:

	  ceph osd primary-temp	<pgid> <id>

       Subcommand repair initiates repair on a specified osd.

       Usage:

	  ceph osd repair <who>

       Subcommand reweight reweights osd to 0.0	< <weight> < 1.0.

       Usage:

	  osd reweight <int[0-]> <float[0.0-1.0]>

       Subcommand reweight-by-pg  reweight  OSDs  by  PG  distribution	[over-
       load-percentage-for-consideration, default 120].

       Usage:

	  ceph osd reweight-by-pg {<int[100-]>}	{<poolname> [<poolname...]}
	  {--no-increasing}

       Subcommand  reweight-by-utilization reweight OSDs by utilization	[over-
       load-percentage-for-consideration, default 120].

       Usage:

	  ceph osd reweight-by-utilization {<int[100-]>}
	  {--no-increasing}

       Subcommand rm removes osd(s) <id> [<id>...] in the cluster.

       Usage:

	  ceph osd rm <ids> [<ids>...]

       Subcommand scrub	initiates scrub	on specified osd.

       Usage:

	  ceph osd scrub <who>

       Subcommand set sets <key>.

       Usage:

	  ceph osd set full|pause|noup|nodown|noout|noin|nobackfill|
	  norebalance|norecover|noscrub|nodeep-scrub|notieragent

       Subcommand setcrushmap sets crush map from input	file.

       Usage:

	  ceph osd setcrushmap

       Subcommand setmaxosd sets new maximum osd value.

       Usage:

	  ceph osd setmaxosd <int[0-]>

       Subcommand stat prints summary of OSD map.

       Usage:

	  ceph osd stat

       Subcommand tier is used for managing tiers.  It	uses  some  additional
       subcommands.

       Subcommand  add	adds the tier <tierpool> (the second one) to base pool
       <pool> (the first one).

       Usage:

	  ceph osd tier	add <poolname> <poolname> {--force-nonempty}

       Subcommand add-cache adds a cache <tierpool> (the second	one)  of  size
       <size> to existing pool <pool> (the first one).

       Usage:

	  ceph osd tier	add-cache <poolname> <poolname>	<int[0-]>

       Subcommand cache-mode specifies the caching mode	for cache tier <pool>.

       Usage:

	  ceph osd tier	cache-mode <poolname> none|writeback|forward|readonly|
	  readforward|readproxy

       Subcommand  remove  removes  the	 tier <tierpool> (the second one) from
       base pool <pool>	(the first one).

       Usage:

	  ceph osd tier	remove <poolname> <poolname>

       Subcommand remove-overlay  removes  the	overlay	 pool  for  base  pool
       <pool>.

       Usage:

	  ceph osd tier	remove-overlay <poolname>

       Subcommand  set-overlay set the overlay pool for	base pool <pool> to be
       <overlaypool>.

       Usage:

	  ceph osd tier	set-overlay <poolname> <poolname>

       Subcommand tree prints OSD tree.

       Usage:

	  ceph osd tree	{<int[0-]>}

       Subcommand unpause unpauses osd.

       Usage:

	  ceph osd unpause

       Subcommand unset	unsets <key>.

       Usage:

	  ceph osd unset full|pause|noup|nodown|noout|noin|nobackfill|
	  norebalance|norecover|noscrub|nodeep-scrub|notieragent

   pg
       It is used for managing the placement groups in OSDs. It	uses some  ad-
       ditional	subcommands.

       Subcommand debug	shows debug info about pgs.

       Usage:

	  ceph pg debug	unfound_objects_exist|degraded_pgs_exist

       Subcommand deep-scrub starts deep-scrub on <pgid>.

       Usage:

	  ceph pg deep-scrub <pgid>

       Subcommand  dump	 shows	human-readable	versions of pg map (only 'all'
       valid with plain).

       Usage:

	  ceph pg dump {all|summary|sum|delta|pools|osds|pgs|pgs_brief}	[{all|summary|sum|delta|pools|osds|pgs|pgs_brief...]}

       Subcommand dump_json shows human-readable version of  pg	 map  in  json
       only.

       Usage:

	  ceph pg dump_json {all|summary|sum|delta|pools|osds|pgs|pgs_brief} [{all|summary|sum|delta|pools|osds|pgs|pgs_brief...]}

       Subcommand dump_pools_json shows	pg pools info in json only.

       Usage:

	  ceph pg dump_pools_json

       Subcommand dump_stuck shows information about stuck pgs.

       Usage:

	  ceph pg dump_stuck {inactive|unclean|stale|undersized|degraded [inactive|unclean|stale|undersized|degraded...]}
	  {<int>}

       Subcommand force_create_pg forces creation of pg	<pgid>.

       Usage:

	  ceph pg force_create_pg <pgid>

       Subcommand getmap gets binary pg	map to -o/stdout.

       Usage:

	  ceph pg getmap

       Subcommand ls lists pg with specific pool, osd, state

       Usage:

	  ceph pg ls {<int>} {active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale| remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized [active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale|remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized...]}

       Subcommand ls-by-osd lists pg on	osd [osd]

       Usage:

	  ceph pg ls-by-osd <osdname (id|osd.id)> {<int>}
	  {active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale| remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized [active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale|remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized...]}

       Subcommand ls-by-pool lists pg with pool	= [poolname]

       Usage:

	  ceph pg ls-by-pool <poolstr> {<int>} {active|
	  clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale| remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized [active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale|remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized...]}

       Subcommand ls-by-primary	lists pg with primary =	[osd]

       Usage:

	  ceph pg ls-by-primary	<osdname (id|osd.id)> {<int>}
	  {active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale| remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized [active|clean|down|replay|splitting|
	  scrubbing|scrubq|degraded|inconsistent|peering|repair|
	  recovery|backfill_wait|incomplete|stale|remapped|
	  deep_scrub|backfill|backfill_toofull|recovery_wait|
	  undersized...]}

       Subcommand map shows mapping of pg to osds.

       Usage:

	  ceph pg map <pgid>

       Subcommand repair starts	repair on <pgid>.

       Usage:

	  ceph pg repair <pgid>

       Subcommand scrub	starts scrub on	<pgid>.

       Usage:

	  ceph pg scrub	<pgid>

       Subcommand set_full_ratio sets ratio at which pgs are considered	full.

       Usage:

	  ceph pg set_full_ratio <float[0.0-1.0]>

       Subcommand  set_backfillfull_ratio  sets	ratio at which pgs are consid-
       ered too	full to	backfill.

       Usage:

	  ceph pg set_backfillfull_ratio <float[0.0-1.0]>

       Subcommand set_nearfull_ratio sets ratio	at which  pgs  are  considered
       nearly full.

       Usage:

	  ceph pg set_nearfull_ratio <float[0.0-1.0]>

       Subcommand stat shows placement group status.

       Usage:

	  ceph pg stat

   quorum
       Cause MON to enter or exit quorum.

       Usage:

	  ceph quorum enter|exit

       Note:  this  only  works	 on  the MON to	which the ceph command is con-
       nected.	If you want a specific MON to enter or exit quorum,  use  this
       syntax:

	  ceph tell mon.<id> quorum enter|exit

   quorum_status
       Reports status of monitor quorum.

       Usage:

	  ceph quorum_status

   report
       Reports full status of cluster, optional	title tag strings.

       Usage:

	  ceph report {<tags> [<tags>...]}

   scrub
       Scrubs the monitor stores.

       Usage:

	  ceph scrub

   status
       Shows cluster status.

       Usage:

	  ceph status

   sync	force
       Forces sync of and clear	monitor	store.

       Usage:

	  ceph sync force {--yes-i-really-mean-it} {--i-know-what-i-am-doing}

   tell
       Sends a command to a specific daemon.

       Usage:

	  ceph tell <name (type.id)> <args> [<args>...]

   version
       Show mon	daemon version

       Usage:

	  ceph version

OPTIONS
       -i infile
	      will  specify an input file to be	passed along as	a payload with
	      the command to the monitor cluster. This is only used  for  spe-
	      cific monitor commands.

       -o outfile
	      will  write any payload returned by the monitor cluster with its
	      reply to outfile.	 Only  specific	 monitor  commands  (e.g.  osd
	      getmap) return a payload.

       -c ceph.conf, --conf=ceph.conf
	      Use   ceph.conf	configuration  file  instead  of  the  default
	      /etc/ceph/ceph.conf  to  determine  monitor   addresses	during
	      startup.

       --id CLIENT_ID, --user CLIENT_ID
	      Client id	for authentication.

       --name CLIENT_NAME, -n CLIENT_NAME
	      Client name for authentication.

       --cluster CLUSTER
	      Name of the Ceph cluster.

       --admin-daemon ADMIN_SOCKET, daemon DAEMON_NAME
	      Submit admin-socket commands via admin sockets in	/var/run/ceph.

       --admin-socket ADMIN_SOCKET_NOPE
	      You probably mean	--admin-daemon

       -s, --status
	      Show cluster status.

       -w, --watch
	      Watch live cluster changes.

       --watch-debug
	      Watch debug events.

       --watch-info
	      Watch info events.

       --watch-sec
	      Watch security events.

       --watch-warn
	      Watch warning events.

       --watch-error
	      Watch error events.

       --version, -v
	      Display version.

       --verbose
	      Make verbose.

       --concise
	      Make less	verbose.

       -f {json,json-pretty,xml,xml-pretty,plain}, --format
	      Format of	output.

       --connect-timeout CLUSTER_TIMEOUT
	      Set a timeout for	connecting to the cluster.

       --no-increasing
	      --no-increasing  is off by default. So increasing	the osd	weight
	      is    allowed    using	the	reweight-by-utilization	    or
	      test-reweight-by-utilization  commands.	If this	option is used
	      with these commands, it will help	not  to	 increase  osd	weight
	      even the osd is under utilized.

AVAILABILITY
       ceph  is	 part  of Ceph,	a massively scalable, open-source, distributed
       storage	system.	 Please	  refer	  to   the   Ceph   documentation   at
       http://ceph.com/docs for	more information.

SEE ALSO
       ceph-mon(8), ceph-osd(8), ceph-mds(8)

COPYRIGHT
       2010-2014,  Inktank Storage, Inc. and contributors. Licensed under Cre-
       ative Commons BY-SA

dev				 Jul 08, 2017			       CEPH(8)

NAME | SYNOPSIS | DESCRIPTION | COMMANDS | OPTIONS | AVAILABILITY | SEE ALSO | COPYRIGHT

Want to link to this manual page? Use this URL:
<https://www.freebsd.org/cgi/man.cgi?query=ceph&sektion=8&manpath=FreeBSD+12.0-RELEASE+and+Ports>

home | help