ceph维护.docx

上传人:b****6 文档编号:8596809 上传时间:2023-02-01 格式:DOCX 页数:22 大小:23.43KB
下载 相关 举报
ceph维护.docx_第1页
第1页 / 共22页
ceph维护.docx_第2页
第2页 / 共22页
ceph维护.docx_第3页
第3页 / 共22页
ceph维护.docx_第4页
第4页 / 共22页
ceph维护.docx_第5页
第5页 / 共22页
点击查看更多>>
下载资源
资源描述

ceph维护.docx

《ceph维护.docx》由会员分享,可在线阅读,更多相关《ceph维护.docx(22页珍藏版)》请在冰豆网上搜索。

ceph维护.docx

ceph维护

Ceph维护命令(持续更新中)eddy

一、集群

1、启动一个ceph进程

启动mon进程

servicecephstart mon.node1

启动msd进程

servicecephstartmds.node1

启动osd进程

 servicecephstartosd.0

2、查看机器的监控状态

[root@client~]#cephhealth

HEALTH_OK

3、查看ceph的实时运行状态

[root@client~]#ceph-w

  clusterbe1756f2-54f7-4d8f-8790-820c82721f17

   healthHEALTH_OK

   monmape2:

3monsat{node1=10.240.240.211:

6789/0,node2=10.240.240.212:

6789/0,node3=10.240.240.213:

6789/0},electionepoch294,quorum0,1,2node1,node2,node3

   mdsmape95:

1/1/1up{0=node2=up:

active},1up:

standby

   osdmape88:

3osds:

3up,3in

   pgmapv1164:

448pgs,4pools,10003MBdata,2520objects

      23617MBused,37792MB/61410MBavail

         448active+clean

2014-06-3000:

48:

28.756948mon.0[INF]pgmapv1163:

448pgs:

448active+clean;10003MBdata,23617MBused,37792MB/61410MBavail

4、检查信息状态信息

[root@client~]#ceph-s

  clusterbe1756f2-54f7-4d8f-8790-820c82721f17

   healthHEALTH_OK

   monmape2:

3monsat{node1=10.240.240.211:

6789/0,node2=10.240.240.212:

6789/0,node3=10.240.240.213:

6789/0},electionepoch294,quorum0,1,2node1,node2,node3

   mdsmape95:

1/1/1up{0=node2=up:

active},1up:

standby

   osdmape88:

3osds:

3up,3in

   pgmapv1164:

448pgs,4pools,10003MBdata,2520objects

      23617MBused,37792MB/61410MBavail

         448active+clean

[root@client~]# 

5、查看ceph存储空间

[root@client~]#cephdf

GLOBAL:

  SIZE   AVAIL   RAWUSED  %RAWUSED 

  61410M  37792M  23617M   38.46   

POOLS:

  NAME    ID  USED   %USED  OBJECTS 

  data    0   10000M  16.28  2500  

  metadata  1   3354k   0    20   

  rbd     2   0     0    0    

  jiayuan   3   0     0    0    

[root@client~]# 

6、删除一个节点的所有的ceph数据包

[root@node1~]#ceph-deploypurgenode1

[root@node1~]#ceph-deploypurgedatanode1

7、为ceph创建一个admin用户并为admin用户创建一个密钥,把密钥保存到/etc/ceph目录下:

cephauthget-or-createclient.adminmds'allow'osd'allow*'mon'allow*'>/etc/ceph/ceph.client.admin.keyring

cephauthget-or-createclient.adminmds'allow'osd'allow*'mon'allow*'-o/etc/ceph/ceph.client.admin.keyring

8、为osd.0创建一个用户并创建一个key

cephauthget-or-createosd.0mon'allowrwx'osd'allow*'-o/var/lib/ceph/osd/ceph-0/keyring

9、为mds.node1创建一个用户并创建一个key

cephauthget-or-createmds.node1mon'allowrwx'osd'allow*'mds'allow*'-o/var/lib/ceph/mds/ceph-node1/keyring

10、查看ceph集群中的认证用户及相关的key

cephauthlist

11、删除集群中的一个认证用户

cephauthdelosd.0

12、查看集群的详细配置

[root@node1~]#cephdaemonmon.node1configshow|more

13、查看集群健康状态细节

[root@admin~]#cephhealthdetail

HEALTH_WARN12pgsdown;12pgspeering;12pgsstuckinactive;12pgsstuckunclean

pg3.3bisstuckinactivesinceforever,currentstatedown+peering,lastacting[1,2]

pg3.36isstuckinactivesinceforever,currentstatedown+peering,lastacting[1,2]

pg3.79isstuckinactivesinceforever,currentstatedown+peering,lastacting[1,0]

pg3.5isstuckinactivesinceforever,currentstatedown+peering,lastacting[1,2]

pg3.30isstuckinactivesinceforever,currentstatedown+peering,lastacting[1,2]

pg3.1aisstuckinactivesinceforever,currentstatedown+peering,lastacting[1,0]

pg3.2disstuckinactivesinceforever,currentstatedown+peering,lastacting[1,0]

pg3.16isstuckinactivesinceforever,currentstatedown+peering,lastacting[1,2]

14、查看cephlog日志所在的目录

[root@node1~]#ceph-conf--namemon.node1--show-config-valuelog_file

/var/log/ceph/ceph-mon.node1.log

 

二、mon

1、查看mon的状态信息

[root@client~]#cephmonstat

e2:

3monsat{node1=10.240.240.211:

6789/0,node2=10.240.240.212:

6789/0,node3=10.240.240.213:

6789/0},electionepoch294,quorum0,1,2node1,node2,node3

2、查看mon的选举状态

[root@client~]#cephquorum_status

{"election_epoch":

294,"quorum":

[0,1,2],"quorum_names":

["node1","node2","node3"],"quorum_leader_name":

"node1","monmap":

{"epoch":

2,"fsid":

"be1756f2-54f7-4d8f-8790-820c82721f17","modified":

"2014-06-2618:

43:

51.671106","created":

"0.000000","mons":

[{"rank":

0,"name":

"node1","addr":

"10.240.240.211:

6789\/0"},{"rank":

1,"name":

"node2","addr":

"10.240.240.212:

6789\/0"},{"rank":

2,"name":

"node3","addr":

"10.240.240.213:

6789\/0"}]}}

3、查看mon的映射信息

[root@client~]#cephmondump

dumpedmonmapepoch2

epoch2

fsidbe1756f2-54f7-4d8f-8790-820c82721f17

last_changed2014-06-2618:

43:

51.671106

created0.000000

0:

10.240.240.211:

6789/0mon.node1

1:

10.240.240.212:

6789/0mon.node2

2:

10.240.240.213:

6789/0mon.node3

4、删除一个mon节点

[root@node1~]#cephmonremovenode1

removedmon.node1at10.39.101.1:

6789/0,therearenow3monitors

2014-07-0718:

11:

04.9741887f4d16bfd700 0monclient:

huntingfornewmon

5、获得一个正在运行的monmap,并保存在1.txt文件中

[root@node3~]#cephmongetmap-o1.txt

gotmonmapepoch6

6、查看上面获得的map

[root@node3~]# monmaptool--print1.txt 

monmaptool:

monmapfile1.txt

epoch6

fsid92552333-a0a8-41b8-8b45-c93a8730525e

last_changed2014-07-0718:

22:

51.927205

created0.000000

0:

10.39.101.1:

6789/0mon.node1

1:

10.39.101.2:

6789/0mon.node2

2:

10.39.101.3:

6789/0mon.node3

[root@node3~]#

7、把上面的monmap注入新加入的节点

ceph-mon-inode4--inject-monmap1.txt

8、查看mon的aminsocket

root@node1~]#ceph-conf--namemon.node1--show-config-valueadmin_socket

/var/run/ceph/ceph-mon.node1.asok 

9、查看mon的详细状态

[root@node1~]#cephdaemonmon.node1 mon_status 

{"name":

"node1",

 "rank":

0,

 "state":

"leader",

 "election_epoch":

96,

 "quorum":

[

    0,

    1,

    2],

 "outside_quorum":

[],

 "extra_probe_peers":

[

    "10.39.101.4:

6789\/0"],

 "sync_provider":

[],

 "monmap":

{"epoch":

6,

   "fsid":

"92552333-a0a8-41b8-8b45-c93a8730525e",

   "modified":

"2014-07-0718:

22:

51.927205",

   "created":

"0.000000",

   "mons":

[

      {"rank":

0,

       "name":

"node1",

       "addr":

"10.39.101.1:

6789\/0"},

      {"rank":

1,

       "name":

"node2",

       "addr":

"10.39.101.2:

6789\/0"},

      {"rank":

2,

       "name":

"node3",

       "addr":

"10.39.101.3:

6789\/0"}]}

10、删除一个mon节点

[root@os-node1~]#cephmonremoveos-node1

removedmon.os-node1at10.40.10.64:

6789/0,therearenow3monitors

三、msd

1、查看msd状态

[root@client~]#cephmdsstat

e95:

1/1/1up{0=node2=up:

active},1up:

standby

2、查看msd的映射信息

[root@client~]#cephmdsdump

dumpedmdsmapepoch95

epoch 95

flags 0

created2014-06-2618:

41:

57.686801

modified    2014-06-3000:

24:

11.749967

tableserver  0

root  0

session_timeout60

session_autoclose   300

max_file_size 1099511627776

last_failure  84

last_failure_osd_epoch 81

compat compat={},rocompat={},incompat={1=basev0.20,2=clientwriteableranges,3=defaultfilelayoutsondirs,4=dirinodeinseparateobject,5=mdsusesversionedencoding,6=dirfragisstoredinomap}

max_mds1

in   0

up   {0=5015}

failed

stopped

data_pools   0

metadata_pool 1

inline_data  disabled

5015:

 10.240.240.212:

6808/3032'node2'mds.0.12up:

activeseq30

5012:

 10.240.240.211:

6807/3459'node1'mds.-1.0up:

standbyseq38

3、删除一个mds节点

[root@node1~]#ceph mdsrm0mds.node1

 mdsgid0dne

四、osd

1、查看cephosd运行状态

[root@client~]#cephosdstat

   osdmape88:

3osds:

3up,3in

2、查看osd映射信息

[root@client~]#cephosddump

epoch88

fsidbe1756f2-54f7-4d8f-8790-820c82721f17

created2014-06-2618:

41:

57.687442

modified2014-06-3000:

46:

27.179793

flags 

pool0'data'replicatedsize2min_size1crush_ruleset0object_hashrjenkinspg_num64pgp_num64last_change1owner0flagshashpspoolcrash_replay_interval45stripe_width0

pool1'metadata'replicatedsize2min_size1crush_ruleset0object_hashrjenkinspg_num64pgp_num64last_change1owner0flagshashpspoolstripe_width0

pool2'rbd'replicatedsize2min_size1crush_ruleset0object_hashrjenkinspg_num64pgp_num64last_change1owner0flagshashpspoolstripe_width0

pool3'jiayuan'replicatedsize2min_size1crush_ruleset0object_hashrjenkinspg_num256pgp_num256last_change73owner0flagshashpspoolstripe_width0

max_osd3

osd.0up in weight1up_from65up_thru75down_at64last_clean_interval[53,55)10.240.240.211:

6800/308910.240.240.211:

6801/308910.240.240.211:

6802/308910.240.240.211:

6803/3089exists,up8a24ad16-a483-4bac-a56a-6ed44ab74ff0

osd.1up in weight1up_from59up_thru74down_at58last_clean_interval[31,55)10.240.240.212:

6800/269610.240.240.212:

6801/269610.240.240.212:

6802/269610.240.240.212:

6803/2696exists,up8619c083-0273-4203-ba57-4b1dabb89339

osd.2up in weight1up_from62up_thru74down_at61last_clean_interval[39,55)10.240.240.213:

6800/266210.240.240.213:

6801/266210.240.240.213:

6802/266210.240.240.213:

6803/2662exists,upf8107c04-35d7-4fb8-8c82-09eb885f0e58

[root@client~]# 

3、查看osd的目录树

[root@client~]#cephosdtree

#id  weight typename   up/downreweight

-1   3   rootdefault

-2   1       hostnode1

0   1           osd.0 up   1

-3   1       hostnode2

1   1           osd.1 up   1

-4   1       hostnode3

2   1           osd.2 up   1

4、down掉一个osd硬盘

[root@node1~]#cephosddown0 #down掉osd.0节点

5、在集群中删除一个osd硬盘

[root@node4~]#cephosdrm0

removedosd.0

6、在集群中删除一个osd硬盘crushmap

[root@node1~]#cephosdcrushrmosd.0

7、在集群中删除一个osd的host节点

[root@node1~]#cephosdcrushrmnode1

removeditemid-2name'node1'fromcrushmap

查看最大osd的个数 

[root@node1~]#cephosdgetmaxosd

max_osd=4inepoch514     #默认最大是4个osd节点

8、设置最大的osd的个数(当扩大osd节点的时候必须扩大这个值)

[root@node1~]#cephosdsetmaxosd10

9、设置osdcrush的权重为1.0

cephosdcrushset{id}{weight}[{loc1}[{loc2}...]]

例如:

[root@admin~]#cephosdcrushset33.0host=node

展开阅读全文
相关资源
猜你喜欢
相关搜索

当前位置:首页 > 经管营销 > 生产经营管理

copyright@ 2008-2022 冰豆网网站版权所有

经营许可证编号:鄂ICP备2022015515号-1