Thanks for all buddies that replied my messages.
Indeed I used
ceph osd primary-affinity <osd-id> <weight>
And we felt some performance increment.
What's help here is that we have 6 proxmox ceph server:
ceph01 - HDD with 5 900 rpm
ceph02 - HDD with 7 200 rpm
ceph03 - HDD with 7 200 rpm
ceph04 - HDD with 7 200 rpm
ceph05 - HDD with 5 900 rpm
ceph06 - HDD with 5 900 rpm
So what I do is define weight 0 to HDD's with 5 900 rpm and define weight 1
to HDD's with 7 200 rpm.
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 62.31059 root default
-3 14.55438 host pve-ceph01
0 hdd 3.63860 osd.0 up 1.00000 0
1 hdd 3.63860 osd.1 up 1.00000 0
2 hdd 3.63860 osd.2 up 1.00000 0
3 hdd 3.63860 osd.3 up 1.00000 0
-5 10.91559 host pve-ceph02
4 hdd 2.72890 osd.4 up 1.00000 1.00000
5 hdd 2.72890 osd.5 up 1.00000 1.00000
6 hdd 2.72890 osd.6 up 1.00000 1.00000
7 hdd 2.72890 osd.7 up 1.00000 1.00000
-7 7.27708 host pve-ceph03
8 hdd 2.72890 osd.8 up 1.00000 1.00000
9 hdd 2.72890 osd.9 up 1.00000 1.00000
10 hdd 1.81929 osd.10 up 1.00000 1.00000
-9 7.27716 host pve-ceph04
11 hdd 1.81929 osd.11 up 1.00000 1.00000
12 hdd 1.81929 osd.12 up 1.00000 1.00000
13 hdd 1.81929 osd.13 up 1.00000 1.00000
14 hdd 1.81929 osd.14 up 1.00000 1.00000
-11 14.55460 host pve-ceph05
15 hdd 7.27730 osd.15 up 1.00000 0
16 hdd 7.27730 osd.16 up 1.00000 0
-13 7.73178 host pve-ceph06
17 hdd 0.90959 osd.17 up 1.00000 0
18 hdd 2.72890 osd.18 up 1.00000 0
19 hdd 1.36440 osd.19 up 1.00000 0
20 hdd 2.72890 osd.20 up 1.00000 0
Tha's it! Thanks again.
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36
Post by Phil SchwarzHope you did change a single disk at a time !
Be warned (if not) that moving an OSD from a server to another triggers
a rebalancing of almost the complete datas stored upon in order to
follow crushmap.
For instance exchanging two OSDs between servers result in a complete
rebalance of the two OSDS,a ccording to my knowledge.
16% of misplaced datas could be acceptable or not depending on your
needs of redundancy and throughput, but it's not a low value that could
be underestimated.
Best regards
Post by Gilberto NunesRight now the ceph are very slow
343510/2089155 objects misplaced (16.443%)
Status
HEALTH_WARN
Monitors
OSDs
In Out
Up 21 0
Down 0 0
Total: 21
PGs
157
1
82
2
8
Usage
7.68 TiB of 62.31 TiB
<http://www.proxmox.com/products/proxmox-ve/subscription-service-plans>
()
Degraded data redundancy: 21495/2089170 objects degraded (1.029%), 8 pgs
degraded, 8 pgs undersized
pg 21.0 is stuck undersized for 63693.346103, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,9]
pg 21.2 is stuck undersized for 63693.346973, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,10]
pg 21.6f is stuck undersized for 62453.277248, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,5]
pg 21.8b is stuck undersized for 63693.361835, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,8]
pg 21.c3 is stuck undersized for 63693.321337, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,9]
pg 21.c5 is stuck undersized for 66587.797684, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,8]
pg 21.d4 is stuck undersized for 62453.047415, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,6]
pg 21.e1 is stuck undersized for 62453.276631, current state
active+undersized+degraded+remapped+backfill_wait, last acting [2,5]
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36
Post by Gilberto NunesSO, what you guys think about this HDD distribuiton?
CEPH-01
1x 3 TB
1x 2 TB
CEPH-02
1x 4 TB
1x 3 TB
CEPH-03
1x 4 TB
1x 3 TB
CEPH-04
1x 4 TB
1x 3 TB
1x 2 TB
CEPH-05
1x 8 TB
1x 2 TB
CEPH-06
1x 3 TB
1x 1 TB
1x 8 TB
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36