> > Removed 488729 objects
> > Clean up completed and total clean up time :8.84114
> >
> >
> >
> >
> >
> >
> >
> >
> > On Wed, Aug 23, 2023 at 1:25 PM Adam King wrote:
> >
> >> this should be possible by specifying a
ut of "ceph
> orch device ls --format json | jq" to see things like what cephadm
> considers the model, size etc. for the devices to be for use in the
> filtering.
>
> On Wed, Aug 23, 2023 at 1:13 PM Satish Patel wrote:
>
>> Folks,
>>
>> I have 3 nodes w
Folks,
I have 3 nodes with each having 1x NvME (1TB) and 3x 2.9TB SSD. Trying to
build ceph storage using cephadm on Ubuntu 22.04 distro.
If I want to use NvME for Journaling (WAL/DB) for my SSD based OSDs then
how does cephadm handle it?
Trying to find a document where I can tell cephadm to
Folks,
I am trying to install ceph on 10 node clusters and planning to use
cephadm. My question is if next year i will add new nodes to this cluster
then what docker image version cephadm will use to add new nodes?
Are there any local registry can i create one to copy images locally? How
does
Folks,
Any idea what is going on, I am running 3 node quincy version of openstack
and today suddenly i noticed the following error. I found reference link
but not sure if that is my issue or not
https://tracker.ceph.com/issues/51974
root@ceph1:~# ceph -s
cluster:
id:
adm rm-daemon —name osd.3‘. Or you try it with the
> orchestrator: ceph orch daemon rm…
> I don’t have the exact command at the moment, you should check the docs.
>
> Zitat von Satish Patel :
>
> > Hi Eugen,
> >
> > I have delected osd.3 directory from datastorn4 n
them. Do
> you see the osd directory in /var/lib/ceph//osd.3 on datastorn4?
> Just remove the osd.3 directory, then cephadm won't try to activate it.
>
>
> Zitat von Satish Patel :
>
> > Folks,
> >
> > I have deployed 15 OSDs node clusters using cephadm and
Folks,
I have deployed 15 OSDs node clusters using cephadm and encount duplicate
OSD on one of the nodes and am not sure how to clean that up.
root@datastorn1:~# ceph health
HEALTH_WARN 1 failed cephadm daemon(s); 1 pool(s) have no replicas
configured
osd.3 is duplicated on two nodes, i would
;
> Wat happens if you try to start a cephadm shell on that node?
>
>
> > -Oorspronkelijk bericht-
> > Van: Satish Patel
> > Verzonden: donderdag 29 september 2022 21:45
> > Aan: ceph-users
> > Onderwerp: [ceph-users] Re: strange osd error during add disk
not be your work hours. Please do not feel the
> need to respond during a time that is not convenient for you.
> --
> Great people talk about ideas,
> ordinary people talk about things,
> small people talk... about other people.
>
&
Bump! Any suggestions?
On Wed, Sep 28, 2022 at 4:26 PM Satish Patel wrote:
> Folks,
>
> I have 15 nodes for ceph and each node has a 160TB disk attached. I am
> using cephadm quincy release and all 14 nodes have been added except one
> node which is giving a very strange erro
Folks,
I have 15 nodes for ceph and each node has a 160TB disk attached. I am
using cephadm quincy release and all 14 nodes have been added except one
node which is giving a very strange error during adding it. I have put all
logs here https://paste.opendev.org/show/bbSKwlSLyANMbrlhwzXL/
In
ipe the disk properly first.
>
> Zitat von Satish Patel :
>
>> Folks,
>>
>> I have created a new lab using cephadm and installed a new 1TB spinning
>> disk which is trying to add in a cluster but somehow ceph is not detecting
>> it.
>>
>> $ part
Folks,
I have created a new lab using cephadm and installed a new 1TB spinning
disk which is trying to add in a cluster but somehow ceph is not detecting
it.
$ parted /dev/sda print
Model: ATA WDC WD10EZEX-00B (scsi)
Disk /dev/sda: 1000GB
Sector size (logical/physical): 512B/4096B
Partition
mon service and now i can't do anything because i
don't have mon. ceph -s and all other command hangs
Try to find out how to get back mon :)
On Fri, Sep 2, 2022 at 3:34 PM Satish Patel wrote:
> Yes, i have stopped upgrade and those log before upgrade
>
> On Fri, Sep 2, 2022 at 3:2
re having stopped the upgrade?
>
> On Fri, Sep 2, 2022 at 2:53 PM Satish Patel wrote:
>
>> Do you think this is because I have only a single MON daemon running? I
>> have only two nodes.
>>
>> On Fri, Sep 2, 2022 at 2:39 PM Satish Patel wrote:
>>
>>
Do you think this is because I have only a single MON daemon running? I
have only two nodes.
On Fri, Sep 2, 2022 at 2:39 PM Satish Patel wrote:
> Adam,
>
> I have enabled debug and my logs flood with the following. I am going to
> try some stuff from your provided mailing
e get more info on why it's not performing the
> redeploy from those debug logs. Just remember to set the log level back
> after 'ceph config set mgr mgr/cephadm/log_to_cluster_level info' as debug
> logs are quite verbose.
>
> On Fri, Sep 2, 2022 at 11:39 AM Satish Patel wrote:
>
a "ceph orch redeploy mgr" and then "ceph orch upgrade start --image
> quay.io/ceph/ceph:v16.2.10" and see if it goes better.
>
> On Fri, Sep 2, 2022 at 10:36 AM Satish Patel wrote:
>
>> Hi Adam,
>>
>> I run the following command to upgrade but i
r
progress:
Upgrade to quay.io/ceph/ceph:v16.2.10 (0s)
[........]
On Fri, Sep 2, 2022 at 10:25 AM Satish Patel wrote:
> It Looks like I did it with the following command.
>
> $ ceph orch daemon add mgr ceph2:10.73.0.192
>
> Now i can see two wit
:c08064dde4bba4e72a1f55d90ca32df9ef5aafab82efe2e0a0722444a5aaacca
93146564743f 294fd6ab6c97
On Fri, Sep 2, 2022 at 10:19 AM Satish Patel wrote:
> Let's come back to the original question: how to bring back the second mgr?
>
> root@ceph1:~# ceph orch apply mgr 2
> Scheduled mgr update...
>
> Nothing happened
-02T14:18:21.127973+ mgr.ceph1.smfvfd (mgr.334626) 17010 :
cephadm [INF] refreshing ceph2 facts
On Fri, Sep 2, 2022 at 10:15 AM Satish Patel wrote:
> Hi Adam,
>
> Wait..wait.. now it's working suddenly without doing anything.. very odd
>
> root@ceph1:~# ceph
e5a616e4b9cf
osd.osd_spec_default 4/0 5s ago -
quay.io/ceph/ceph:v15
93146564743f
prometheus1/1 5s ago 2w count:1
quay.io/prometheus/prometheus:v2.18.1
On Fri, Sep 2, 2022 at 10:13 AM Satish Patel wrote:
> I can see that in the out
remove the files.
>
> On Fri, Sep 2, 2022 at 9:57 AM Satish Patel wrote:
>
>> Hi Adam,
>>
>> I have deleted file located here - rm
>> /var/lib/ceph/f270ad9e-1f6f-11ed-b6f8-a539d87379ea/cephadm.7ce656a8721deb5054c37b0cfb90381522d521dde51fb0c5a2142314d663f63d
>&g
c37b0cfb9038
> 1522d521dde51fb0c5a2142314d663f63d (and any others like it) file would be
> the way forward to get orch ls working again.
>
> On Fri, Sep 2, 2022 at 9:44 AM Satish Patel wrote:
>
>> Hi Adam,
>>
>> In cephadm ls i found the following service but i believe it was there
>> b
thinks is this "cephadm" service that it has deployed.
> Lastly, you could try having the mgr you manually deploy be a 16.2.10 one
> instead of 15.2.17 (I'm assuming here, but the line numbers in that
> traceback suggest octopus). The 16.2.10 one is just much less likely to
>
gr/orchestrator/_interface.py", line 642, in
raise_if_exception
raise e
AssertionError: cephadm
On Fri, Sep 2, 2022 at 1:32 AM Satish Patel wrote:
> nevermind, i found doc related that and i am able to get 1 mgr up -
> https://docs.ceph.com/en/quincy/cephadm/troubleshooting/#ma
nevermind, i found doc related that and i am able to get 1 mgr up -
https://docs.ceph.com/en/quincy/cephadm/troubleshooting/#manually-deploying-a-mgr-daemon
On Fri, Sep 2, 2022 at 1:21 AM Satish Patel wrote:
> Folks,
>
> I am having little fun time with cephadm and it's very annoyin
Folks,
I am having little fun time with cephadm and it's very annoying to deal
with it
I have deployed a ceph cluster using cephadm on two nodes. Now when i was
trying to upgrade and noticed hiccups where it just upgraded a single mgr
with 16.2.10 but not other so i started messing around and
a relatively
> sage operation). In the end, the IP cephadm lists for each host in "ceph
> orch host ls" must be an IP that allows correctly ssh-ing to the host.
>
> On Thu, Sep 1, 2022 at 9:17 PM Satish Patel wrote:
>
>> Hi Adam,
>>
>> You are correct, look like
e on both ceph1 and ceph2 and some sort of networking issue is
> the only thing I'm aware of currently that causes something like that.
>
> On Thu, Sep 1, 2022 at 6:30 PM Satish Patel wrote:
>
>> Hi Adam,
>>
>> I have also noticed a very strange thing which is Duplicat
?
https://achchusnulchikam.medium.com/deploy-ceph-cluster-with-cephadm-on-centos-8-257b300e7b42
On Thu, Sep 1, 2022 at 6:20 PM Satish Patel wrote:
> Hi Adam,
>
> Getting the following error, not sure why it's not able to find it.
>
> root@ceph1:~# ceph orch daemon redeploy mgr.ceph1
loy mgr.ceph1.xmbvsb`?
>
> On Thu, Sep 1, 2022 at 5:12 PM Satish Patel wrote:
>
>> Hi Adam,
>>
>> Here is requested output
>>
>> root@ceph1:~# ceph health detail
>> HEALTH_WARN 4 stray daemon(s) not managed by cephadm
>> [WRN] CEPHADM_STRAY_DAEMON:
s reported as
> being on ceph1 in the orch ps output in your original message in this
> thread. I'm interested in what `ceph health detail` is reporting now as
> well, as it says there are 4 stray daemons. Also, the `ceph orch host ls`
> output just to get a better grasp of the topolo
"container_image_name": "quay.io/ceph/ceph:v15",
"container_image_id":
"93146564743febec815d6a764dad93fc07ce971e88315403ac508cb5da6d35f4",
"version": "15.2.17",
"started": "2022-08-31T03:2
Folks,
I am playing with cephadm and life was good until I started upgrading from
octopus to pacific. My upgrade process stuck after upgrading mgr and in
logs now i can see following error
root@ceph1:~# ceph log last cephadm
2022-09-01T14:40:45.739804+ mgr.ceph2.hmbdla (mgr.265806) 8 :
Hi,
I have a small cluster in the lab which has only two nodes. I have a single
monitor and two OSD nodes.
Running upgrade but somehow it stuck after upgrading mgr
ceph orch upgrade start --ceph-version 16.2.10
root@ceph1:~# ceph -s
cluster:
id: f270ad9e-1f6f-11ed-b6f8-a539d87379ea
Hi,
I believe only advantage of running dockerize which isolate binaries from
OS and as you said upgrade is easier, In my case i am running OSD/MON role
on same servers so it provide greater isolation when i want to upgrade
component.
cephadm uses containers to deploy ceph clusters in
des do you have for your cluster size? Do you have dedicated
or shared MDS with OSDs?
> I don't know if this is optimum, we are in testing process...
>
> - Mail original -
> > De: "Stefan Kooman"
> > À: "Jake Grimmett" , "Christian Wuerdig" &
NVMe drives.
>
>
Can you share what company NvME drives are you using?
> best regards,
>
> Jake
>
> On 20/06/2022 08:22, Stefan Kooman wrote:
> > On 6/19/22 23:23, Christian Wuerdig wrote:
> >> On Sun, 19 Jun 2022 at 02:29, Satish Patel
> wr
Greeting folks,
We are planning to build Ceph storage for mostly cephFS for HPC workload
and in future we are planning to expand to S3 style but that is yet to be
decided. Because we need mass storage, we bought the following HW.
15 Total servers and each server has a 12x18TB HDD (spinning disk)
41 matches
Mail list logo