[ceph-users] Re: Files listed in radosgw BI but is not available in ceph

2021-07-22 Thread Rafael Lopez
Thanks for further clarification Dan. Boris, if you have a test/QA environment on the same code as production, you can confirm if the problem is as above. *Do NOT do this in production* - if the problem exists it might result in losing production data. 1. Upload large S3 object that would take

[ceph-users] Re: How to make CephFS a tiered file system?

2021-07-22 Thread Patrick Donnelly
Yes. On Thu, Jul 22, 2021 at 3:23 PM Calhoun, Patrick wrote: > > Do I understand correctly that to relocate a single file to a different pool, > the process would be: > > setfattr -n ceph.dir.layout.pool -v NewPool original_file_name > cp -a original_file_name .hidden_file_name && mv -f

[ceph-users] Re: How to make CephFS a tiered file system?

2021-07-22 Thread Calhoun, Patrick
Do I understand correctly that to relocate a single file to a different pool, the process would be: setfattr -n ceph.dir.layout.pool -v NewPool original_file_name cp -a original_file_name .hidden_file_name && mv -f .hidden_file_name original_file_name -Patrick

[ceph-users] Re: How to make CephFS a tiered file system?

2021-07-22 Thread Patrick Donnelly
On Wed, Jul 21, 2021 at 1:49 PM huxia...@horebdata.cn wrote: > > Dear Patrick, > > Thanks a lot for pointing out the HSM ticket. We will see whether we have the > resource to do something with the ticket. > > I am thinking of a temporary solution for HSM using cephfs client commands. > The

[ceph-users] Re: Pacific 16.2.5 Dashboard minor regression

2021-07-22 Thread Chris Palmer
Ok... We haven't used cephadm/orch at all - it was a very straightforward manual install. But looking at the PR it seems that container-related complexities are now spilling over into traditional deployments too. This one is easily worked around. But I hope we don't keep getting these

[ceph-users] Installing and Configuring RGW to an existing cluster

2021-07-22 Thread Matt Dunavant
Hi all, We are currently using a ceph cluster for block storage on version 14.2.16. We would like to start experimenting with object storage but the ceph documentation doesn't seem to cover a lot of the installation or configuration of the RGW piece. Does anybody know where I may be able to

[ceph-users] Re: Pacific 16.2.5 Dashboard minor regression

2021-07-22 Thread Ernesto Puerta
Hey Chris, Nothing dumb from your end. Cephadm has recently changed the way it configures the network endpoints favoring IP addresses over hostnames. I think this was the change: https://github.com/ceph/ceph/pull/41483/ I may suggest you to use a SAN with the IP address/es in your certificate?

[ceph-users] Re: RHCS 4.1 with grafana and prometheus with Node exporter.

2021-07-22 Thread Ernesto Puerta
Hi Ramanathan, As Sebastian said, if you're using a paid subscription I'd recommend you to follow that channel. That said, Grafana 8 is not going to work with the provided Dashboards, as AFAIK it doesn't support some of the plugins we use here. Master/pacific is tested against Grafana 6.7.

[ceph-users] Re: Procedure for changing IP and domain name of all nodes of a cluster

2021-07-22 Thread Eugen Block
Note that there's a similar field in the nova database (connection_info): ---snip--- MariaDB [nova]> select connection_info from block_device_mapping where instance_uuid='bbc33a1d-10c0-47b1-8179-304899c4546c';

[ceph-users] Can't clear UPGRADE_REDEPLOY_DAEMON after fix

2021-07-22 Thread Arnaud MARTEL
Hi, I just upgraded my cluster from 16.2.4 to 16.2.5 and I had an error during the upgrade of the first osd daemon (cf below). I fixed the error (I just created the missing directory on all hosts), then resume the upgrade. Now, everything is OK but I still have a warning:" [WRN]

[ceph-users] Re: Using CephFS in High Performance (and Throughput) Compute Use Cases

2021-07-22 Thread Mark Nelson
Hi Dan, Ah, that's fantastic regarding IOR.  Have you tried the libcephfs backend?  That might be another route for easy testing (and at least on our previous test setup I saw higher large sequential IO throughput with it vs the kernel client).  Lazy IO is definitely worth it if you have an

[ceph-users] Pacific 16.2.5 Dashboard minor regression

2021-07-22 Thread Chris Palmer
Since updating from Pacific 16.2.4 -> 16.2.5 I've noticed a behaviour change in the Dashboard. If I connect to the active MGR, it is fine. However if I connect to a standby MGR, it redirects to the active one by placing the active IP address in the URL, rather than the active hostname as it

[ceph-users] Re: Fwd: Re: Issues with Ceph network redundancy using L2 MC-LAG

2021-07-22 Thread Frank Schilder
Hi, I'm a bit late to the party. I use 6x10G active/active LACP bonds on Dell switches and servers and also observe very bad behaviour when a link is flapping. First, I get "long ping time" warnings and a lot of ops are stuck. It usually takes several minutes until the kernel/switch starts

[ceph-users] Cephadm: How to remove a stray daemon ghost

2021-07-22 Thread Kai Stian Olstad
Hi I have a warning that says "1 stray daemon(s) not managed by cephadm" What i did is the following. I have 3 nodes that the mon should run on, but because of a bug in 16.2.4 I couldn't run on then since they are in different subnet. But this was fixed in 16.2.5 so i upgraded without issues.

[ceph-users] Re: Huge headaches with NFS and ingress HA failover

2021-07-22 Thread Jeff Layton
On Wed, 2021-07-21 at 18:28 +0200, Andreas Weisker wrote: > Hi, > > we recently set up a new pacific cluster with cephadm. > Deployed nfs on two hosts and ingress on two other hosts. (ceph orch > apply for nfs and ingress like on the docs page) > > So far so good. ESXi with NFS41 connects, but

[ceph-users] Re: Files listed in radosgw BI but is not available in ceph

2021-07-22 Thread Dan van der Ster
Boris, To check if your issue is related to Rafael's, could you check your access logs for requests on the missing objects which lasted longer than one hour? I ask because Nautilus also has rgw_gc_obj_min_wait (2hr by default), which is the main config option related to

[ceph-users] Re: Files listed in radosgw BI but is not available in ceph

2021-07-22 Thread Dan van der Ster
Hi Rafael, AFAIU, that gc issue was not relevant for N -- the bug is in the new rgw_gc code which landed in Octopus and was not backported to N. Well, RHCEPH had the new rgw_gc cls backported to it, and RHCEPH has the bugfix you refer to: * Wed Dec 02 2020 Ceph Jenkins 2:14.2.11-86 - rgw:

[ceph-users] Re: Files listed in radosgw BI but is not available in ceph

2021-07-22 Thread Rafael Lopez
hi boris, We hit an issue late last year that sounds similar to what you are experiencing. I am not sure if the fix was backported to nautilus, I can't see any reference to a nautilus backport so it's possible it was only backported to octopus (15.x), exception being red hat ceph nautilus.

[ceph-users] Re: Using CephFS in High Performance (and Throughput) Compute Use Cases

2021-07-22 Thread Dan van der Ster
Hi Mark and all, The key point is to consider your users' write requirements: do your applications need to write concurrently to the same file from several cephfs mounts? or does each job write to a separate file? If your use-case is predominantly the latter, you'll have a lot of success right

[ceph-users] Re: Procedure for changing IP and domain name of all nodes of a cluster

2021-07-22 Thread Burkhard Linke
Hi, On 7/21/21 8:30 PM, Konstantin Shalygin wrote: Hi, On 21 Jul 2021, at 10:53, Burkhard Linke > wrote: One client with special needs is openstack cinder. The database entries contain the mon list for volumes Another question: do

[ceph-users] Re: nobody in control of ceph csi development?

2021-07-22 Thread Ilya Dryomov
On Wed, Jul 21, 2021 at 4:30 PM Marc wrote: > > Crappy code continues to live on? > > This issue has been automatically marked as stale because it has not had > recent activity. It will be closed in a week if no further activity occurs. > Thank you for your contributions. Hi Marc, Which issue