On Wed, Jun 28, 2017 at 2:10 PM Matt Benjamin <[email protected]> wrote:
> Hi, > > A supported way to access S3 objects from a filesystem mount is with RGW > NFS. That is, RGW now exports the S3 namespace directly as files and > directories, one consumer is an nfs-ganesha NFS driver. > This supports a very specific subset of use cases/fs operations though, right? You can use it if you're just doing bulk file shuffling but it's not a way to upload via S3 and then perform filesystem update-in-place operations in any reasonable fashion (which is what I think was described in the original query). -Greg > > Regards, > > Matt > > ----- Original Message ----- > > From: "David Turner" <[email protected]> > > To: "Jonathan Lefman" <[email protected]>, > [email protected] > > Sent: Wednesday, June 28, 2017 2:59:12 PM > > Subject: Re: [ceph-users] Mapping data and metadata between rados and > cephfs > > > > CephFS is very different from RGW. You may be able to utilize s3fs-fuse > to > > interface with RGW, but I haven't heard of anyone using that on the ML > > before. > > > > On Wed, Jun 28, 2017 at 2:57 PM Lefman, Jonathan < > [email protected] > > > wrote: > > > > > > > > > > > > Thanks for the prompt reply. I was hoping that there would be an s3fs ( > > https://github.com/s3fs-fuse/s3fs-fuse ) equivalent for Ceph since > there are > > numerous functional similarities. Ideally one would be able to upload > data > > to a bucket and have the file synced to the local filesystem mount of > that > > bucket. This is similar to the idea of uploading data through RadosGW and > > have the data be available in CephFS. > > > > > > > > -Jon > > > > > > > > From: David Turner [mailto: [email protected] ] > > Sent: Wednesday, June 28, 2017 2:51 PM > > > > > > > > To: Lefman, Jonathan < [email protected] >; > [email protected] > > Subject: Re: [ceph-users] Mapping data and metadata between rados and > cephfs > > > > > > > > > > > > CephFS and RGW store data differently. I have never heard of, nor do I > > believe that it's possible, to have CephFS and RGW sharing the same data > > pool. > > > > > > > > > > > > On Wed, Jun 28, 2017 at 2:48 PM Lefman, Jonathan < > [email protected] > > > wrote: > > > > > > > > > > > > Yes, sorry. I meant the RadosGW. I still do not know what the mechanism > is to > > enable the mapping between data inserted by the rados component and the > > cephfs component. I hope that makes sense. > > > > > > > > -Jon > > > > > > > > From: David Turner [mailto: [email protected] ] > > Sent: Wednesday, June 28, 2017 2:46 PM > > To: Lefman, Jonathan < [email protected] >; > [email protected] > > Subject: Re: [ceph-users] Mapping data and metadata between rados and > cephfs > > > > > > > > > > > > You want to access the same data via a rados API and via cephfs? Are you > > thinking RadosGW? > > > > > > > > > > > > On Wed, Jun 28, 2017 at 1:54 PM Lefman, Jonathan < > [email protected] > > > wrote: > > > > > > > > > > > > Hi all, > > > > > > > > I would like to create a 1-to-1 mapping between rados and cephfs. Here's > the > > usage scenario: > > > > > > > > 1. Upload file via rest api through rados compatible APIs > > > > 2. Run "local" operations on the file delivered via rados on the linked > > cephfs mount > > > > 3. Retrieve/download file via rados API on newly created data available > on > > the cephfs mount > > > > > > > > I would like to know whether this is possible out-of-the-box; this will > never > > work; or this may work with a bit of effort. If this is possible, can > this > > be achieved in a scalable manner to accommodate multiple (10s to 100s) > users > > on the same system? > > > > > > > > I asked this question in #ceph and #ceph-devel. So far, there have not > been > > replies with a way to accomplish this. Thank you. > > > > > > > > -Jon > > > > > > > > > > _______________________________________________ > > ceph-users mailing list > > [email protected] > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > > ceph-users mailing list > > [email protected] > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > -- > Matt Benjamin > Red Hat, Inc. > 315 West Huron Street, Suite 140A > Ann Arbor, Michigan 48103 > > http://www.redhat.com/en/technologies/storage > > tel. 734-821-5101 <(734)%20821-5101> > fax. 734-769-8938 <(734)%20769-8938> > cel. 734-216-5309 <(734)%20216-5309> > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
