For what it's worth here, after I added a hot tier to the pool, the brick
sizes are now reporting the correct size of all bricks combined instead of
just one brick.

Not sure if that gives you any clues for this... maybe adding another brick
to the pool would have a similar effect?


On Thu, Dec 21, 2017 at 11:44 AM, Tom Fite <[email protected]> wrote:

> Sure!
>
> > 1 - output of gluster volume heal <volname> info
>
> Brick pod-sjc1-gluster1:/data/brick1/gv0
> Status: Connected
> Number of entries: 0
>
> Brick pod-sjc1-gluster2:/data/brick1/gv0
> Status: Connected
> Number of entries: 0
>
> Brick pod-sjc1-gluster1:/data/brick2/gv0
> Status: Connected
> Number of entries: 0
>
> Brick pod-sjc1-gluster2:/data/brick2/gv0
> Status: Connected
> Number of entries: 0
>
> Brick pod-sjc1-gluster1:/data/brick3/gv0
> Status: Connected
> Number of entries: 0
>
> Brick pod-sjc1-gluster2:/data/brick3/gv0
> Status: Connected
> Number of entries: 0
>
> > 2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log
>
> Attached
>
> > 3 - output of gluster volume <volname> info
>
> [root@pod-sjc1-gluster2 ~]# gluster volume info
>
> Volume Name: gv0
> Type: Distributed-Replicate
> Volume ID: d490a9ec-f9c8-4f10-a7f3-e1b6d3ced196
> Status: Started
> Snapshot Count: 13
> Number of Bricks: 3 x 2 = 6
> Transport-type: tcp
> Bricks:
> Brick1: pod-sjc1-gluster1:/data/brick1/gv0
> Brick2: pod-sjc1-gluster2:/data/brick1/gv0
> Brick3: pod-sjc1-gluster1:/data/brick2/gv0
> Brick4: pod-sjc1-gluster2:/data/brick2/gv0
> Brick5: pod-sjc1-gluster1:/data/brick3/gv0
> Brick6: pod-sjc1-gluster2:/data/brick3/gv0
> Options Reconfigured:
> performance.cache-refresh-timeout: 60
> performance.stat-prefetch: on
> server.allow-insecure: on
> performance.flush-behind: on
> performance.rda-cache-limit: 32MB
> network.tcp-window-size: 1048576
> performance.nfs.io-threads: on
> performance.write-behind-window-size: 4MB
> performance.nfs.write-behind-window-size: 512MB
> performance.io-cache: on
> performance.quick-read: on
> features.cache-invalidation: on
> features.cache-invalidation-timeout: 600
> performance.cache-invalidation: on
> performance.md-cache-timeout: 600
> network.inode-lru-limit: 90000
> performance.cache-size: 4GB
> server.event-threads: 16
> client.event-threads: 16
> features.barrier: disable
> transport.address-family: inet
> nfs.disable: on
> performance.client-io-threads: on
> cluster.lookup-optimize: on
> server.outstanding-rpc-limit: 1024
> auto-delete: enable
> You have new mail in /var/spool/mail/root
>
> > 4 - output of gluster volume <volname> status
>
> [root@pod-sjc1-gluster2 ~]# gluster volume status gv0
> Status of volume: gv0
> Gluster process                             TCP Port  RDMA Port  Online
> Pid
> ------------------------------------------------------------
> ------------------
> Brick pod-sjc1-gluster1:/data/
> brick1/gv0                                  49152     0          Y
>  3198
> Brick pod-sjc1-gluster2:/data/
> brick1/gv0                                  49152     0          Y
>  4018
> Brick pod-sjc1-gluster1:/data/
> brick2/gv0                                  49153     0          Y
>  3205
> Brick pod-sjc1-gluster2:/data/
> brick2/gv0                                  49153     0          Y
>  4029
> Brick pod-sjc1-gluster1:/data/
> brick3/gv0                                  49154     0          Y
>  3213
> Brick pod-sjc1-gluster2:/data/
> brick3/gv0                                  49154     0          Y
>  4036
> Self-heal Daemon on localhost               N/A       N/A        Y
>  17869
> Self-heal Daemon on pod-sjc1-gluster1.exava
> ult.com                                     N/A       N/A        Y
>  3183
>
> Task Status of Volume gv0
> ------------------------------------------------------------
> ------------------
> There are no active volume tasks
>
>
> > 5 - Also, could you try unmount the volume and mount it again and check
> the size?
>
> I have done this a few times but it doesn't seem to help.
>
> On Thu, Dec 21, 2017 at 11:18 AM, Ashish Pandey <[email protected]>
> wrote:
>
>>
>> Could youplease provide following -
>>
>> 1 - output of gluster volume heal <volname> info
>> 2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log
>> 3 - output of gluster volume <volname> info
>> 4 - output of gluster volume <volname> status
>> 5 - Also, could you try unmount the volume and mount it again and check
>> the size?
>>
>>
>>
>>
>>
>>
>>
>> ------------------------------
>> *From: *"Teknologeek Teknologeek" <[email protected]>
>> *To: *[email protected]
>> *Sent: *Wednesday, December 20, 2017 2:54:40 AM
>> *Subject: *[Gluster-users] Wrong volume size with df
>>
>>
>> I have a glusterfs setup with distributed disperse volumes 5 * ( 4 + 2 ).
>>
>> After a server crash, "gluster peer status" reports all peers as
>> connected.
>>
>> "gluster volume status detail" shows that all bricks are up and running
>> with the right size, but when I use df from a client mount point, the size
>> displayed is about 1/6 of the total size.
>>
>> When browsing the data, they seem to be ok tho.
>>
>> I need some help to understand what's going on as i can't delete the
>> volume and recreate it from scratch.
>>
>> _______________________________________________
>> Gluster-users mailing list
>> [email protected]
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>>
>> _______________________________________________
>> Gluster-users mailing list
>> [email protected]
>> http://lists.gluster.org/mailman/listinfo/gluster-users
>>
>
>
_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to