> 19 марта 2015 г., в 20:30, Sergey Shchukin <shchukin....@gmail.com> 
> написал(а):
> 
> 17.03.2015 13:22, Sergey Shchukin пишет:
>> 05.03.2015 11:25, Jim Nasby пишет:
>>> On 2/27/15 5:11 AM, Sergey Shchukin wrote: 
>>>> 
>>>> show max_standby_streaming_delay; 
>>>>   max_standby_streaming_delay 
>>>> ----------------------------- 
>>>>   30s 
>>> 
>>> We both need to be more clear about which server we're talking about 
>>> (master or replica). 
>>> 
>>> What are max_standby_streaming_delay and max_standby_archive_delay set to 
>>> *on the replica*? 
>>> 
>>> My hope is that one or both of those is set to somewhere around 8 minutes 
>>> on the replica. That would explain everything. 
>>> 
>>> If that's not the case then I suspect what's happening is there's something 
>>> running on the replica that isn't checking for interrupts frequently 
>>> enough. That would also explain it. 
>>> 
>>> When replication hangs, is the replication process using a lot of CPU? Or 
>>> is it just sitting there? What's the process status for the replay process 
>>> show? 
>>> 
>>> Can you get a trace of the replay process on the replica when this is 
>>> happening to see where it's spending all it's time? 
>>> 
>>> How are you generating these log lines? 
>>>  Tue Feb 24 15:05:07 MSK 2015 Stream: MASTER-masterdb:79607161592048 
>>> SLAVE:79607161550576 Replay:79607160986064 :: REPLAY 592 KBytes 
>>> (00:00:00.398376 seconds) 
>>> 
>>> Do you see the confl_* fields in pg_stat_database_conflicts on the 
>>> *replica* increasing? 
>> 
>> Hi Jim,
>> 
>> max_standby_streaming_delay and max_standby_archive_delay  both are 30s on 
>> master and replica dbs
>> 
>> I don't see any specific or heavy workload during this issue with a hanging 
>> apply process. Just a normal queries as usual. 
>> 
>> But I see an increased disk activity during the time when the apply issue is 
>> ongoing
>> 
>> DSK |          sdc  |              | busy     61%  | read   11511 |          
>>      | write   4534 | KiB/r     46  |              |  KiB/w      4 | MBr/s  
>> 52.78 |               | MBw/s   1.88 |  avq     1.45 |              |  avio 
>> 0.38 ms |
>> DSK |          sde  |              | busy     60%  | read   11457 |          
>>      | write   4398 | KiB/r     46  |              |  KiB/w      4 | MBr/s  
>> 51.97 |               | MBw/s   1.83 |  avq     1.47 |              |  avio 
>> 0.38 ms |
>> DSK |          sdd  |              | busy     60%  | read    9673 |          
>>      | write   4538 | KiB/r     61  |              |  KiB/w      4 | MBr/s  
>> 58.24 |               | MBw/s   1.88 |  avq     1.47 |              |  avio 
>> 0.42 ms |
>> DSK |          sdj  |              | busy     59%  | read    9576 |          
>>      | write   4177 | KiB/r     63  |              |  KiB/w      4 | MBr/s  
>> 59.30 |               | MBw/s   1.75 |  avq     1.48 |              |  avio 
>> 0.43 ms |
>> DSK |          sdh  |              | busy     59%  | read    9615 |          
>>      | write   4305 | KiB/r     63  |              |  KiB/w      4 | MBr/s  
>> 59.23 |               | MBw/s   1.80 |  avq     1.48 |              |  avio 
>> 0.42 ms |
>> DSK |          sdf  |              | busy     59%  | read    9483 |          
>>      | write   4404 | KiB/r     63  |              |  KiB/w      4 | MBr/s  
>> 59.11 |               | MBw/s   1.83 |  avq     1.47 |              |  avio 
>> 0.42 ms |
>> DSK |          sdi  |              | busy     59%  | read   11273 |          
>>      | write   4173 | KiB/r     46  |              |  KiB/w      4 | MBr/s  
>> 51.50 |               | MBw/s   1.75 |  avq     1.43 |              |  avio 
>> 0.38 ms |
>> DSK |          sdg  |              | busy     59%  | read   11406 |          
>>      | write   4297 | KiB/r     46  |              |  KiB/w      4 | MBr/s  
>> 51.66 |               | MBw/s   1.80 |  avq     1.46 |              |  avio 
>> 0.37 ms |
>> 
>> Although it's not seems to be an upper IO limit.
>> 
>> Normally disks are busy at 20-45%
>> 
>> DSK |          sde  |              | busy     29%  | read    6524 |          
>>      | write  14426 | KiB/r     26  |              |  KiB/w      5 | MBr/s  
>> 17.08 |               | MBw/s   7.78 |  avq    10.46 |              |  avio 
>> 0.14 ms |
>> DSK |          sdi  |              | busy     29%  | read    6590 |          
>>      | write  14391 | KiB/r     26  |              |  KiB/w      5 | MBr/s  
>> 17.19 |               | MBw/s   7.76 |  avq     8.75 |              |  avio 
>> 0.14 ms |
>> DSK |          sdg  |              | busy     29%  | read    6547 |          
>>      | write  14401 | KiB/r     26  |              |  KiB/w      5 | MBr/s  
>> 16.94 |               | MBw/s   7.60 |  avq     7.28 |              |  avio 
>> 0.14 ms |
>> DSK |          sdc  |              | busy     29%  | read    6835 |          
>>      | write  14283 | KiB/r     27  |              |  KiB/w      5 | MBr/s  
>> 18.08 |               | MBw/s   7.74 |  avq     8.77 |              |  avio 
>> 0.14 ms |
>> DSK |          sdf  |              | busy     23%  | read    3808 |          
>>      | write  14391 | KiB/r     36  |              |  KiB/w      5 | MBr/s  
>> 13.49 |               | MBw/s   7.78 |  avq    12.88 |              |  avio 
>> 0.13 ms |
>> DSK |          sdd  |              | busy     23%  | read    3747 |          
>>      | write  14229 | KiB/r     33  |              |  KiB/w      5 | MBr/s  
>> 12.32 |               | MBw/s   7.74 |  avq    10.07 |              |  avio 
>> 0.13 ms |
>> DSK |          sdj  |              | busy     23%  | read    3737 |          
>>      | write  14336 | KiB/r     36  |              |  KiB/w      5 | MBr/s  
>> 13.16 |               | MBw/s   7.76 |  avq    10.48 |              |  avio 
>> 0.13 ms |
>> DSK |          sdh  |              | busy     23%  | read    3793 |          
>>      | write  14362 | KiB/r     35  |              |  KiB/w      5 | MBr/s  
>> 13.29 |               | MBw/s   7.60 |  avq     8.61 |              |  avio 
>> 0.13 ms |
>> 
>> 
>> Also during the issue perf shows [k] copy_user_generic_string on the top 
>> positions
>>  14.09%     postmaster  postgres               [.] 0x00000000001b4569
>>  10.25%     postmaster  [kernel.kallsyms]      [k] copy_user_generic_string
>>   4.15%     postmaster  postgres               [.] 
>> hash_search_with_hash_value
>>   2.08%     postmaster  postgres               [.] SearchCatCache
>>   1.79%     postmaster  postgres               [.] LWLockAcquire
>>   1.18%     postmaster  libc-2.12.so           [.] memcpy
>>   1.12%     postmaster  postgres               [.] mdnblocks
>> 
>> Issue starts: at 19:43
>> Mon Mar 16 19:43:04 MSK 2015 Stream: MASTER-rdb04d:70837172337784 
>> SLAVE:70837172314864 Replay:70837172316512 :: REPLAY 21 KBytes 
>> (00:00:00.006225 seconds)
>> Mon Mar 16 19:43:09 MSK 2015 Stream: MASTER-rdb04d:70837177455624 
>> SLAVE:70837177390968 Replay:70837176794376 :: REPLAY 646 KBytes 
>> (00:00:00.367305 seconds)
>> Mon Mar 16 19:43:14 MSK 2015 Stream: MASTER-rdb04d:70837185005120 
>> SLAVE:70837184961280 Replay:70837183253896 :: REPLAY 1710 KBytes 
>> (00:00:00.827881 seconds)
>> Mon Mar 16 19:43:19 MSK 2015 Stream: MASTER-rdb04d:70837190417984 
>> SLAVE:70837190230232 Replay:70837183253896 :: REPLAY 6996 KBytes 
>> (00:00:05.873169 seconds)
>> Mon Mar 16 19:43:24 MSK 2015 Stream: MASTER-rdb04d:70837198538232 
>> SLAVE:70837198485000 Replay:70837183253896 :: REPLAY 14926 KBytes 
>> (00:00:11.025561 seconds)
>> Mon Mar 16 19:43:29 MSK 2015 Stream: MASTER-rdb04d:70837209961192 
>> SLAVE:70837209869384 Replay:70837183253896 :: REPLAY 26081 KBytes 
>> (00:00:16.068014 seconds)
>> 
>> We see [k] copy_user_generic_string
>> 
>>  12.90%     postmaster  [kernel.kallsyms]      [k] copy_user_generic_string
>>  11.49%     postmaster  postgres               [.] 0x00000000001f40c1
>>   4.74%     postmaster  postgres               [.] 
>> hash_search_with_hash_value
>>   1.86%     postmaster  postgres               [.] mdnblocks
>>   1.73%     postmaster  postgres               [.] LWLockAcquire
>>   1.67%     postmaster  postgres               [.] SearchCatCache
>> 
>> 
>>  25.71%     postmaster  [kernel.kallsyms]      [k] copy_user_generic_string
>>   7.89%     postmaster  postgres               [.] 
>> hash_search_with_hash_value
>>   4.66%     postmaster  postgres               [.] 0x00000000002108da
>>   4.51%     postmaster  postgres               [.] mdnblocks
>>   3.36%     postmaster  [kernel.kallsyms]      [k] put_page
>> 
>> Issue stops: at 19:51:39
>> Mon Mar 16 19:51:24 MSK 2015 Stream: MASTER-rdb04d:70838904179344 
>> SLAVE:70838903934392 Replay:70837183253896 :: REPLAY 1680591 KBytes 
>> (00:08:10.384679 seconds)
>> Mon Mar 16 19:51:29 MSK 2015 Stream: MASTER-rdb04d:70838929994336 
>> SLAVE:70838929873624 Replay:70837183253896 :: REPLAY 1705801 KBytes 
>> (00:08:15.428773 seconds)
>> Mon Mar 16 19:51:34 MSK 2015 Stream: MASTER-rdb04d:70838951993624 
>> SLAVE:70838951899768 Replay:70837183253896 :: REPLAY 1727285 KBytes 
>> (00:08:20.472567 seconds)
>> Mon Mar 16 19:51:39 MSK 2015 Stream: MASTER-rdb04d:70838975297912 
>> SLAVE:70838975180384 Replay:70837208050872 :: REPLAY 1725827 KBytes 
>> (00:08:10.256935 seconds)
>> Mon Mar 16 19:51:44 MSK 2015 Stream: MASTER-rdb04d:70839001502160 
>> SLAVE:70839001412616 Replay:70837260116984 :: REPLAY 1700572 KBytes 
>> (00:07:49.849511 seconds)
>> Mon Mar 16 19:51:49 MSK 2015 Stream: MASTER-rdb04d:70839022866760 
>> SLAVE:70839022751184 Replay:70837276732880 :: REPLAY 1705209 KBytes 
>> (00:07:42.307364 seconds)
>> 
>> And copy_user_generic_string goes down
>> +  13.43%       postmaster  postgres                 [.] 0x000000000023dc9a
>> +   3.71%       postmaster  [kernel.kallsyms]        [k] 
>> copy_user_generic_string
>> +   2.46%             init  [kernel.kallsyms]        [k] intel_idle
>> +   2.30%       postmaster  postgres                 [.] 
>> hash_search_with_hash_value
>> +   2.01%       postmaster  postgres                 [.] SearchCatCache
>> 
>> 
>> Could you clarify what types of traces did you mean? GDB?
>> 
>> To calculate slave and apply lag I use the following query at the replica 
>> site
>> 
>> slave_lag=$($psql -U monitor  -h$s_host -p$s_port -A -t -c "SELECT 
>> pg_xlog_location_diff(pg_last_xlog_receive_location(), '0/0') AS receive" 
>> $p_db)
>> replay_lag=$($psql -U monitor -h$s_host -p$s_port -A -t -c "SELECT 
>> pg_xlog_location_diff(pg_last_xlog_replay_location(), '0/0') AS replay" 
>> $p_db)
>> replay_timediff=$($psql -U monitor -h$s_host -p$s_port -A -t -c "SELECT 
>> NOW() - pg_last_xact_replay_timestamp() AS replication_delay" $p_db)
>> master_lag=$($psql -U monitor -h$p_host -p$p_port -A -t -c "SELECT 
>> pg_xlog_location_diff(pg_current_xlog_location(), '0/0') AS offset" $p_db)
>> echo "$(date) Stream: MASTER-$p_host:$master_lag SLAVE:$slave_lag 
>> Replay:$replay_lag :: REPLAY $(bc <<< $master_lag/1024-$replay_lag/1024) 
>> KBytes (${replay_timediff} seconds)" 
>> 
>> - 
>> Best regards,
>> Sergey Shchukin 
>> 
> 
> One more thing
> 
> We have upgraded one of our shards to 9.4.1 and expectedly that did not help.
> 
> A few things to notice which may be useful.
> 
> 1. When replay stops, startup process reads a lot from array with $PGDATA. In 
> iotop and iostat we see the following:
> 
> Total DISK READ: 490.42 M/s | Total DISK WRITE: 3.82 M/s
>   TID  PRIO  USER    DISK READ>  DISK WRITE  SWAPIN      IO    COMMAND
>    3316 be/4 postgres  492.34 M/s    0.00 B/s  0.00 % 39.91 % postgres: 
> startup process
>    <...>
> 
>    Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s avgrq-sz 
> avgqu-sz   await  svctm  %util
>    <...>
>    md2               0.00     0.00 6501.00    7.00   339.90     0.03   106.97 
>     0.00    0.00   0.00   0.00
>    md3               0.00     0.00    0.00 1739.00     0.00     6.79     8.00 
>     0.00    0.00   0.00   0.00
> 
>    root@rpopdb04g ~ # fgrep 9.4 /proc/mounts
>    /dev/md2 /var/lib/pgsql/9.4/data ext4 
> rw,noatime,nodiratime,barrier=1,stripe=64,data=ordered 0 0
>    /dev/md3 /var/lib/pgsql/9.4/data/pg_xlog ext4 
> rw,noatime,nodiratime,barrier=0,stripe=64,data=ordered 0 0
>    root@rpopdb04g ~ #
> 
> 2. The state of the startup process is changing in such a way:
> 
>    root@rpopdb04g ~ # while true; do ps aux | grep '[s]tartup'; sleep 1; done
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Rs   18:04   8:11 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Ts   18:04   8:11 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Rs   18:04   8:12 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Ds   18:04   8:12 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Rs   18:04   8:13 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Rs   18:04   8:13 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Rs   18:04   8:14 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Ts   18:04   8:14 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Ds   18:04   8:15 postgres: 
> startup process
>    postgres  3316 26.6  3.2 4732052 4299260 ?     Rs   18:04   8:15 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Ds   18:04   8:15 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Ds   18:04   8:16 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Rs   18:04   8:16 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Rs   18:04   8:17 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Ts   18:04   8:17 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Rs   18:04   8:18 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Rs   18:04   8:18 postgres: 
> startup process
>    postgres  3316 26.7  3.2 4732052 4299260 ?     Ds   18:04   8:19 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Rs   18:04   8:19 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Rs   18:04   8:20 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Rs   18:04   8:20 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Ds   18:04   8:21 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Ds   18:04   8:22 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Rs   18:04   8:22 postgres: 
> startup process
>    postgres  3316 26.8  3.2 4732052 4299260 ?     Rs   18:04   8:23 postgres: 
> startup process
>    postgres  3316 26.9  3.2 4732052 4299260 ?     Rs   18:04   8:23 postgres: 
> startup process
>    postgres  3316 26.9  3.2 4732052 4299260 ?     Rs   18:04   8:24 postgres: 
> startup process
>    postgres  3316 26.9  3.2 4732052 4299260 ?     Ts   18:04   8:24 postgres: 
> startup process
>    postgres  3316 26.9  3.2 4732052 4299260 ?     Rs   18:04   8:25 postgres: 
> startup process
>    ^C
>    root@rpopdb04g ~ #
> 
> 3. confl* fields in pg_stat_database_conflicts are always zero during the 
> pausing of replay.
> 
> 4. The stack-traces taken with GDB are not really informative. We will 
> recompile PostgreSQL with —enable-debug option and run it on one of our 
> replicas if needed. Since it is a production system we would like to do it 
> last of all. But we will do it if anybody would not give us any ideas.

We did it. Most of the backtraces (taken while replay_location was not 
changing) looks like that:

[Thread debugging using libthread_db enabled]
0x00007f54a71444c0 in __read_nocancel () from /lib64/libc.so.6
#0  0x00007f54a71444c0 in __read_nocancel () from /lib64/libc.so.6
#1  0x000000000065d2f5 in FileRead (file=<value optimized out>, 
buffer=0x7f5470e6ba20 "\004", amount=8192) at fd.c:1286
#2  0x000000000067acad in mdread (reln=<value optimized out>, forknum=<value 
optimized out>, blocknum=311995, buffer=0x7f5470e6ba20 "\004") at md.c:679
#3  0x0000000000659b4e in ReadBuffer_common (smgr=<value optimized out>, 
relpersistence=112 'p', forkNum=MAIN_FORKNUM, blockNum=311995, 
mode=RBM_NORMAL_NO_LOG, strategy=0x0, hit=0x7fff898a912f "") at bufmgr.c:476
#4  0x000000000065a61b in ReadBufferWithoutRelcache (rnode=..., 
forkNum=MAIN_FORKNUM, blockNum=311995, mode=<value optimized out>, 
strategy=<value optimized out>) at bufmgr.c:287
#5  0x00000000004cfb78 in XLogReadBufferExtended (rnode=..., 
forknum=MAIN_FORKNUM, blkno=311995, mode=RBM_NORMAL_NO_LOG) at xlogutils.c:324
#6  0x00000000004a3651 in btree_xlog_vacuum (lsn=71742288638464, 
record=0x1e48b78) at nbtxlog.c:522
#7  btree_redo (lsn=71742288638464, record=0x1e48b78) at nbtxlog.c:1144
#8  0x00000000004c903a in StartupXLOG () at xlog.c:6827
#9  0x000000000062f8bf in StartupProcessMain () at startup.c:224
#10 0x00000000004d3e9a in AuxiliaryProcessMain (argc=2, argv=0x7fff898a98a0) at 
bootstrap.c:416
#11 0x000000000062a99c in StartChildProcess (type=StartupProcess) at 
postmaster.c:5146
#12 0x000000000062e9e2 in PostmasterMain (argc=3, argv=<value optimized out>) 
at postmaster.c:1237
#13 0x00000000005c7d68 in main (argc=3, argv=0x1e22910) at main.c:228

So the problem seems to be in this part of code - 
http://git.postgresql.org/gitweb/?p=postgresql.git;a=blob;f=src/backend/access/nbtree/nbtxlog.c;h=5f9fc49e78ca1388ab482e24c8b5a873238ae0b6;hb=d0f83327d3739a45102fdd486947248c70e0249d#l507.
 I suppose, that answers the question why startup process reads a lot from disk 
while paused replay.

So the questions are:
1. Is there anything we can tune right now? Except for not reading from 
replicas and partitioning this table.
2. Isn’t there still a function to determine that a buffer is not pinned in 
shared_buffers without reading it from disk? To optimize current behaviour in 
the future.

> 
> 5. In one of the experiments replay stopped on 4115/56126DC0 xlog position. 
> Here is a bit of pg_xlogdump output:
> 
>     rpopdb04d/rpopdb M # select pg_xlogfile_name('4115/56126DC0');
>      pg_xlogfile_name
>      --------------------------
>       000000060000411500000056
>       (1 row)
> 
>       Time: 0.496 ms
>       rpopdb04d/rpopdb M #
> 
>     root@pg-backup04h /u0/rpopdb04/wals/0000000600004115 # 
> /usr/pgsql-9.4/bin/pg_xlogdump 000000060000411500000056 
> 000000060000411500000056 | fgrep 4115/56126DC0 -C10
>     rmgr: Heap        len (rec/tot):     36/  3948, tx: 3930143874, lsn: 
> 4115/561226C8, prev 4115/56122698, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 196267/4 xmax 3930143874 ; new tid 196267/10 xmax 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143874, lsn: 
> 4115/56123638, prev 4115/561226C8, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.725158 MSK
>     rmgr: Heap        len (rec/tot):     58/    90, tx: 3930143875, lsn: 
> 4115/56123668, prev 4115/56123638, bkp: 0000, desc: hot_update: rel 
> 1663/16420/16737; tid 23624/3 xmax 3930143875 ; new tid 23624/21 xmax 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143875, lsn: 
> 4115/561236C8, prev 4115/56123668, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.726432 MSK
>     rmgr: Heap        len (rec/tot):     36/  2196, tx: 3930143876, lsn: 
> 4115/561236F8, prev 4115/561236C8, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 123008/4 xmax 3930143876 ; new tid 123008/5 xmax 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143876, lsn: 
> 4115/56123F90, prev 4115/561236F8, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.727088 MSK
>     rmgr: Heap        len (rec/tot):     36/  7108, tx: 3930143877, lsn: 
> 4115/56123FC0, prev 4115/56123F90, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 34815/6 xmax 3930143877 ; new tid 34815/16 xmax 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143877, lsn: 
> 4115/56125BA0, prev 4115/56123FC0, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.728178 MSK
>     rmgr: Heap        len (rec/tot):     36/  4520, tx: 3930143878, lsn: 
> 4115/56125BD0, prev 4115/56125BA0, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 147863/5 xmax 3930143878 ; new tid 147863/16 xmax 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143878, lsn: 
> 4115/56126D90, prev 4115/56125BD0, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.728339 MSK
>     rmgr: Btree       len (rec/tot):     20/    52, tx:          0, lsn: 
> 4115/56126DC0, prev 4115/56126D90, bkp: 0000, desc: vacuum: rel 
> 1663/16420/16796; blk 31222118, lastBlockVacuumed 0
>     rmgr: Heap        len (rec/tot):     36/  6112, tx: 3930143879, lsn: 
> 4115/56126DF8, prev 4115/56126DC0, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 23461/26 xmax 3930143879 ; new tid 23461/22 xmax 0
>     rmgr: Heap2       len (rec/tot):     24/  8160, tx:          0, lsn: 
> 4115/561285F0, prev 4115/56126DF8, bkp: 1000, desc: clean: rel 
> 1663/16420/16782; blk 21997709 remxid 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143879, lsn: 
> 4115/5612A5E8, prev 4115/561285F0, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.728805 MSK
>     rmgr: Heap2       len (rec/tot):     20/  8268, tx:          0, lsn: 
> 4115/5612A618, prev 4115/5612A5E8, bkp: 1000, desc: visible: rel 
> 1663/16420/16782; blk 21997709
>     rmgr: Heap        len (rec/tot):     36/  7420, tx: 3930143880, lsn: 
> 4115/5612C680, prev 4115/5612A618, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 37456/8 xmax 3930143880 ; new tid 37456/29 xmax 0
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143880, lsn: 
> 4115/5612E398, prev 4115/5612C680, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.729141 MSK
>     rmgr: Heap        len (rec/tot):     36/  7272, tx: 3930143881, lsn: 
> 4115/5612E3C8, prev 4115/5612E398, bkp: 1000, desc: hot_update: rel 
> 1663/16420/16737; tid 23614/2 xmax 3930143881 ; new tid 23614/22 xmax 0
>     rmgr: Heap        len (rec/tot):    150/   182, tx:          0, lsn: 
> 4115/56130048, prev 4115/5612E3C8, bkp: 0000, desc: inplace: rel 
> 1663/16420/12764; tid 11/31
>     rmgr: Transaction len (rec/tot):     12/    44, tx: 3930143881, lsn: 
> 4115/56130100, prev 4115/56130048, bkp: 0000, desc: commit: 2015-03-19 
> 18:26:27.729340 MSK
>     rmgr: Heap        len (rec/tot):     43/    75, tx: 3930143882, lsn: 
> 4115/56130130, prev 4115/56130100, bkp: 0000, desc: insert: rel 
> 1663/16420/16773; tid 10159950/26
>     rmgr: Btree       len (rec/tot):     42/    74, tx: 3930143882, lsn: 
> 4115/56130180, prev 4115/56130130, bkp: 0000, desc: insert: rel 
> 1663/16420/16800; tid 12758988/260
>     root@pg-backup04h /u0/rpopdb04/wals/0000000600004115 #
> 
> Any help would be really appropriate. Thanks in advance.
> 


--
May the force be with you…
https://simply.name

Reply via email to