Lacey Powers ha scritto:
Silvio Brandani wrote:
We have a postgres 8.3.8 on linux
We get following messages int /var/log/messages:
May 6 22:31:01 pgblade02 kernel: postgres invoked oom-killer:
gfp_mask=0x201d2, order=0, oomkilladj=0
May 6 22:31:01 pgblade02 kernel:
May 6 22:31:01
Greg Spiegelberg ha scritto:
On Fri, May 7, 2010 at 8:26 AM, Silvio Brandani
silvio.brand...@tech.sdb.it mailto:silvio.brand...@tech.sdb.it wrote:
We have a postgres 8.3.8 on linux
We get following messages int /var/log/messages:
May 6 22:31:01 pgblade02 kernel: postgres invoked
Silvio Brandani ha scritto:
Lacey Powers ha scritto:
Silvio Brandani wrote:
We have a postgres 8.3.8 on linux
We get following messages int /var/log/messages:
May 6 22:31:01 pgblade02 kernel: postgres invoked oom-killer:
gfp_mask=0x201d2, order=0, oomkilladj=0
May 6 22:31:01 pgblade02
Silvio Brandani ha scritto:
Greg Spiegelberg ha scritto:
On Fri, May 7, 2010 at 8:26 AM, Silvio Brandani
silvio.brand...@tech.sdb.it mailto:silvio.brand...@tech.sdb.it
wrote:
We have a postgres 8.3.8 on linux
We get following messages int /var/log/messages:
May 6 22:31:01
Giles Lean ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
yes the server is dedicated to PostgreSQL.
Could be a bug of PostgreSQL the fact that the system went Out of
Memory?? Wath can be the cause of it?
Regards,
Silvio
The out of memory killer is a totally bogus
.log:2010-05-12 06:30:57 CEST [26617]:
[289-1] LOG: unexpected EOF on client connection
How is it possible to eliminate this disconnection problems? are they
related to postgres configuration or something related to linux or other
stuff ?
Any suggestion higly appreciated
Silvio Brandani
Kevin Grittner ha scritto:
Silvio Brandani wrote:
postgresql-2010-05-12_053937.log:2010-05-12 06:30:57 CEST [26617]:
[288-1] LOG: could not receive data from client: Connection reset
by peer
postgresql-2010-05-12_053937.log:2010-05-12 06:30:57 CEST [26617]:
[289-1] LOG: unexpected EOF
, attachment_id) TO stdout;
I have already try to reindex and vacuum full the table. but problem
still stand.
How can we fix this error and get a good dump??--
Any help higly appreciated.
--Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
Kevin Grittner ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
We have a standby database
During pg_dump
Hmm... I just noticed that word standby in there. Can you
elaborate on what you mean by that?
-Kevin
It means it is an istance refreshed (via
Kevin Grittner ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
We have a standby database
During pg_dump
Hmm... I just noticed that word standby in there. Can you
elaborate on what you mean by that?
-Kevin
It means it is an istance refreshed (via
We need to upgrade the postgres running on our production system under
Red Hat Enterprise Linux Server release 5.1 from version 8.1.21 to
version 8.3.6.
we could have a stop/maintenance window of 3/4 our the sum of size of
databases is around 1G .
Which is the best practice to execute such
Gabriele Bartolini ha scritto:
Ciao Silvio,
On Tue, 13 Jul 2010 22:52:51 +0200, Iñigo Martinez Lasala
You should script database migration process in order to make it faster.
Upgrading binaries is really simple. A yum upgrade should be enough.
Yes, faster and repeatable. Something
Hi all,
I need to start a postgres istance on a samba filesystems /data which
have rwxrwxrwx (777) permissions, Ic ould not change such permissions.
If I try to start postgres it complain that should be 700 permission .
How can I skip this control by postgres startup process and be able to
David Bear ha scritto:
this doesn't make sense. why go through the overhead of smb to get to
a database cluster over a network connection when pg can already work
over tcpip sockets and work much more switfly on a local file system?
2010/7/15 Silvio Brandani silvio.brand...@tech.sdb.it
Hi alls,
we have a performance issue on deleting data from a table.
the delete is on the index key so should be very fast but the table
have 13 contraints of type foreign keys and the slowness seems due to
all the works done with the constraint.
Trying to drop the constraint and execute the
Kevin Grittner ha scritto:
Silvio Brandani wrote:
we have a performance issue on deleting data from a table.
the delete is on the index key so should be very fast but the table
have 13 contraints of type foreign keys and the slowness seems due
to all the works done with the constraint
Hi,
a query on our production database give following errror:
2010-08-05 10:52:40 CEST [12106]: [278-1] ERROR: out of memory
2010-08-05 10:52:40 CEST [12106]: [279-1] DETAIL: Failed on request of
size 48.
any suggestion ?
--
Silvio Brandani
Infrastructure Administrator
SDB Information
max_stack_depth = 2MB
[]´s
Victor Hugo P.Clemente
Brazil
2010/8/5 Silvio Brandani silvio.brand...@tech.sdb.it:
Hi,
a query on our production database give following errror:
2010-08-05 10:52:40 CEST [12106]: [278-1] ERROR: out of memory
2010-08-05 10:52:40 CEST [12106]: [279-1] DETAIL: Failed
Tom Lane ha scritto:
Kevin Grittner kevin.gritt...@wicourts.gov writes:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
a query on our production database give following errror:
2010-08-05 10:52:40 CEST [12106]: [278-1] ERROR: out of memory
2010-08-05 10:52:40 CEST [12106]: [279-1
Bob Lunney ha scritto:
Silvio ,
I had a similar problem when starting the database from an account that didn't
have the appropriate ulimits set. Check the ulimit values using ulimit -a.
HTH,
Bob Lunney
--- On Thu, 8/5/10, Silvio Brandani silvio.brand...@tech.sdb.it wrote:
From
Silvio Brandani ha scritto:
Bob Lunney ha scritto:
Silvio ,
I had a similar problem when starting the database from an account
that didn't have the appropriate ulimits set. Check the ulimit
values using ulimit -a.
HTH,
Bob Lunney
--- On Thu, 8/5/10, Silvio Brandani silvio.brand
Hi all,
we are looking for a solution to create a replicated database to be used
as reporting database with same data of production ( we can accept a lag
of some minutes).
we already have the Point In Time Recovery but we need a solution where
the standby is always open in readonly , but
Is it possible to have multiple host destinations in the archive_command
parameter of postgresql.conf to be able to mantain 2 standby datbase
with write ahead log???
Can we use something like:
archive_command = '/usr/bin/scp -Cv %p
host_standby1:/var/lib/pgsql/archivelog/%f /usr/bin/scp -Cv
in 3 blocks; 24480 free (18 chunks); 96 used
Any suggestion higly appreciated
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo a terzi.
Informazioni dettagliate, anche in ordine al
Silvio Brandani ha scritto:
Still problems of Out of Memory:
the query is the following and if I run it from psql is working fine,
but from application I get error :
SELECT MAX(oec.ctnr_nr) ::char(13) as Ctnr_nr,MAX(oec.file_ref)
::char(7) as File_Ref,MAX(oec.move_type) ::char(5
Tom Lane ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it writes:
Still problems of Out of Memory:
the query is the following and if I run it from psql is working fine,
but from application I get error :
Is it really the *exact* same query both ways, or are you doing
Tom Lane ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it writes:
Tom Lane ha scritto:
Is it really the *exact* same query both ways, or are you doing
something like parameterizing the query in the application?
Is it exactly the same, the query text is from
Tom Lane ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it writes:
Tom Lane ha scritto:
Is it really the *exact* same query both ways, or are you doing
something like parameterizing the query in the application?
Is it exactly the same, the query text is from
Simon Riggs ha scritto:
On Tue, 2010-08-24 at 15:46 +0200, Silvio Brandani wrote:
Hi all,
we are looking for a solution to create a replicated database to be used
as reporting database with same data of production ( we can accept a lag
of some minutes).
we already have the Point In Time
on our production
system.
Is it possible to softly kill a vacuum process without risk a panic ?
thanks a lot
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo a terzi
Kevin Grittner ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
we have develop a script to execute the vacuum full on all tables
of our very big database , since it is a 24 x 7 available system
we have not a timeframe to exec the vacuum full.
so we try with this script
Kevin Grittner ha scritto:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
we have develop a script to execute the vacuum full on all tables
of our very big database , since it is a 24 x 7 available system
we have not a timeframe to exec the vacuum full.
so we try with this script
Tom Lane ha scritto:
Kevin Grittner kevin.gritt...@wicourts.gov writes:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
we have develop a script to execute the vacuum full on all tables
of our very big database , since it is a 24 x 7 available system
we have not a timeframe to exec
Kevin Grittner ha scritto:
Scott Marlowe scott.marl...@gmail.com wrote:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
we have develop a script to execute the vacuum full on all tables
Vacuum full is more of a recovery / offline command and is to be
used sparingly
Silvio Brandani ha scritto:
Kevin Grittner ha scritto:
Scott Marlowe scott.marl...@gmail.com wrote:
Silvio Brandani silvio.brand...@tech.sdb.it wrote:
we have develop a script to execute the vacuum full on all tables
Vacuum full is more of a recovery / offline command
a reindex in a share lock mode ?? we have a 8.3
version. In higher postgres version the reindex mode change?
thanks a lot
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo
Kevin Grittner ha scritto:
Silvio Brandani wrote:
Is it possible to use a reindex in a share lock mode ??
Check out the CONCURRENTLY keyword in CREATE INDEX:
http://www.postgresql.org/docs/current/interactive/sql-createindex.html
You could create a duplicate index concurrently
During a server restart the mount point /var/lib/pgsql/data lost the
0700 permission and the automatic restart of postgres failed.
How can I set the permission of this mount point to be not lost during
restart of server ?? Have I to set specific option in fstab??
thanks
---
: out of memory
DETAIL: Failed on request of size 16.
ERROR: current transaction is aborted, commands ignored until end of
transaction block
STATEMENT: close SQL_CUR02B79258
LOG: statement: ROLLBACK
--
Silvio Brandani
Infrastructure Administrator
SDB Information Technology
Phone: +39.055.3811222
error also with 9.0.1 server version.
Regarding the ODBC driver we test all last versions (9.0.2, 8.4.0.2) .
The encoding used is UTF8 users work on a Turkish language.
--
Silvio Brandani
Infrastructure Administrator
SDB Information Technology
Phone: +39.055.3811222
Fax: +39.055.5201119
server version.
any comment higly appreciated.
--
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo a terzi.
Informazioni dettagliate, anche in ordine al Suo diritto di accesso e agli
Our database seems not using index anymore, please help with, is a
production database.
is there a script to check missing index on foreign key ??
thanks a lot
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando
Ok,
the problem was on a big table on query like this:
select outmessage0_.out_msg_id as out1_0_
from edi.out_messages outmessage0_, edi.transaction_set_partners
transactio1_
where outmessage0_.transaction_set_partner=transactio1_.trn_set_prtn_id
and outmessage0_.status_id='TOSND'
and
)
Filter: (aida_reference_id = 3145)
Total runtime: 1457.696 ms
already executed the vacuum ,reindex.
Please help
--
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo a terzi
table is 959818 records,
I create a copy of the table with create tabase as select ... and
then indexed, the new table use the indexes ...
thanks
Il 18/11/2011 15.19, Szymon Guz ha scritto:
2011/11/18 Silvio Brandani silvio.brand...@tech.sdb.it
mailto:silvio.brand...@tech.sdb.it
it.
I have vacuum full the table and still not work, how can I rebuild this
table ? this is a production database..
thanks
--
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo
I have a system with archive log shipping to a standby server read only
with replication streaming.
is possible to use the truncate in the primary or it will affect the log
shipping architecture and will corrupt the standby server??
thanks
--
Silvio Brandani
---
Utilizziamo i dati
can I reset the
pg_statistics for this table???
Any comment higly appreciated.
--
Silvio Brandani
---
Utilizziamo i dati personali che la riguardano esclusivamente per nostre
finalità amministrative e contabili, anche quando li comunichiamo a terzi.
Informazioni dettagliate, anche
rows=117295 width=11)
Filter: (abs((amount - total_amo)) 0.1::double precision)
Il 09/01/2012 15.59, Szymon Guz ha scritto:
On 9 January 2012 15:41, Silvio Brandani silvio.brand...@tech.sdb.it
mailto:silvio.brand...@tech.sdb.it wrote:
In the last few hours we get
49 matches
Mail list logo