I have twice set up pg hot standbys ala the docs at
http://www.postgresql.org/docs/9.1/interactive/hot-standby.html
The third time I'm trying this I'm running into trouble. The first two
times were with actual servers. This time I'm trying to set up two pg
instances on my desktop for
I have a vacuum process that is sitting around and apparently not doing
anything. It's been around over 2000 seconds and is eating up no cpu.
It isn't waiting on a lock. Backtrace is this:
#0 0x00367aed4ff7 in semop () from /lib64/libc.so.6
#1 0x005d2a83 in PGSemaphoreLock
All was fine until:
LOG: statement: select _devel.cleanupEvent('10 minutes'::interval,
'false'::boolean);
ERROR: could not open file base/16406/2072097_fsm: Permission denied
STATEMENT: select _devel.cleanupEvent('10 minutes'::interval,
'false'::boolean);
WARNING: AbortTransaction while
On 12/08/2011 12:54 AM, Tom Lane wrote:
Joseph Shraibmanj...@selectacast.net writes:
All was fine until:
LOG: statement: select _devel.cleanupEvent('10 minutes'::interval,
'false'::boolean);
ERROR: could not open file base/16406/2072097_fsm: Permission denied
That's pretty weird. What
This query is taking much longer on 9.1 than it did on 8.4. Why is it
using a seq scan?
= explain verbose SELECT status,EXISTS(SELECT 1 FROM eventlog e WHERE
e.uid = ml.uid AND e.jobid = ml.jobid AND type = 4),EXISTS(SELECT 1 FROM
eventlog e WHERE e.uid = ml.uid AND e.jobid = ml.jobid AND type =
On 11/17/2011 03:30 PM, Michael Glaesemann wrote:
On Nov 17, 2011, at 14:24, Joseph Shraibman wrote:
This query is taking much longer on 9.1 than it did on 8.4. Why is it
using a seq scan?
Without seeing the table definition (including indexes) as well as the output
of EXPLAIN for 8.4
Umm r/static/stable
--
Sent via pgsql-general mailing list (pgsql-general@postgresql.org)
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-general
After I sent my last email, a light bulb went off. I remembered a
similar problem I had a while ago with parts of postgres not having read
permission. Sure enough after I
[EMAIL PROTECTED] /usr/local/pgsql]# chmod -R a+r *
then restart postgres everything is fine.
This is a IMHO a bug in
Sorry, I didn't realize what you were asking.
[local]:owl=# SHOW TimeZone;
TimeZone
--
EST5EDT
(1 row)
Tom Lane wrote:
Joseph S [EMAIL PROTECTED] writes:
Tom Lane wrote:
Works for me ... what have you got TimeZone set to?
/etc/localtime - /usr/share/zoneinfo/US/Eastern
You
Tom Lane wrote:
Joseph S [EMAIL PROTECTED] writes:
... and when I notice that the tuplesperpage for the indexes is low (or
that the indexes are bigger then the tables themselves) I know it is
time for a VACUUM FULL and REINDEX on that table.
If you are taking the latter as a blind
I've had trouble with NFS files on nfs filesystems disappearing for a
second and reappearing. I had to add a retry loop with a delay in my
code that does file reading. I wouldn't try running a production level
postgres over nfs.
---(end of
Richard Huxton wrote:
Joseph Shraibman wrote:
I'm running:
PostgreSQL 8.2.3 on x86_64-unknown-linux-gnu, compiled by GCC gcc
(GCC) 3.4.6 20060404 (Red Hat 3.4.6-3)
My memory settings are:
work_mem = 64MB
shared_buffers = 128MB
temp_buffers = 32MB
I ran a query that was SELECT field
Richard Huxton wrote:
Joseph Shraibman wrote:
I ran a query that was SELECT field, count(*) INTO TEMP temptable
and it grew to be 10gig (as reported by top)
What was the real query?
First I selected 90634 rows (3 ints) into the first temp table, then I
did select intfield1, count
I'm running:
PostgreSQL 8.2.3 on x86_64-unknown-linux-gnu, compiled by GCC gcc (GCC)
3.4.6 20060404 (Red Hat 3.4.6-3)
My memory settings are:
work_mem = 64MB
shared_buffers = 128MB
temp_buffers = 32MB
I ran a query that was SELECT field, count(*) INTO TEMP temptable and
it grew to be 10gig
Well 1) I'd like to avoid the performance penalty for including debug
symbols and 2) I already built the binary and it is running on a live
system, and I'd like to get debug symbols w/o restarting.
Peter Eisentraut wrote:
Am Mittwoch, 28. März 2007 03:00 schrieb Joseph S:
I don't use rpms, I
Robert Treat wrote:
If you are running pre-8.0 versions you need to update your operating system
(as you indicated). If you running an any 8.x version, you need to be on the
most current corresponding 8.x.y release.
So what happens if you have an old os with a new postgresql install?
Will
8.0.x has the problem that VACUUM FULL on a table does not reclaim space
from the indexes, and I have to issue a separate REINDEX command. Has
this been fixed in later versions?
---(end of broadcast)---
TIP 5: don't forget to increase your free
See example below. At the very least the documentation needs to tell
users that savepoints use shared memory, and the cofusing HINT string
needs to be changed to something more useful.
When run on a machine running 8.2b3
version: PostgreSQL 8.2beta3 on i686-pc-linux-gnu, compiled by GCC gcc
I'm running postgres 8.0.8. I have a table that is updated very
rapidly, so I vacuum it every 10 minutes. The problem is that I
sometimes have transactions that hang out for a long time without doing
anything. These transactions are preventing VACUUM from cleaning up
tuples that were
I have a trigger that updates a count table, based on status. The count
table looks like this:
key status count
a1 300
a2 400
b1 100
b2 200
The problem is that for large updates when I do UPDATE table SET status
= 1 WHERE status = 2 and key =
I'm trying to do this:
IF TG_OP = \'INSERT\' OR (TG_OP = \'UPDATE\' AND OLD.status
NEW.status) THEN
..but pg is complaining:
ERROR: record old is not assigned yet
DETAIL: The tuple structure of a not-yet-assigned record is indeterminate.
CONTEXT: PL/pgSQL function set_dir_count line 4
db:db=explain select * from elog where id = eds('2006-01-01');
QUERY PLAN
---
Seq Scan on elog (cost=0.00..1894975.10 rows=1 width=204)
Filter: (id =
It is STABLE, which I finally figured out. I had to find section 31.6
of the docs, which is nowhere near the part about writing functions.
Merlin Moncure wrote:
On 7/13/06, Joseph Shraibman jks@selectacast.net wrote:
db:db=explain select * from elog where id = eds('2006-01-01
NM I found the documentation.
Joseph Shraibman wrote:
Joshua D. Drake wrote:
Secondly this sounds like a perfect time for you to consider upgrading
to 8.1
and making use of table partitioning.
How does that work, exactly?
---(end of broadcast
Joshua D. Drake wrote:
Secondly this sounds like a perfect time for you to consider upgrading to 8.1
and making use of table partitioning.
How does that work, exactly?
---(end of broadcast)---
TIP 3: Have you checked our extensive FAQ?
Alex Turner wrote:
As an aside note, I would consider a 13 disk RAID 5 a high risk
solution. If you loose just two drives of 13 at the same time, your
data is all gone. If you loose one drive, your array goes into degraded
mode and your read and write performance goes to hell, and your
I'm running 8.0.8 on a raid 5 over 13 disks, and select performance on a
query that needs to join two large tables is very bad. top shows pg
using 2 to 4 percent cpu. Doing a query on one big table uses 30 to 45
percent cpu.
This is RHEL 4 running kernel 2.6.9-22.ELsmp, using an LSI fiber
The verbose output shows the table being vacuumed last. Maybe it
changed after 8.0
Greg Stark wrote:
Jim Nasby [EMAIL PROTECTED] writes:
My RFE: When vacuuming a table, pg should try to vacuum the primary key
first. If that results in 0 recovered entries, then assume the table has no
I like to make sure the vacuum takes place during off peak times, which
is why I don't use autovacuum.
Jim Nasby wrote:
On Jun 22, 2006, at 7:12 PM, Joseph Shraibman wrote:
I'm running a 8.0 database. I have a very large log table that is
rarely updated or deleted from. The nightly vacuum
Is there any performance impact of releasing savepoints?
---(end of broadcast)---
TIP 5: don't forget to increase your free space map settings
Is there a way to get the ip address of the connections listed in
pg_stat_activity?
---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
subscribe-nomail command to [EMAIL PROTECTED] so that your
I have a method in my rmi server that takes a query and returns an
Object[][]. I had this query:
SELECT (select relname from pg_catalog.pg_class where relfilenode =
relation) as relname, * FROM pg_locks;
After upgrading from 7.4 to 8.0 I was getting this problem:
WARNING: Servlet.service()
Oliver Jowett wrote:
Joseph Shraibman wrote:
Is it a jdbc bug that is returning the answer as
org.postgresql.util.PGobject instead of some kind of Number?
The column's type is 'xid' which the driver doesn't currently handle, so
it gets put into the wrap it in PGobject bucket.
Is xid
So basically what needs to be changed is TypeInfoCache.java Oid.java
Alvaro Herrera wrote:
On Thu, Aug 18, 2005 at 03:55:43PM +1200, Oliver Jowett wrote:
Joseph Shraibman wrote:
The column's type is 'xid' which the driver doesn't currently handle,
so it gets put into the wrap
I have this index:
directory_lower_username_seg_key unique, btree (lower(username)
text_pattern_ops, seg)
... but my query refuses to use that index.
[local]:owl=explain select * from directory where lower(username) =
'jks@selectacast.net';
QUERY PLAN
Madison Kelly wrote:
Joseph Shraibman wrote:
What happens if you 'SET enable_seqscan TO OFF' and try the query
again? I've had a couple of instances where the planner just doesn't
like my index but once it is told to use it I get a nice performance boost.
It still does a seqscan
Stephan Szabo wrote:
On Tue, 26 Jul 2005, Joseph Shraibman wrote:
I have this index:
directory_lower_username_seg_key unique, btree (lower(username)
text_pattern_ops, seg)
... but my query refuses to use that index.
text_pattern_ops is an opclass for doing LIKE queries using the index
Stephan Szabo wrote:
It is for the operators ~~, ~=~, ~=~, ~=~, ~~ (for like optimization).
The docs seem to say that it does a character by character comparison
rather than one using the collation thus being better for pattern
matching. I'd think letting it do , =, =, =, would have it
INFO: free space map: 195 relations, 96448 pages stored; 417104 total
pages needed
DETAIL: Allocated FSM size: 1000 relations + 9 pages = 588 kB
shared memory.
I'm confused, do I need to set my fsm settings to 96448 or 417104 based
on this output?
Are fsm settings updated during a
How come when a share lock is held and update can't be done on the
table, but a SELECT FOR UPDATE can be done? I can't SELECT FOR UPDATE
the same row in two transactions, but I can SELECT FOR UPDATE a row that
I will won't be able to update because the other table is held in a
SHARE lock.
is that one transaction can lock the rows, and the other
transaction locks the table, which leads to a deadlock.
Tom Lane wrote:
Joseph Shraibman jks@selectacast.net writes:
How come when a share lock is held and update can't be done on the
table, but a SELECT FOR UPDATE can be done? I can't
I want to do the following:
BEGIN;
SELECT ... FROM table WHERE a = 1 FOR UPDATE;
UPDATE table SET ... WHERE a = 1;
if that resturns zero then
INSERT INTO table (...) VALUES (...);
END;
The problem is that I need to avoid race conditions. Sometimes I get
primary key exceptions on the INSERT.
I'm running:
PostgreSQL 7.4.7 on i686-pc-linux-gnu, compiled by GCC gcc (GCC) 3.2.2
20030222 (Red Hat Linux 3.2.2-5)
I do this:
BEGIN;
SELECT count(*) FROM u, d WHERE u.id = d.id AND ... ;
DECLARE cname CURSOR FOR SELECT u.field, d.field FROM u, d WHERE u.id =
d.id AND ... ;
At the end of the
Doesn't this need to be ammened to say before 8.0?
---(end of broadcast)---
TIP 9: the planner will ignore your desire to choose an index scan if your
joining column's datatypes do not match
---(end of
Scott Marlowe wrote:
Only if you set transaction isolation to serializable.
So am I getting data that was updated up until the time of the FETCH or
the DECLARE CURSOR?
---(end of broadcast)---
TIP 2: you can get off all lists at once with the
How come when I paste a large query into psql it starts off fast but
then slows to a crawl eating up cpu just echoing the query back to me?
I'm using psql 7.4.7
---(end of broadcast)---
TIP 3: if posting/reading through Usenet, please send an
--hopefully your query doesn't start
with rm -rf / : )
Regards,
Paul Tillotson
Joseph Shraibman wrote:
How come when I paste a large query into psql it starts off fast but
then slows to a crawl eating up cpu just echoing the query back to me?
I'm using psql 7.4.7
---(end
http://www.faqs.org/docs/ppbook/r26943.htm
-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Joseph
Shraibman
Sent: Thursday, April 14, 2005 6:19 PM
To: pgsql-general
Subject: [GENERAL] psql performance
How come when I paste a large query into psql it starts off
Alvaro Herrera wrote:
Incidentally when I did that I only got back one row. What's up with that?
Try with union all instead of plain union.
Talk about serendipity. The problem I've been struggling with for the
last few hours has been why my query wasn't producing sorted output even
though I
server: rpm -qa | grep readline
readline-devel-4.3-3
readline-4.3-3
[EMAIL PROTECTED] ~]$ rpm -qa | grep readline
readline-4.3-13
readline-devel-4.3-13
Tom Lane wrote:
Joseph Shraibman jks@selectacast.net writes:
It doesn't matter what the query is. The problem happens before it even
runs the query
Joseph Shraibman wrote:
So the question what is the difference between konsole and xterm that is
causing cpu to be eating up on the server?
Scratch that. I wasn't using the same input for both queries. Both of
them are slow.
I discovered that adding newlines to the query speeds things up
Uwe C. Schroeder wrote:
Don't see a problem pasting this one.
Neither to a local nor to a remote ssh (running psql certainly).
This is 7.4.7 on redhat and mandrake linux'es
I'd suspect it has nothing to do with psql. Can you paste that into a normal
ssh / terminal ?
It is slow just pasting to
How come relation_size() doesn't work on an index?
---(end of broadcast)---
TIP 3: if posting/reading through Usenet, please send an appropriate
subscribe-nomail command to [EMAIL PROTECTED] so that your
message can get through to the
row)
[local]:o=select relation_size('pg_am_oid_index');
ERROR: pg_am_oid_index is an index
Michael Fuhr wrote:
On Tue, Apr 05, 2005 at 12:24:02PM -0400, Joseph Shraibman wrote:
How come relation_size() doesn't work on an index?
Could you define doesn't work? An example with the expected
Can I take the new .c file, do a make install, and have it work in 7.4.7 ?
Michael Fuhr wrote:
On Tue, Apr 05, 2005 at 01:25:01PM -0400, Joseph Shraibman wrote:
local]:o=select version();
version
Is there a function I can call to see if the current user has
permissions on a certain table?
---(end of broadcast)---
TIP 7: don't forget to increase your free space map settings
Has this error from 7.4.6 been fixed in 8.0?
ERROR: Unicode characters greater than or equal to 0x1 are not supported
---(end of broadcast)---
TIP 7: don't forget to increase your free space map settings
I have a table that is usually really small (currently 316 rows) but
goes through spasams of updates in a small time window. Therefore I
have a vacuum full run every hour on this table.
Last night one of these vacuum fulls deadlocked with a query on this
table. Both were stuck doing nothing
Why then when I did a kill -INT on the vacuuming backends did everything
unfreeze?
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
Last night one of these vacuum fulls deadlocked with a query on this
table. Both were stuck doing nothing until I did a kill -INT on the
backends doing
That is what I wanted to know, how to get the evidence for next time.
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
Why then when I did a kill -INT on the vacuuming backends did everything
unfreeze?
You could have had other stuff backed up behind the VACUUM FULL lock
requests.
It's
be executed. Thus, the parsing,
rewriting, and planning stages are only performed once, instead of every
time the statement is executed.
Markus Bertheau wrote:
On Fri, 27 Aug 2004 11:09:26 -0400, Joseph Shraibman
[EMAIL PROTECTED] wrote:
How does EXECUTE solve the cached-plan business?
It re-plans
Martijn van Oosterhout wrote:
The classic issue is what encoding are the databases. Anything other
than C and like won't use indexes.
Unless you use text_pattern_ops. See
http://www.postgresql.org/docs/7.4/static/indexes-opclass.html
I think this needs to be in the faq.
Greg Stark wrote:
Alex [EMAIL PROTECTED] writes:
There won't be anything to VACUUM after the insert, but perhaps you still want
to run ANALYZE. Note that a plain ANALYZE uses a statistical sample which is
much faster, whereas VACUUM ANALYZE has to look at every record anyways so
it's slower but
I recently dumped and restored a 7.4.2 database. It took 30 minutes for
the data to load (6 gig) and 45 for the indexes to be created (3 gig).
Why are the primary keys created after the other indexes? That means
that the table data had been evicted from the cache and has to be reloaded.
What
I have a query like this:
SELECT ... FROM u, d WHERE d.ukey = u.ukey AND restrictions on u AND
(d.status = 3 OR (u.status = 3 AND d.status IN(2,5)));
explain shows:
- Aggregate (cost=126787.04..126787.04 rows=1 width=4)
- Hash Join (cost=39244.00..126786.07 rows=387 width=4)
Does ALTER TABLE table ALTER field SET STATISTICS 100; lock the
table? I just tried to do that while a query is running and the ALTER
is hanging.
---(end of broadcast)---
TIP 7: don't forget to increase your free space map settings
Bruce Momjian wrote:
I have events in the next few weeks in New York City, Copenhagen, Paris,
and Atlanta. Check the News section on the web site for more
information. I will also be in Amsterdam February 2-3, though I have no
public events scheduled there.
You mean the events section, don't
Tom Lane wrote:
Kaare Rasmussen [EMAIL PROTECTED] writes:
Not sure about your position here. You claimed that it would be a good idea to
freeze the on disk format for at least a couple of versions.
I said it would be a good idea to freeze the format of user tables (and
indexes) across multiple
Ron Johnson wrote:
Who's want to build a 40-year-old rocket?
You'd be surpised. Some plans for replacing the shuttle call for going back to Saturn
V's. NASA went with the shuttle design in the first place because resusable was supposed
to be cheaper, but it hasn't turned out that way.
Is there a way to get random rows besides ORDER BY random()? The problem with ORDER BY
random() is that is has to get all the rows from the table before the results are returned.
---(end of broadcast)---
TIP 9: the planner will ignore your desire
I'm having a wierd problem with pg 7.3.3
PostgreSQL 7.3.3 on i686-pc-linux-gnu, compiled by GCC gcc (GCC) 3.2
20020903 (Red Hat Linux 8.0 3.2-7)
I have a bunch of inserts being done by JDBC. One of them if causing a
problem.
LOG: query: INSERT INTO mailtextlog (mlid,tlog,cdate)
Richard Huxton wrote:
On Monday 04 August 2003 04:29, Joseph Shraibman wrote:
Hmm. I didn't work for me. I'll try and figure this out.
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
I notice in 7.3.3 the planner can't tell when a LIKE has no wildcards
and is in reality
Hmm. I didn't work for me. I'll try and figure this out.
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
I notice in 7.3.3 the planner can't tell when a LIKE has no wildcards
and is in reality an '='. Is this an easy change to make?
On what do you base that conclusion
The link at the end of 1.6 in the faq does not work.
---(end of broadcast)---
TIP 3: if posting/reading through Usenet, please send an appropriate
subscribe-nomail command to [EMAIL PROTECTED] so that your
message can get through to the
I have a query where I want to select the usertable records that have a matching entry in
an event table. There are two ways to do this.
1) SELECT COUNT(DISTINCT u.uid) FROM usertable u, eventlog e WHERE u.uid = e.uid AND
e.type = XX;
2) SELECT COUNT(u.uid) FROM usertable u WHERE EXISTS(SELECT
playpen=# create table jm(
playpen(# jid int NOT NULL,
playpen(# mid int ,
playpen(# UNIQUE(jid, mid)
playpen(# );
NOTICE: CREATE TABLE/UNIQUE will create implicit index 'jm_jid_key' for table 'jm'
CREATE
Why isn't the index created called 'jm_jid_mid_key' ?
--
Joseph Shraibman
[EMAIL
user db as
'simon' .
the above error is using 'simon' + password and connection to 'simon' db
I'm trying to trace that from the log. there've /var/log/pgsql which is
0 bytes .
Am I open the right log file of postgresql ?? Thx for reading this
message !!
--
Joseph Shraibman
Philip Crotwell wrote:
Hi
I have a very large database of seismic data. It is about 27 Gb now, and
growing at about the rate of 1 Gb every 3-4 days. I am running
snip
Out of curiosity, how long does it take you to vacuum that?
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal
postgres.
CREATE DATABASE
You are now connected to database mydb as user postgres.
psql:/local/dumpall-8-14:22: \connect: FATAL 1: user
username-goes-here does not exist
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
---(end
Bruce Momjian wrote:
6. Can databases be partitioned over multiple physical files. Can
You have to use symlinks to move to other file systems.
That's not what he asked. He asked about files, and the answer is yes.
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio
Doug McNaught wrote:
[HACKERS removed from CC: list]
Joseph Shraibman [EMAIL PROTECTED] writes:
Doing a dumpall for a backup is taking a long time, the a restore from
the dump files doesn't leave the database in its original state. Could
a command be added that locks all the files
It appears that all my data was restored except for the table which had
the error(which happened to be my biggest table) of which none was
restored.
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
---(end of broadcast
No, I have a redhat 6.x system and I built the postgres myself from the
7.0.3 source.
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
psql:dumpall-2001-4-27:8452: ERROR: copy: line 8933, Bad timestamp
external representation '2001-01-17 19:37:60.00-05'
Are you on Mandrake
, or the query rewriter output to the server log.
DEBUG_PRETTY_PRINT selects
are nicer but longer output format.
The order of the lists do not seem to match.
BTW can I send to the bugs list w/o subscribing? I'm not sure so I'm
sending to general.
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase
If I turn off fsync on 7.1 does that mean that the wal file is sync'd
(according to WAL_SYNC_METHOD in the log file) and other files are not?
Or does fsync apply to all file equally?
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
broadcast)---
TIP 1: subscribe and unsubscribe commands go to [EMAIL PROTECTED]
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
---(end of broadcast)---
TIP 5: Have you c
now better), that the indexes could be
updated on the COMMIT.
Please don't hurt me too bad...
Rob
I imagine because the transaction might do a select on data it just
inserted/updated.
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.ta
of parsing.
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
Ian Lance Taylor wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
A caveat on this reply: I've been studying the Postgres internals, but
I have not mastered them.
I understand that keeping different views for different open
transactions can be difficult, but after a transaction
information to do that. If it doesn't have to worry about
rows that aren't visible.
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
Maybe I'm not making myself understood. Another way of asking the same
thing:
Say there is a transaction that is looking at a non-current
Stephan Szabo wrote:
On Tue, 20 Feb 2001, Joseph Shraibman wrote:
Stephan Szabo wrote:
Where are you seeing something that says the estimator/planner using the
index to get an upper bound? The estimator shouldn't be asking either the
index or the heap for anything, it should
Ian Lance Taylor wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
I understand that keeping different views for different open
transactions can be difficult, but after a transaction that updates a
row is over why isn't the row marked as 'universally visible' for all
new
Mitterrand wrote:
On Thu, Feb 15, 2001 at 04:45:02PM -0500, Joseph Shraibman wrote:
Yes, I have read that thread qnd implemented a similar trigger in the
past.
What I meant to do is detect a change at the _table_ level, not the row
level. Is there such a field somewhere
om session;
count
401094
(1 row)
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
Joseph Shraibman wrote:
Can't postgres do the index lookup first and find out there are only a
few tuples that might match?
Actually it looks like postgres is doing this:
o=# explain select * from usertable where p = 33;
NOTICE: QUERY PLAN:
Seq Scan on usertable (cost=0.00..30.54 rows
I've been thinking on some new pesudo-types like SERIAL.
1) LAST_MODIFIED is the timestamp of that last time this row was
modified. Easy enough to do currently with triggers.
2) TIME_CREATED is the timestamp of when this row was first created with
an INSERT. I'm not sure how to do this because
Look at http://www.selectacast.net/~jks/postgres/gdb2.txt and gdb3.txt
It looks like script put some garbage in the file where I used file
completion in the shell, so just ignore that.
Tom Lane wrote:
Joseph Shraibman [EMAIL PROTECTED] writes:
Postgres has core dumped on me a few
E b _http://www.kimo.com.tw
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
Anand Raman wrote:
there is a white paper present on the great bridge site
(http://www.greatbridge.com/) where u can find some tuning tips .
Where, persactly?
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signal to noise ratio. http://www.targabot.com
It's been done for months...
I've been wondering why we haven't seen 7.1 before now then. I mean why
are you waiting on whatever you are waiting on? Why not release 7.1 now
and 7.2 in January with all the other features you want to add?
--
Joseph Shraibman
[EMAIL PROTECTED]
Increase signa
1 - 100 of 119 matches
Mail list logo