[GENERAL] invalid connection type listen_addresses='*'

2014-07-10 Thread Aram Fingal
I just tried to set up a PostgreSQL server on an existing instillation of Ubuntu 13.10 server but I am getting an error trying to start the server and I am not finding anything relevant to the error searching the web. Here’s what I did to install: $ sudo apt-get install postgresql $ sudo

Re: [GENERAL] invalid connection type listen_addresses='*'

2014-07-10 Thread Aram Fingal
listen_addresses='*' parameter doesn't belong in pg_hba.conf This parameter should be in postgresql.conf Thanks. That was really unclear, at least the way I followed the online documentation: http://www.postgresql.org/docs/9.1/static/auth-pg-hba-conf.html …even after following the

Re: [GENERAL] Convert an XML database

2014-05-21 Thread Aram Fingal
a generic ETL (extract, transform, load) tool that supports XML and postgresql should be able to do it, but its not something built into postgresql as-is.I can't recommend any specific ETL tool to do this as I've never needed to do this exactly. Thanks. That at least tells me that I

[GENERAL] Convert an XML database

2014-05-20 Thread Aram Fingal
I want to set up a new PostgreSQL database from an XML database file. It seems like there should be an obvious way to do this but I can't seem to find any directions anywhere. How is this sort of thing normally done? The database is available to the public here:

[GENERAL] Alternate input for user defined data type

2013-08-20 Thread Aram Fingal
I want to create a user defined data type but have flexible input just like, for example, the boolean data type where TRUE can be entered as any of (TRUE, true, T, t, YES, yes, Y, y, 1...) and it will be interpreted as the same thing. So suppose I have days of the week: CREATE TYPE

Re: [GENERAL] Multiple indexes, huge table

2012-09-07 Thread Aram Fingal
On Sep 7, 2012, at 11:15 AM, Marti Raudsepp wrote: There's a pg_bulkload extension which does much faster incremental index updates for large bulk data imports, so you get best of both worlds: http://pgbulkload.projects.postgresql.org/ Thanks, I'll have to check that out. This is going to

[GENERAL] Multiple indexes, huge table

2012-09-06 Thread Aram Fingal
I have a table which currently has about 500 million rows. For the most part, the situation is going to be that I will import a few hundred million more rows from text files once every few months but otherwise there won't be any insert, update or delete queries. I have created five indexes,

Re: [GENERAL] Multiple indexes, huge table

2012-09-06 Thread Aram Fingal
On Sep 6, 2012, at 5:54 PM, Tom Lane wrote: That sounds like you lack an index on the referencing column of the foreign key constraint. Postgres doesn't require you to keep such an index, but it's a really good idea if you ever update the referenced column. Thanks. You're right. That

Re: [GENERAL] crosstab

2012-09-04 Thread Aram Fingal
Has anyone tried developing a PLR procedure to do the equivalent of crosstab but using the Reshape library from R? The reason I ask is that I have been extracting tables like this with an R script, using RPostgreSQL, then functions from Reshape and writing the output to a .csv file. I wonder

Re: [GENERAL] crosstab

2012-09-04 Thread Aram Fingal
On Sep 4, 2012, at 2:51 PM, Vincent Veyron wrote: Le mardi 04 septembre 2012 à 08:39 -0700, punnoose a écrit : hi all How could i use crostab to display variable number of columns. in the output There could be variable number of columns see the documentation for Additional Supplied

Re: [GENERAL] crosstab

2012-09-04 Thread Aram Fingal
On Sep 4, 2012, at 3:26 PM, Joe Conway wrote: On 09/04/2012 12:17 PM, Aram Fingal wrote: On Sep 4, 2012, at 2:51 PM, Vincent Veyron wrote: see the documentation for Additional Supplied Modules, in your case tablefunc : http://www.postgresql.org/docs/9.1/static/tablefunc.html I

Re: [GENERAL] crosstab

2012-09-04 Thread Aram Fingal
On Sep 4, 2012, at 3:56 PM, Joe Conway wrote: On 09/04/2012 12:48 PM, Aram Fingal wrote: So, are you saying that if I do something like this: copy(crosstab(source_sql, category_sql)) to '/output.csv' with csv; Then I don't have to list what the columns are going to be? In other words

Re: [GENERAL] crosstab

2012-09-04 Thread Aram Fingal
On Sep 4, 2012, at 4:18 PM, Misa Simic wrote: Inside PL/R you can take the same table as it is (unpivoted) as your data.frame and then pivot it inside R using reshape package,,, And then inside PL/R function do whatever you would like to do with data i.e export to whatever... - but you

Re: [GENERAL] crosstab

2012-09-04 Thread Aram Fingal
On Sep 4, 2012, at 4:36 PM, A.M. wrote: Or you could return the heatmap/plot as BYTEA data or use arrays as necessary. I was actually thinking exactly the same thing. Part of the reason I use PostgreSQL for all my bioinformatics work is that there is a need to correctly associate analysis

Re: [GENERAL] Interval to months

2012-08-15 Thread Aram Fingal
know the exact standard used by the source for a month, though I could ask if it becomes an issue. -Aram Fingal

[GENERAL] Interval to months

2012-08-07 Thread Aram Fingal
I have a field which contains an interval value and I sometimes need to represent the full interval (not a part) as a decimal number of months. For example, 5 years 6 mons 3 days as 66.1 months. I've been trying to figure out how to do this and haven't found a definitive answer. The

[GENERAL] Group by with insensitive order

2011-01-19 Thread Aram Fingal
as acetaminophen, aspirin and I want these grouped together in the query. Is there a simple way to do this? --Aram Fingal -- Sent via pgsql-general mailing list (pgsql-general@postgresql.org) To make changes to your subscription: http://www.postgresql.org/mailpref/pgsql-general

Re: [GENERAL] Group by with insensitive order

2011-01-19 Thread Aram Fingal
Easy way is something like SELECT LEAST(drug1, drug2), GREATEST(drug1, drug2), AVG(response) FROM data GROUP BY 1, 2 though it'd be a PITA to scale that to more than 2 drugs. regards, tom lane Thanks, Tom and Hubert, who said the same thing. For the foreseeable

[GENERAL] Alter table to on update cascade

2010-11-17 Thread Aram Fingal
I have a table where I should have declared a foreign key with ON UPDATE CASCADE and didn't. Now I want to fix that. From the documentation on www.postgresql.org, about ALTER TABLE it's not at all clear how to do this or even whether you can do this. -Aram -- Sent via pgsql-general

Re: [GENERAL] Alter table to on update cascade

2010-11-17 Thread Aram Fingal
On Nov 17, 2010, at 12:42 PM, Richard Broersma wrote: ALTER TABLE foo DROP CONSTRAINT your_constraint, ADD CONSTRAINT your_constraint FOREIGN KEY ... ON UPDATE CASCADE ON DELETE RESTRICT; Thanks. That worked. -Aram -- Sent via pgsql-general mailing list

Re: [GENERAL] Schema tool

2010-11-15 Thread Aram Fingal
Thanks, each of you for all the suggestions on schema generating tools. The idea is to have something which will connect to the database and automatically make a schema from what you've got. Here's what I have had a chance to evaluate so far... DBVisualizer - It does a good job with the

Re: [GENERAL] Schema tool

2010-11-15 Thread Aram Fingal
On Nov 11, 2010, at 4:56 PM, Thomas Kellerer wrote: Actually I think it would be worthwhile documenting your experience in the PostgreSQL Wiki as well: http://wiki.postgresql.org/wiki/Community_Guide_to_PostgreSQL_GUI_Tools Thanks, I will post something there after I have done some more

[GENERAL] Schema tool

2010-11-11 Thread Aram Fingal
A while back, I thought I remembered seeing a Mac OS X client for PostgreSQL which had a feature where it would display a graphic schema of whatever database you connect to but I can't seem to find it again (web searching.)I did come across one post which said that this was a planned

Re: [GENERAL] Schema tool

2010-11-11 Thread Aram Fingal
Thanks, each of you for all the suggestions on schema generating tools. I haven't had a chance to evaluate them all yet but DBVisualizer looks pretty good. In the meanwhile I found SQL Power Architect, which is also free/open source, and can do this kind of diagraming but is not as good as

[GENERAL] Installing Contrib Modules with a Precompiled Binary

2010-09-20 Thread Aram Fingal
I'm using the OS X precompiled binary from EnterpriseDB and want to add the tablefunc contrib module. I haven't been able to find any documentation about how to do this or even whether modules can be added to this binary version. I suppose that it may be a good idea to export my databases

Re: [GENERAL] Installing Contrib Modules with a Precompiled Binary

2010-09-20 Thread Aram Fingal
On Sep 20, 2010, at 2:12 PM, Dave Page wrote: On Mon, Sep 20, 2010 at 7:06 PM, Aram Fingal fin...@multifactorial.com wrote: I'm using the OS X precompiled binary from EnterpriseDB and want to add the tablefunc contrib module. I haven't been able to find any documentation about how to do

Re: [GENERAL] Transposing rows and columns

2010-09-17 Thread Aram Fingal
On Sep 17, 2010, at 9:00 AM, Steve Clark wrote: I think excel 2007 can handle more than 65,535 rows. You may be right. I'm actually using NeoOffice (Mac enhanced version of OpenOffice) and that can handle something like 1,048,000 rows.I wouldn't be surprised if newer versions of Excel

[GENERAL] Transposing rows and columns

2010-09-16 Thread Aram Fingal
I'm working with some people who live and breath Excel. I need to be able to move data back and forth between formats which make sense for Excel and for PostgreSQL. In some cases, this is just to accommodate what people are used to. In other cases, like statistical clustering, it's something

Re: [GENERAL] Transposing rows and columns

2010-09-16 Thread Aram Fingal
On Sep 16, 2010, at 12:28 PM, Sam Mason wrote: On Thu, Sep 16, 2010 at 11:42:21AM -0400, Aram Fingal wrote: create table results( expt_no int references experiments(id), subject int references subjects(id), drug text references drugs(name), dose numeric, response numeric ) What's

Re: [GENERAL] Transposing rows and columns

2010-09-16 Thread Aram Fingal
On Sep 16, 2010, at 4:37 PM, John R Pierce wrote: On 09/16/10 10:44 AM, Aram Fingal wrote: I have thought about that but later on, when we do the full sized experiments, there will be too many rows for Excel to handle. if you insist on this transposing, won't that mean you'll end up