On Tue, Feb 3, 2015 at 9:33 PM, BladeOfLight16 bladeofligh...@gmail.com
wrote:
This is why ORMs are bad. They make hard problems *much* harder, and the
only benefit is that they maybe make easy problems a little quicker. The
cost/savings is *heavily* skewed toward the cost, since there's no
On Mon, Feb 2, 2015 at 1:16 AM, Sam Saffron sam.saff...@gmail.com wrote:
However, the contortions on the above query make it very un-ORM
friendly as I would need to define a view for it but would have no
clean way to pass limits and offsets in.
This is why ORMs are bad. They make hard
Note: I still consider this a bug/missing feature of sorts since the
planner could do better here, and there is no real clean way of
structuring a query to perform efficiently here, which is why I
erroneously cross posted this to hacker initially:
# create table testing(id serial primary key,
On Tue, Feb 3, 2015 at 3:33 PM, Gordon Haverland
ghave...@materialisations.com wrote:
TheRegister is running an article about someone breaking into a dbase,
taking control of the encryption key, and 6 or so months later
demanding ransom from the owner of the dbase.
Hi Adrian,
From: Adrian Klaver adrian.kla...@aklaver.com
Subject: Re: [GENERAL] Ungroup data for import into PostgreSQL
On 01/15/2015 04:56 PM, Jim Nasby wrote:
On 1/15/15 9:43 AM, George Weaver wrote:
Hi List,
I need to import data from a large Excel spreadsheet into a PostgreSQL
table.
On 02/03/2015 04:49 AM, Ramesh T wrote:
Hi ,
i created type on postgres
CREATE TYPE order_list AS (order_id bigint);
it works fine.
then, i try to create a other table type using above created type.
like,
--create or replace type suborder_list_table as
TheRegister is running an article about someone breaking into a dbase,
taking control of the encryption key, and 6 or so months later
demanding ransom from the owner of the dbase.
http://www.theregister.co.uk/2015/02/03/web_ransomware_scum_now_lay_waste_to_your_backups/
Anyone want to comment on
On 02/03/2015 05:16 AM, Ramesh T wrote:
hi,
How to run dbms_scheduler.create_job in postgres and is it
available postgres..?
Postgres != Oracle.
There is no dbms_scheduler in Postgres. You can use cron or pgAgent:
http://www.pgadmin.org/docs/1.20/pgagent.html
As has been pointed
On 2/3/15 7:03 AM, holger.friedrich-fa-triva...@it.nrw.de wrote:
On Tuesday, February 03, 2015 3:58 AM, Jim Nasby wrote:
Note that the recursive grep starts at the current directory, so make sure
you're actually in the pgsql source code when you use it.
cat ~/bin/pg_grep
#!/bin/sh
grep -r
Hi list,
We run a website. We once stored all sorts of files in pg_largeobject,
which grew to 266GB. This is on an m1.large on Amazon EC2 on a single,
magnetic, non-provisioned-IO volume. In that context, 266GB is a lot.
We've since moved all but 60GB of that data to S3. We plan to reduce
that
Sorry for the late reply...life interefered...
From: Jim Nasby jim.na...@bluetreble.com
On 1/15/15 9:43 AM, George Weaver wrote:
Hi List,
I need to import data from a large Excel spreadsheet into a PostgreSQL
table. I have a program that uses ODBC to connect to Excel and extract
data
On Tuesday, February 03, 2015 3:58 AM, Jim Nasby wrote:
Note that the recursive grep starts at the current directory, so make sure
you're actually in the pgsql source code when you use it.
cat ~/bin/pg_grep
#!/bin/sh
grep -r $* * | grep -iv TAGS: | grep -v 'Binary file' | grep -v '.deps/'
On 02/03/2015 07:50 AM, Ramesh T wrote:
Am CCing the list.
CREATE TYPE order_list AS (order_id bigint);
i created above type
Not sure that the above does anything.
and i am using order_list, trying creating table type (datatype)
*create or replace type order_list_table as table of
On Tue, 3 Feb 2015 10:53:11 -0500
Adam Hooper a...@adamhooper.com wrote:
Hi list,
We run a website. We once stored all sorts of files in pg_largeobject,
which grew to 266GB. This is on an m1.large on Amazon EC2 on a single,
magnetic, non-provisioned-IO volume. In that context, 266GB is a
On Tue, Feb 3, 2015 at 12:58 PM, Bill Moran wmo...@potentialtech.com wrote:
On Tue, 3 Feb 2015 10:53:11 -0500
Adam Hooper a...@adamhooper.com wrote:
This plan won't work: Step 2 will be too slow because pg_largeobject
still takes 266GB. We tested `VACUUM FULL pg_largeobject` on our
staging
On Tue, 3 Feb 2015 14:17:03 -0500
Adam Hooper a...@adamhooper.com wrote:
On Tue, Feb 3, 2015 at 12:58 PM, Bill Moran wmo...@potentialtech.com wrote:
On Tue, 3 Feb 2015 10:53:11 -0500
Adam Hooper a...@adamhooper.com wrote:
This plan won't work: Step 2 will be too slow because
On Tue, Feb 3, 2015 at 2:29 PM, Bill Moran wmo...@potentialtech.com wrote:
On Tue, 3 Feb 2015 14:17:03 -0500
Adam Hooper a...@adamhooper.com wrote:
My recommendation here would be to use Slony to replicate the data to a
new server, then switch to the new server once the data has synchornized.
On Tue, 3 Feb 2015 14:48:17 -0500
Adam Hooper a...@adamhooper.com wrote:
On Tue, Feb 3, 2015 at 2:29 PM, Bill Moran wmo...@potentialtech.com wrote:
On Tue, 3 Feb 2015 14:17:03 -0500
Adam Hooper a...@adamhooper.com wrote:
My recommendation here would be to use Slony to replicate the data
BEGIN
EXECUTE IMMEDIATE 'DROP TABLE CONTAINER';
EXCEPTION
WHEN OTHERS THEN
IF SQLCODE != -942 THEN
RAISE;
END IF;
END;
Jim nailed it. In PostgreSQL, this is just
DROP TABLE IF EXISTS CONTAINER;
One line. No dynamic SQL, exception block, or even a block
19 matches
Mail list logo