Thanks all,
I went reading about the nfile and it does look like it is the
problem.. I could rework the script to just dump to a flat file (which I
had in the past) and have another script read it in after the fact with
only one db connection.. but that's kind of inefficient..
Travis
-----Original Message-----
From: Tom Lane [mailto:tgl@sss.pgh.pa.us]
Sent: Friday, October 18, 2002 10:42 PM
To: Martijn van Oosterhout
Cc: Williams, Travis L, NPONS; pgsql-general@postgresql.org
Subject: Re: [GENERAL] table overflow question
Martijn van Oosterhout <kleptog@svana.org> writes:
> Looks like you've run into a open file limit. If you're using linux
you
> should look in /proc/sys/fs to make sure you can actually open the
number of
> files you need. You should estimate at least 40 files per server.
> I think file-max is the one you want.
He said he was using HPUX. On HPUX 10.20, the kernel parameters NFILE
and NINODE would be the things to bump up; I suspect 11 is the same.
The other direction to attack it from is to reduce PG's parameter
MAX_FILES_PER_PROCESS, but if you have to set that lower than 100
or so then you'd be better advised to fix the kernel.
regards, tom lane