Обсуждение: SQL/MED estimated time of arrival?
Hi SQL/MED developers,<br /><br /> Our company has just finished development of a database extension for Informix that providestabular access to various types of structured files (NetCDF and HDF5, with more types to come). We would like toport this logic to run on PostgreSQL, since many of our potential customers use PostgreSQL.<br /><br /> On Informix, wewere able to take advantage of the VTI (Virtual Table Interface) feature to support "table" scans and indexing. (See <aeudora="autourl" href="http://www.ibm.com/developerworks/data/zones/informix/library/techarticle/db_vti.html"> http://www.ibm.com/developerworks/data/zones/informix/library/techarticle/db_vti.html</a>.) Do you have any idea of how longit will be before SQL/MED on PostgreSQL will be available, and perhaps how similar it will be to Informix VTI?<br /><br/> Thanks,<br /> Eric.<br /><p> ********************************************** <br /> Eric Davies, M.Sc. <br /> SeniorProgrammer Analyst<br /> Barrodale Computing Services Ltd. <br /> 1095 McKenzie Ave., Suite 418<br /> Victoria BC V8P2L5<br /> Canada<br /><br /> Tel: (250) 704-4428<br /> Web: <a eudora="autourl" href="http://www.barrodale.com/"> <fontcolor="#0000FF"><u>http://www.barrodale.com</u></font></a><br /> Email: eric@barrodale.com <br /> **********************************************<br/><br /><br />
On Wed, 03 Nov 2010 13:32:18 -0700 Eric Davies <eric@barrodale.com> wrote: > On Informix, we were able to take advantage of the VTI (Virtual Table > Interface) feature to support "table" scans and indexing. (See > http://www.ibm.com/developerworks/data/zones/informix/library/techarticle/db_vti.html > .) Do you have any idea of how long it will be before SQL/MED on > PostgreSQL will be available, and perhaps how similar it will be to > Informix VTI? SQL/MED is now under discussion/development for PostgreSQL 9.1, and 9.1 would be released one year after 9.0, maybe around Sep 2011? For detail of release schedule, please see the development plan of PostgreSQL 9.1. http://wiki.postgresql.org/wiki/PostgreSQL_9.1_Development_Plan I looked into VTI documents you've pointed. ISTM that VTI and SQL/MED would have a lot of common ideas, and most of VTI items would be able to be mapped to one of SQL/MED items, except features about updating data and indexing. For example: * PRIMARY ACCESS_METHOD -> HANDLER of FOREIGN DATA WRAPPER * am_scancost() -> FdwRoutine.EstimateCosts() * am_open() -> FdwRoutine.Open() * am_beginscan() -> first call of FdwRoutine.Iterate()? * am_getnext() -> FdwRoutine.Iterate() * am_rescan() -> FdwRoutine.ReOpen() * am_close() -> FdwRoutine.Close() * Table descriptor -> Relation, Form_pg_class * Qual descriptor -> PlanState.qual I hope the summary of SQL/MED described in wiki page helps you. http://wiki.postgresql.org/wiki/SQL/MED Any comments and questions are welcome. Regards, -- Shigeru Hanada
On Thu, Nov 4, 2010 at 6:04 PM, Shigeru HANADA <hanada@metrosystems.co.jp> wrote: > For example: > * PRIMARY ACCESS_METHOD -> HANDLER of FOREIGN DATA WRAPPER > * am_scancost() -> FdwRoutine.EstimateCosts() > * am_open() -> FdwRoutine.Open() > * am_beginscan() -> first call of FdwRoutine.Iterate()? It might be good to have a separated "beginscan" method if we use asynchronous scans in multiple foreign servers in one query because multiple foreign servers can run their queries in parallel. (Imagine that pushing-down aggregate function into each foreign server.) I think it is different from "open" because it is called before query execution, for example by EXPLAIN. > * am_getnext() -> FdwRoutine.Iterate() > * am_rescan() -> FdwRoutine.ReOpen() > * am_close() -> FdwRoutine.Close() > * Table descriptor -> Relation, Form_pg_class > * Qual descriptor -> PlanState.qual Do you think you have all counterpart methods for VTI AMs? If so, it's a good news ;-) We could support foreign table features as same level as Informix. -- Itagaki Takahiro
On Thu, 4 Nov 2010 18:22:52 +0900 Itagaki Takahiro <itagaki.takahiro@gmail.com> wrote: > On Thu, Nov 4, 2010 at 6:04 PM, Shigeru HANADA > <hanada@metrosystems.co.jp> wrote: > > For example: > > * PRIMARY ACCESS_METHOD -> HANDLER of FOREIGN DATA WRAPPER > > * am_scancost() -> FdwRoutine.EstimateCosts() > > * am_open() -> FdwRoutine.Open() > > * am_beginscan() -> first call of FdwRoutine.Iterate()? > > It might be good to have a separated "beginscan" method if we use > asynchronous scans in multiple foreign servers in one query > because multiple foreign servers can run their queries in parallel. > (Imagine that pushing-down aggregate function into each foreign server.) You mean that separated beginscan (FdwRoutine.BeginScan?) starts asynchronous query and returns immediately, and FdwRoutine.Iterate returns result of that query? Pushing aggregate down to foreign server would be efficient, but need another hook which can create one ForeignScan node which have "Agg + ForeignScan" functionality. Same optimization would be able to apply for Sort and Limit. Such optimization should be done in optimizer with estimated costs? Or FDW's hook function may change plan tree which was created by planner? > I think it is different from "open" because it is called > before query execution, for example by EXPLAIN. Right, I've misunderstood. VTI programmer's guide says that am_open is called before processing SQL to initialize input or output, and called for not only SELECT but also other queries using a virtual table such as INSERT and DROP TABLE. The am_open would have no counterpart in SQL/MED. > Do you think you have all counterpart methods for VTI AMs? > If so, it's a good news ;-) We could support foreign table > features as same level as Informix. Not all, but most of them for read-only access. VTI supports updating external data and various management tasks via SQL, but SQL/MED supports (at least in standard) only read access. The full set of ACCESS_METHOD functions are: am_create CREATE FOREIGN TABLE am_drop DROP TABLE am_stats gather statistics (ANALYZE) am_check verify data structure and index consistency am_open initialize access to a virtual table (might connect to external server) am_close finalize access to a virtual table am_scancost estimate cost of a scan am_beginscan initialize scan am_getbyid get a tuple by row-id am_getnext get next tuple(s) am_rescan reset state of scanning am_endscan finalize scan am_insert insert a tuple and return row-id am_update update a tuple by row-id am_delete delete atuple by row-id am_truncate truncate table VTI might be similar to storage engine of MySQL or heap-am of PG, rather than SQL/MED of PG. Like FOREIGN INDEX of HiRDB, Informix has Virtual Index Interface, and am_getbyid is used to get a tuple by row-id. I'll research more about VTI and VII for revising design of SQL/MED. Regards, -- Shigeru Hanada
On Fri, Nov 5, 2010 at 4:00 PM, Shigeru HANADA <hanada@metrosystems.co.jp> wrote: >> > * am_beginscan() -> first call of FdwRoutine.Iterate()? >> It might be good to have a separated "beginscan" method if we use >> asynchronous scans in multiple foreign servers in one query > > You mean that separated beginscan (FdwRoutine.BeginScan?) starts > asynchronous query and returns immediately, and FdwRoutine.Iterate > returns result of that query? Yes. Each BeginScan() in the executor node tree will be called at the beginning of executor's run. The callback should not block the caller. OTOH, Iterate() are called at the first time tuples in the node are required. PL/Proxy has a similar functionality with RUN ON ALL to start queries in parallel. So, I think it's a infrastructure commonly required. -- Itagaki Takahiro
On Fri, 5 Nov 2010 16:27:49 +0900 Itagaki Takahiro <itagaki.takahiro@gmail.com> wrote: > On Fri, Nov 5, 2010 at 4:00 PM, Shigeru HANADA > <hanada@metrosystems.co.jp> wrote: > >> > * am_beginscan() -> first call of FdwRoutine.Iterate()? > >> It might be good to have a separated "beginscan" method if we use > >> asynchronous scans in multiple foreign servers in one query > > > > You mean that separated beginscan (FdwRoutine.BeginScan?) starts > > asynchronous query and returns immediately, and FdwRoutine.Iterate > > returns result of that query? > > Yes. Each BeginScan() in the executor node tree will be called at > the beginning of executor's run. The callback should not block > the caller. OTOH, Iterate() are called at the first time tuples > in the node are required. Thanks, now I see your point. Current FdwRoutine has no appropriate function because Open is called from ExecutorStart which is used by EXPLAIN too. But then we have mismatch between executor node interface and FDW interface about BeginScan. Should we add new function such as ExecBeginNode and call ExecBeginXXX for each plan node? New Query Processing Control Flow would be: # based on README of executor directory CreateQueryDesc ExecutorStart CreateExecutorState creates per-query context switch to per-query context to run ExecInitNode ExecInitNode --- recursively scans plan tree CreateExprContext creates per-tuplecontext ExecInitExpr ExecutorRun ExecBeginNode(new) --- recursively scans plan tree call ExecBeginXXXS for each plan node ExecProcNode --- recursively called in per-query context ExecEvalExpr --- called in per-tuple context ResetExprContext --- to free memory ExecutorEnd ExecEndNode --- recursively releases resources FreeExecutorState frees per-query contextand child contexts FreeQueryDesc > PL/Proxy has a similar functionality with RUN ON ALL to start queries > in parallel. So, I think it's a infrastructure commonly required. I noticed the lack of consideration about cache invalidation from reading PL/Proxy source, thanks for your mention about PL/Proxy. :-) Regards, -- Shigeru Hanada
Shigeru HANADA <hanada@metrosystems.co.jp> writes: > Thanks, now I see your point. Current FdwRoutine has no appropriate > function because Open is called from ExecutorStart which is used by > EXPLAIN too. > But then we have mismatch between executor node interface and FDW > interface about BeginScan. Should we add new function such as > ExecBeginNode and call ExecBeginXXX for each plan node? That seems like a massive amount of new code, and wasted cycles during every query startup, to fix a very small problem. There's a flag EXEC_FLAG_EXPLAIN_ONLY that tells node Init functions whether the query is going to be run "for real" or only EXPLAINed. Use that to decide whether to do any real work. regards, tom lane
2010/11/5 Shigeru HANADA <hanada@metrosystems.co.jp>: > On Fri, 5 Nov 2010 16:27:49 +0900 > Itagaki Takahiro <itagaki.takahiro@gmail.com> wrote: >> PL/Proxy has a similar functionality with RUN ON ALL to start queries >> in parallel. So, I think it's a infrastructure commonly required. > I noticed the lack of consideration about cache invalidation from > reading PL/Proxy source, thanks for your mention about PL/Proxy. :-) And if we really make this async query come true, I suggest designing resource (i.e. remote connection) management very carefully. When the executor fails in the middle of its execution, it possibly fails to release its own resource; close() in ExecutorEnd() will never be called. As far as I know files and memory are released automatically in the current mechanism, but MED APIs will use their own resources other than them. Regards, -- Hitoshi Harada
Hitoshi Harada <umi.tanuki@gmail.com> writes: > And if we really make this async query come true, I suggest designing > resource (i.e. remote connection) management very carefully. When the > executor fails in the middle of its execution, it possibly fails to > release its own resource; close() in ExecutorEnd() will never be > called. As far as I know files and memory are released automatically > in the current mechanism, but MED APIs will use their own resources > other than them. The way to fix that is for the FDW to hook into the ResourceOwner mechanism (via RegisterResourceReleaseCallback). Then it can track and clean up things it knows about just as "automatically" as anything else is. Of course, if you lose your network connection to the remote DB, you have to assume it will clean up of its own accord. regards, tom lane
On Fri, 05 Nov 2010 10:43:45 -0400 Tom Lane <tgl@sss.pgh.pa.us> wrote: > Shigeru HANADA <hanada@metrosystems.co.jp> writes: > > Thanks, now I see your point. Current FdwRoutine has no appropriate > > function because Open is called from ExecutorStart which is used by > > EXPLAIN too. > > > But then we have mismatch between executor node interface and FDW > > interface about BeginScan. Should we add new function such as > > ExecBeginNode and call ExecBeginXXX for each plan node? > > That seems like a massive amount of new code, and wasted cycles during > every query startup, to fix a very small problem. Agreed. > There's a flag EXEC_FLAG_EXPLAIN_ONLY that tells node Init functions > whether the query is going to be run "for real" or only EXPLAINed. > Use that to decide whether to do any real work. I missed that flag. That flag would make ExecInitForeignScan be able to skip calling FdwRoutine.BeginScan when the query was an EXPLAIN without ANALYZE. Thanks for the suggestion. Regards, -- Shigeru Hanada
On Sat, 6 Nov 2010 16:04:37 +0900 Hitoshi Harada <umi.tanuki@gmail.com> wrote: > 2010/11/5 Shigeru HANADA <hanada@metrosystems.co.jp>: > > On Fri, 5 Nov 2010 16:27:49 +0900 > > Itagaki Takahiro <itagaki.takahiro@gmail.com> wrote: > >> PL/Proxy has a similar functionality with RUN ON ALL to start queries > >> in parallel. So, I think it's a infrastructure commonly required. > > I noticed the lack of consideration about cache invalidation from > > reading PL/Proxy source, thanks for your mention about PL/Proxy. :-) > > And if we really make this async query come true, I suggest designing > resource (i.e. remote connection) management very carefully. When the > executor fails in the middle of its execution, it possibly fails to > release its own resource; close() in ExecutorEnd() will never be > called. As far as I know files and memory are released automatically > in the current mechanism, but MED APIs will use their own resources > other than them. Yes, managegement of FDW's resources is very important issue. Curren FdwRoutine includes ConnectServer and FreeFSConnection, but they might not be enough to manage FDW's resources by backend in common way. Because connection is not only resource FDW use. Possible resources are: - Files (Virtual File descriptor would help to manage) - Database connections (might be cached) - Server-side cursors (wouldbe released with DB connection?) - Heap memory (for instance, libpq uses malloc) For example, if postgresql_fdw uses server-side cursor to retreive result tuples, it would be required to CLOSE cursors at the end of transaction. Closing cursor at the end of session wouldn't be good idea because clients might pool and reuse connections. How about removing them, ConnectServer and FreeFSConnection, from FdwRoutine and leaving the responsibility of resource management to each FDW? Each FDW would have to use mechanism such as Virtual File and ResourceOwner to manage resources properly, though. Regards, -- Shigeru Hanada
On Fri, 5 Nov 2010 16:27:49 +0900 Itagaki Takahiro <itagaki.takahiro@gmail.com> wrote: > On Fri, Nov 5, 2010 at 4:00 PM, Shigeru HANADA > <hanada@metrosystems.co.jp> wrote: > >> > * am_beginscan() -> first call of FdwRoutine.Iterate()? > >> It might be good to have a separated "beginscan" method if we use > >> asynchronous scans in multiple foreign servers in one query > > > > You mean that separated beginscan (FdwRoutine.BeginScan?) starts > > asynchronous query and returns immediately, and FdwRoutine.Iterate > > returns result of that query? > > Yes. Each BeginScan() in the executor node tree will be called at > the beginning of executor's run. The callback should not block > the caller. OTOH, Iterate() are called at the first time tuples > in the node are required. Please find attached WIP patch for BeginScan. Postgresql_fdw has been changed to use server-side cursor for sample. It's DECLAREd with HOLD option to avoid transaction management, though. Other changes since 20101025 patch are: - Some document fixes. - Don't call ConnectServer from ExecInitForeignScan. Instead, postgresql_fdw calls it from pgOpen(). This change is only trial and would be removed later. - Add "schema" column to output of \det psql command. - New \dE psql command shows list of foreign tables in \d format. - \d+ <foreign table> psql command shows per-column options. If the changes (at least adding BeginScan) are OK, I'll clean the patch up and post it soon. Regards, -- Shigeru Hanada
Вложения
Shigeru HANADA <hanada@metrosystems.co.jp> writes: > How about removing them, ConnectServer and FreeFSConnection, from > FdwRoutine and leaving the responsibility of resource management to > each FDW? Each FDW would have to use mechanism such as Virtual File > and ResourceOwner to manage resources properly, though. For the most part, we expect that ResourceOwners only do something useful during error cleanup. That is, you *should* have a close-connection type of function that is expected to be called during normal query shutdown. The ResourceOwner hooks will operate to compensate for the lack of this call in an error recovery case. The reason for doing things that way is so that we can check for unintentional resource leakage in the non-error code paths. regards, tom lane
On Mon, 08 Nov 2010 10:31:22 -0500 Tom Lane <tgl@sss.pgh.pa.us> wrote: > Shigeru HANADA <hanada@metrosystems.co.jp> writes: > > How about removing them, ConnectServer and FreeFSConnection, from > > FdwRoutine and leaving the responsibility of resource management to > > each FDW? Each FDW would have to use mechanism such as Virtual File > > and ResourceOwner to manage resources properly, though. > > For the most part, we expect that ResourceOwners only do something > useful during error cleanup. That is, you *should* have a > close-connection type of function that is expected to be called during > normal query shutdown. The ResourceOwner hooks will operate to > compensate for the lack of this call in an error recovery case. > The reason for doing things that way is so that we can check for > unintentional resource leakage in the non-error code paths. I fixed postgresql_fdw to use RegisterResourceReleaseCallback() to close all connections in error cases including user interrupt. But I'm not sure if I used the mechanism correctly because all I could find about the API was only few documents, README of resowner and function comments. I tested the codes in cases below and confirmed that all connections have been closed. - remote query error with wrong relation name - user interrupt, Ctrl+C on psql during long query Of course, in non-error case, all connections are closed via normal query shutdown path. During fixing connection cleanup, I've removed connection pooling from postgresql_fdw to make resource management simple. Now postgresql_fdw uses one connection for all of ForeignScan nodes in a local query, but doesn't keep the connection alive beyond queries. Originally, sharing connection is intended to execute multiple remote query in a transaction for consistency. I think external tools such as pgpool or pgbouncer would be better to pool connections. Is it reasonable? Also, ExecInitForeignScan() was fixed to call ConnectServer() and BeginScan() only when the EXEC_FLAG_EXPLAIN_ONLY is not set. Regards, -- Shigeru Hanada
Вложения
Hi Gentlemen,<br /><br /> Thank you for the time estimate and the interface discussion. It sounds like the PostgreSQL SQL/MEDcode will be very useful when it is done. Our product provides read-only access to files, so updates/inserts/deletesaren't an issue for us.<br /><br /> One thing that is not clear to me is indexing support. Will itbe possible to index a SQL/MED table as if it were a regular table? What would be the equivalent of Informix's row ids?<br /><br /> Eric.<br /><br /><br /><p> ********************************************** <br /> Eric Davies, M.Sc. <br/> Senior Programmer Analyst<br /> Barrodale Computing Services Ltd. <br /> 1095 McKenzie Ave., Suite 418<br /> VictoriaBC V8P 2L5<br /> Canada<br /><br /> Tel: (250) 704-4428<br /> Web: <a eudora="autourl" href="http://www.barrodale.com/"><font color="#0000FF"><u>http://www.barrodale.com</u></font></a><br /> Email: eric@barrodale.com<br /> **********************************************<br /><br /><br />
On Fri, 12 Nov 2010 08:27:54 -0800 Eric Davies <eric@barrodale.com> wrote: > Thank you for the time estimate and the interface discussion. It > sounds like the PostgreSQL SQL/MED code will be very useful when it > is done. Our product provides read-only access to files, so > updates/inserts/deletes aren't an issue for us. > > One thing that is not clear to me is indexing support. Will it be > possible to index a SQL/MED table as if it were a regular table? No, SQL/MED would not support indexing foreign tables, at least in first version. Because it would be difficult to use common row id for various FDWs. To support indexing foreign tables might need to change common structure of index tuple to be able to hold virtual row-id, not ItemPointerData. Instead, FDW can handle expressions which are parsed from WHERE clause and JOIN condition of original SQL, and use them to optimize scanning. For example, FDW for PostgreSQL pushes some conditions down to remote side to decrease result tuples to be transferred. I hope this idea helps you. > What > would be the equivalent of Informix's row ids? Answer to the second question would be "ItemPointerData". It consists of a block number and an offset in the block, and consume 6 bytes for each tuple. With this information, PostgreSQL can access to a data tuple directly. Actual definition is: typedef struct ItemPointerData { BlockIdData ip_blkid; OffsetNumber ip_posid; } ItemPointer; Does Informix uses common row-id (AFAIK it's 4 bytes integer) for both of virtual tables and normal tables? Regards, -- Shigeru Hanada
On Mon, Nov 15, 2010 at 12:41, Shigeru HANADA <hanada@metrosystems.co.jp> wrote: > No, SQL/MED would not support indexing foreign tables, at least in > first version. Because it would be difficult to use common row id for > various FDWs. I think the reason is the SQL standard never mention about indexes. It is not a specific issue for SQL/MED. > To support indexing foreign tables might need to change > common structure of index tuple to be able to hold virtual row-id, not > ItemPointerData. I'm not sure we actually need foreign indexes because the query text sent to another server is same whether the foreign table has indexes. Of course, foreign indexes might be useful to calculate costs to scan foreign tables, but the cost also comes from non-index conditions. I think foreign table and foreign index are a model for row-based databases, including postgres. But other DBs might have different cost models. So, it would be better to encapsulate such operations in FDW. -- Itagaki Takahiro
With Informix VTI, indexing is the same for native tables as for virtual tables, except the interpretation of the 32 bitrowid is left up to the developer. When you define the VTI class, you optionally supply a method that can fetch data basedon a 32 bit rowid, and it's the responsibility of your non-indexed scanning methods to provide rowids along with therow tuple. <br /><br /> Having local indexes can be very useful if you have a user that issues queries like:<br /> select count(*) from some_external_table where .... ;<br /> With VTI, the "count" aggregate doesn't get pushed down, meaningthat without a local index, your scanning method has to return as many tuples as match the where clause, which canbe very slow.<br /><br /> Local indexes also affords the opportunity of using specialized indexes built into the database.My guess is that without some form of rowids being passed back and forth, you couldn't define non-materialized viewsof virtual tables that could be indexed.<br /><br /> That said, we implemented our own btree-like index that used thepushed down predicates because fetching data one row at a time wasn't desirable with our design choices, and we wantedto support virtual tables with more than 4 billion rows.<br /><br /> Eric<br /> At 07:41 PM 11/14/2010, Shigeru HANADAwrote:<br /><blockquote cite="" class="cite" type="cite">On Fri, 12 Nov 2010 08:27:54 -0800<br /> Eric Davies <eric@barrodale.com>wrote:<br /> > Thank you for the time estimate and the interface discussion. It <br /> >sounds like the PostgreSQL SQL/MED code will be very useful when it <br /> > is done. Our product provides read-onlyaccess to files, so <br /> > updates/inserts/deletes aren't an issue for us.<br /> > <br /> > One thingthat is not clear to me is indexing support. Will it be <br /> > possible to index a SQL/MED table as if it werea regular table?<br /><br /> No, SQL/MED would not support indexing foreign tables, at least in<br /> first version. Because it would be difficult to use common row id for<br /> various FDWs. To support indexing foreign tables mightneed to change<br /> common structure of index tuple to be able to hold virtual row-id, not<br /> ItemPointerData.<br/><br /> Instead, FDW can handle expressions which are parsed from WHERE clause<br /> and JOIN conditionof original SQL, and use them to optimize scanning. <br /> For example, FDW for PostgreSQL pushes some conditionsdown to remote<br /> side to decrease result tuples to be transferred. I hope this idea<br /> helps you.<br /><br/> > What <br /> > would be the equivalent ofInformix's row ids?<br /><br /> Answer to the second question would be "ItemPointerData". It consists<br /> of a blocknumber and an offset in the block, and consume 6 bytes for<br /> each tuple. With this information, PostgreSQL canaccess to a data<br /> tuple directly. Actual definition is:<br /><br /> typedef struct ItemPointerData<br /> {<br /> BlockIdData ip_blkid;<br /> OffsetNumber ip_posid;<br /> } ItemPointer;<br /><br /> Does Informix uses common row-id(AFAIK it's 4 bytes integer) for<br /> both of virtual tables and normal tables?<br /><br /> Regards,<br /> --<br />Shigeru Hanada</blockquote><p> ********************************************** <br /> Eric Davies, M.Sc. <br /> Senior ProgrammerAnalyst<br /> Barrodale Computing Services Ltd. <br /> 1095 McKenzie Ave., Suite 418<br /> Victoria BC V8P 2L5<br/> Canada<br /><br /> Tel: (250) 704-4428<br /> Web: <a eudora="autourl" href="http://www.barrodale.com/"> <font color="#0000FF"><u>http://www.barrodale.com</u></font></a><br/> Email: eric@barrodale.com <br /> **********************************************<br/><br /><br />
Thanks for the information about Informix VTI. Because I'm not familiar to Informix, I might have missed your point. Would you mind telling me more about Informix VTI? On Mon, 15 Nov 2010 08:45:14 -0800 Eric Davies <eric@barrodale.com> wrote: > With Informix VTI, indexing is the same for native tables as for > virtual tables, except the interpretation of the 32 bit rowid is left > up to the developer. When you define the VTI class, you optionally > supply a method that can fetch data based on a 32 bit rowid, and it's > the responsibility of your non-indexed scanning methods to provide > rowids along with the row tuple. ISTM that index on a VTI table could be inconsistent when original (remote) data was changed in the way other than VTI. Is it assumed that the data source is never updated without VTI interface? > Having local indexes can be very useful if you have a user that > issues queries like: > select count(*) from some_external_table where .... ; > With VTI, the "count" aggregate doesn't get pushed down, meaning that > without a local index, your scanning method has to return as many > tuples as match the where clause, which can be very slow. How can Informix server optimize such kind of query? Counts the index tuple which match the WHERE clause? If so, such optimization seems to be limited to "count" and wouldn't be able to be useful for "max" or "sum". Or, specialized index or VTI class is responsible to the optimization? > Local indexes also affords the opportunity of using specialized > indexes built into the database. My guess is that without some form > of rowids being passed back and forth, you couldn't define > non-materialized views of virtual tables that could be indexed. > > That said, we implemented our own btree-like index that used the > pushed down predicates because fetching data one row at a time wasn't > desirable with our design choices, and we wanted to support virtual > tables with more than 4 billion rows. I couldn't see the way to handle virtual table with more than 4 billion rows with 32 bit rowids in local index. Do you mean that your "btree-like index" searches result rows by predicates directly and skips getbyid()? Regards, -- Shigeru Hanada
At 01:36 AM 11/16/2010, Shigeru HANADA wrote:<br /><blockquote cite="" class="cite" type="cite">Thanks for the informationabout Informix VTI. Because I'm not<br /> familiar to Informix, I might have missed your point. Would you mind<br/> telling me more about Informix VTI?</blockquote><br /><br /><blockquote cite="" class="cite" type="cite">On Mon,15 Nov 2010 08:45:14 -0800<br /> Eric Davies <eric@barrodale.com> wrote:<br /> > With Informix VTI, indexingis the same for native tables as for <br /> > virtual tables, except the interpretation of the 32 bit rowid isleft <br /> > up to the developer. When you define the VTI class, you optionally <br /> > supply a method that canfetch data based on a 32 bit rowid, and it's <br /> > the responsibility of your non-indexed scanning methods to provide<br /> > rowids along with the row tuple.<br /><br /> ISTM that index on a VTI table could be inconsistent whenoriginal<br /> (remote) data was changed in the way other than VTI. Is it assumed<br /> that the data source is neverupdated without VTI interface?</blockquote><br /> Yes, the data sources are assumed to updated only through the VTIinterface.<br /> With our UFI product, the data sources are assumed to be unchanging files, you'd need to re-index themif they changed.<br /><br /><br /><blockquote cite="" class="cite" type="cite">> Having local indexes can be veryuseful if you have a user that <br /> > issues queries like:<br /> > select count(*) from some_external_tablewhere .... ;<br /> > With VTI, the "count" aggregate doesn't get pushed down, meaning that <br /> >without a local index, your scanning method has to return as many <br /> > tuples as match the where clause, whichcan be very slow.<br /><br /> How can Informix server optimize such kind of query? Counts the index<br /> tuple whichmatch the WHERE clause? </blockquote><br /> That would be my assumption.<br /><br /><br /><blockquote cite="" class="cite"type="cite"> If so, such optimization seems to<br /> be limited to "count" and wouldn't be able to be usefulfor "max" or<br /> "sum". Or, specialized index or VTI class is responsible to the<br /> optimization?</blockquote><br/> If there is an index on the column you want to sum/min/max, and your where clause restrictsthe query to a particular set of rows based on that index, Informix can get the values for that column from theindex (which it needed to scan anyhow) without looking at the table. This isn't particular to VTI, it's just a cleveruse of indexes.<br /><br /> Here is a clipping from one of the Informix manuals on the topic: <dl><dd>The way thatthe optimizer chooses to read a table is called an <i>access plan</i>. The simplest method to access a table is to readit sequentially, which is called a <i>table scan</i>. The optimizer chooses a table scan when most of the table mustbe read or the table does not have an index that is useful for the query.<a name="idx2646"></a><dd><a name="idx2647"></a>Theoptimizer can also choose to access the table by an index. If the column in the index is the same asa column in a filter of the query, the optimizer can use the index to retrieve only the rows that the query requires. Theoptimizer can use a <i>key-only index scan</i><a name="idx2647"></a> if the columns requested are within one index onthe table. The database server retrieves the needed data from the index and does not access the associated table. <dd><aname="wq276"></a>Important: <dd><a name="idx2649"></a>The optimizer does not choose a key-only scan for a VARCHAR column.If you want to take advantage of key-only scans, use the ALTER TABLE with the MODFIY<a name="idx2649"></a> clauseto change the column to a CHAR data type.<a name="idx2655"></a><dd>The optimizer compares the cost of each plan todetermine the best one. The database server derives cost from estimates of the number of I/O operations required, calculationsto produce the results, rows accessed, sorting, and so forth.<br /><br /><br /><blockquote cite="" class="cite"type="cite"></blockquote></dl>> Local indexes also affords the opportunity of using specialized <br /> >indexes built into the database. My guess is that without some form <br /> > of rowids being passed back and forth,you couldn't define <br /> > non-materialized views of virtual tables that could be indexed.<br /> > <br /> >That said, we implemented our own btree-like index that used the <br /> > pushed down predicates because fetchingdata one row at a time wasn't <br /> > desirable with our design choices, and we wanted to support virtual <br/> > tables with more than 4 billion rows.<br /><br /> I couldn't see the way to handle virtual table with more than4<br /> billion rows with 32 bit rowids in local index. Do you mean that your<br /> "btree-like index" searches resultrows by predicates directly and<br /> skips getbyid()?<br /> Exactly. Our own "rowids" can be up to 64 bits but arenever seen by Informix. As far as Informix is concerned, it's a regular table scan because the use of our indexes is hidden.<br/><br /><br /><blockquote cite="" class="cite" type="cite">Regards,<br /> --<br /> Shigeru Hanada</blockquote><br/><br /> Cheers,<br /> Eric.<br /><br /><p> ********************************************** <br /> EricDavies, M.Sc. <br /> Senior Programmer Analyst<br /> Barrodale Computing Services Ltd. <br /> 1095 McKenzie Ave., Suite418<br /> Victoria BC V8P 2L5<br /> Canada<br /><br /> Tel: (250) 704-4428<br /> Web: <a eudora="autourl" href="http://www.barrodale.com/"><font color="#0000FF"><u>http://www.barrodale.com</u></font></a><br /> Email: eric@barrodale.com<br /> **********************************************<br /><br /><br />
Thanks for the additional information! On Tue, 16 Nov 2010 09:31:43 -0800 Eric Davies <eric@barrodale.com> wrote: > At 01:36 AM 11/16/2010, Shigeru HANADA wrote: > >On Mon, 15 Nov 2010 08:45:14 -0800 > >Eric Davies <eric@barrodale.com> wrote: > >ISTM that index on a VTI table could be inconsistent when original > >(remote) data was changed in the way other than VTI. Is it assumed > >that the data source is never updated without VTI interface? > > Yes, the data sources are assumed to updated only through the VTI interface. > With our UFI product, the data sources are assumed to be unchanging > files, you'd need to re-index them if they changed. ISTM that it would be difficult for SQL/MED to assume that external data is unchangeable. Also I'm not sure that index types which PostgreSQL has currently are useful for external data. So I think that indexes for external data would be better to be encapsulated into FDWs. > > If so, such optimization seems to > >be limited to "count" and wouldn't be able to be useful for "max" or > >"sum". Or, specialized index or VTI class is responsible to the > >optimization? > > If there is an index on the column you want to sum/min/max, and your > where clause restricts the query to a particular set of rows based on > that index, Informix can get the values for that column from the > index (which it needed to scan anyhow) without looking at the table. > This isn't particular to VTI, it's just a clever use of indexes. <snip> The optimization in Informix is specific to "key-only scan" but not to VTI, isn't it? Then, it would be better to implement such kind of optimization independently (separated from SQL/MED). I found "Index Only Scan" thread in pgsql-hackers, but I don't know its conclusion... > >I couldn't see the way to handle virtual table with more than 4 > >billion rows with 32 bit rowids in local index. Do you mean that your > >"btree-like index" searches result rows by predicates directly and > >skips getbyid()? > > Exactly. Our own "rowids" can be up to 64 bits but are never seen by > Informix. As far as Informix is concerned, it's a regular table scan > because the use of our indexes is hidden. The design of pushing down predicates which you've taken might be similar to the one in the FDW for PostgreSQL. FDW for PostgreSQL generates WHERE clause from predicates and restrict the tuples to be returned. I think that you can port your VTI class to SQL/MED as a FDW for indexed-files, without local index. Regards, -- Shigeru Hanada
Some random comments on the patch: ReleaseConnection is a very generic name for a global function, would be good to prefix it with "pgsqlfdw" or something. Same with any other globally visible functions. Please use the built-in contain_mutable_functions(Node *) instead of custom is_immutable_func(). Or at least func_volatile(Oid) Is it really a good idea to allow LOCK TABLE on foreign tables in its current form? It only locks the local foreign table object, not the table in the remote server. Sorry if this was fiercely discussed already, but I don't think the file FDW belongs in core. I'd rather see it as a contrib module I would've expected the contrib install script to create the foreign data wrapper for me. While you can specify options to a foreign data wrapper, the CREATE FOREIGN DATA WRAPPER seems similar to CREATE LANGUAGE, ie. something that happens when the foreign data wrapper library is installed. How do you specify a foreign table that has a different name in the remote server? For example, if I wanted to create a foreign table called "foo", that fetched rows from a remote table called "bar"? I would really like to see the SQL query that's shipped to the remote host in EXPLAIN. That's essential information for analyzing a query that involves a foreign table. What about transactions? Does the SQL/MED standard have something to say about that? In general, I'm surprised that there's no hook at all into the planning phase. You have this TODO comment postgresql_fdw: > /* > * TODO: omit (deparse to "NULL") columns which are not used in the > * original SQL. > * > * We must parse nodes parents of this ForeignScan node to determine unused > * columns because some columns may be used only in parent Sort/Agg/Limit > * nodes. > */ Parsing the parents of the ForeignScan node seems like a backwards way of solving the problem. The planner should tell the FDW what columns it needs. And there should be some way for the FDW to tell the planner which quals it can handle, so that the executor doesn't need to recheck them. You could make the planner interface infinitely complicated, but that's no excuse for doing nothing at all. The interface needs some thought... -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
Thanks for the comments. Attached patch includes fixes your comments marked (*), and other small fixes such as oid system column support by postgresql_fdw and file_fdw. On Fri, 19 Nov 2010 12:16:00 +0200 Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote: > ReleaseConnection is a very generic name for a global function, would be > good to prefix it with "pgsqlfdw" or something. Same with any other > globally visible functions. (*)Agreed, merged two files and make all functions other than postgresql_fdw_handler() private. Refactored name of public functions defined for file_fdw to have common prefix "FileState". > Please use the built-in contain_mutable_functions(Node *) instead of > custom is_immutable_func(). Or at least func_volatile(Oid) (*)I didn't know the function, thanks. Replaced custom is_immutable_func() with contain_mutable_functions(). > Is it really a good idea to allow LOCK TABLE on foreign tables in its > current form? It only locks the local foreign table object, not the > table in the remote server. The first reason to allow LOCK TABLE is to make pg_dump be able to export definition of foreign tables, and second is allow to lock normal table which has been inherited by foreign table(s). It would be able to allow FDWs to delegate lock requestto remote server with new hook in LockTableRecurse() or somewhere, but IMHO locking remote would be overkill because SQL/MED doesn't mention about lock. > Sorry if this was fiercely discussed already, but I don't think the file > FDW belongs in core. I'd rather see it as a contrib module It's in core from some passive reasons: - The file_fdw shared codes with COPY FROM heavily in first proposal. - Built-in FDW makes creating regression tests easier. Especially CREATE FOREIGN TABLE DDL requires FDW with valid handler. Moving to contrib would need adding "dummy FDW" or something which has valid handler to keep regression tests about DDL in core... > I would've expected the contrib install script to create the foreign > data wrapper for me. While you can specify options to a foreign data > wrapper, the CREATE FOREIGN DATA WRAPPER seems similar to CREATE > LANGUAGE, ie. something that happens when the foreign data wrapper > library is installed. (*)It seems to have been deleted by mistake, fixed. > How do you specify a foreign table that has a different name in the > remote server? For example, if I wanted to create a foreign table called > "foo", that fetched rows from a remote table called "bar"? You can specify name of schema, table and column with generic option of postgresql_fdw objects. object | option name | context --------+-------------+------------------------- schema | nspname | foreign table table | relname | foreign table column | colname | column of foreign table > I would really like to see the SQL query that's shipped to the remote > host in EXPLAIN. That's essential information for analyzing a query that > involves a foreign table. Me too :-) You can see the SQL if you set client_min_messages to debug1 or lower, but it's just debug message. New hook in ExplainNode, or new attribute of ForeignScan, would be necessary to show FWD-specific information in EXPLAIN result. ISTM this issue should be considered with the following issue about planner-hook because such information should be generated in planner phase. > What about transactions? Does the SQL/MED standard have something to say > about that? SQL/MED says nothing about transaction management. It's perhaps because SQL/MED allows only read access. > In general, I'm surprised that there's no hook at all into the planning > phase. You have this TODO comment postgresql_fdw: > > > /* > > * TODO: omit (deparse to "NULL") columns which are not used in the > > * original SQL. > > * > > * We must parse nodes parents of this ForeignScan node to determine unused > > * columns because some columns may be used only in parent Sort/Agg/Limit > > * nodes. > > */ > > Parsing the parents of the ForeignScan node seems like a backwards way > of solving the problem. The planner should tell the FDW what columns it > needs. And there should be some way for the FDW to tell the planner > which quals it can handle, so that the executor doesn't need to recheck > them. > > You could make the planner interface infinitely complicated, but that's > no excuse for doing nothing at all. The interface needs some thought... Optimization about column would be minor issue, rather might be removable. I'll research about foreign path/plan creation again to figure the issues about planning out. Regards, -- Shigeru Hanada
Вложения
On Fri, Nov 19, 2010 at 9:55 AM, Shigeru HANADA <hanada@metrosystems.co.jp> wrote: > [ new SQL/MED patch ] I can't help noticing that this patch adds 8,982 lines and removes 408, making it far larger any other patch I've ever seen on this list.And what that means is that committing all of thisin one go is going to be very, very difficult. Now, on the plus side, as 9000+ line patches go, this one looks pretty well-written, at least after ten seconds of looking at it, which is great as far as it goes, but the sheer size is still going to make it just about impossible for anyone to review it effectively and have real confidence that the whole thing is commit-quality. To have a chance of getting a significant portion of this into PostgreSQL 9.1, it really needs to be broken up into INDEPENDENTLY COMMITTABLE SUB-PATCHES. The key words here are "independently committable". Breaking up a patch into sub-patches by directory, for example, is completely useless - we're not, for example, going to commit the code first and the docs separately. Let me say that again - the ONLY useful way of breaking up a patch is to divide it into pieces such that EACH piece, by itself, would represent a credible commit. Each piece should be posted to a separate thread and a separate discussion should be had about the merits and demerits of each one. Each should have a separate CommitFest entry and, ideally, a separate reviewer. Of course, it may not be possible to fully evaluate a given patch without looking at the other ones, but the extent to which this is necessary should be minimized; otherwise you haven't really broken it up usefully. Ultimately, we probably want and need to get this patch down to chunks of less than 2000 lines each. But for starters, it looks quite simple to break this into three chunks: one for the PostgreSQL FDW, one for the CSV FDW, and one for the core functionality. I think that the CSV FDW, like the PG FDW, should be a loadable module. (I wonder if it would be more sensible to name all the FDWs as "fdw_foo" rather than "foo_fdw", so that they alphabetize together, but I believe that Tom has shot down similar suggestions in the past, so maybe it's not a good idea after all.) So let's do that and then see if we can find anything that we can either simplify (so it takes fewer lines of code) or pull out and commit separately (because, for example, it's some kind of refactoring that is a good idea independently of this patch). -- Robert Haas EnterpriseDB: http://www.enterprisedb.com The Enterprise PostgreSQL Company
On Mon, Nov 22, 2010 at 11:16, Robert Haas <robertmhaas@gmail.com> wrote: > To have a chance of getting a significant portion > of this into PostgreSQL 9.1, it really needs to be broken up into > INDEPENDENTLY COMMITTABLE SUB-PATCHES. Did we discuss about syntax-only patch is not acceptable because it makes the head broken state at the previous commit-fest? I think that's why the patch becomes so large. So, our guideline to submit a large patch would be: * Split patch into commitable sub-patches (2000 lines each), * But submita series of patches at once. Am I understanding correctly? -- Itagaki Takahiro
On Sun, Nov 21, 2010 at 10:14 PM, Itagaki Takahiro <itagaki.takahiro@gmail.com> wrote: > On Mon, Nov 22, 2010 at 11:16, Robert Haas <robertmhaas@gmail.com> wrote: >> To have a chance of getting a significant portion >> of this into PostgreSQL 9.1, it really needs to be broken up into >> INDEPENDENTLY COMMITTABLE SUB-PATCHES. > > Did we discuss about syntax-only patch is not acceptable because > it makes the head broken state at the previous commit-fest? > I think that's why the patch becomes so large. Right, I remember that discussion. Hopefully the distinction between that conversation and this one is clear. > So, our guideline to submit a large patch would be: > * Split patch into commitable sub-patches (2000 lines each), It's not a hard number - it's more important that the patch *make sense* than what the exact line count is. But I think that's a reasonable guideline to shoot for. Ideally, smaller still would probably be even better, but sometimes it just can't be done. Also, note that pulling off small chunks is a valuable way to make progress.For example, if we notice that there's a 100-line refactoringin the FDW patch that stands on its own, by all means let's pull it out and commit it. > * But submit a series of patches at once. When necessary, yes. Of course, the best thing is if you can make them truly independent and submit the one after another. Get one committed, move on to the next. But if you can't, then you can't. In this case, there's not much help for the fact that to decide whether the FDW patch is a good idea you're probably going to at least want to glance at the PGFDW and CSVFDW patches -- but it's possible we could decide to commit the core support first, and then work on getting the implementations committed afterwards, if we're confident that the basic design is all right but more work is needed down in the details. > Am I understanding correctly? I think so. -- Robert Haas EnterpriseDB: http://www.enterprisedb.com The Enterprise PostgreSQL Company
On Sun, 21 Nov 2010 21:16:05 -0500 Robert Haas <robertmhaas@gmail.com> wrote: <snip> > Ultimately, we probably want and need to get this patch down to chunks > of less than 2000 lines each. But for starters, it looks quite simple > to break this into three chunks: one for the PostgreSQL FDW, one for > the CSV FDW, and one for the core functionality. I think that the CSV > FDW, like the PG FDW, should be a loadable module. I've separated the patch into tree parts. They have codes, documents and tests within, and file_fdw and pgsql_fdw can be applied onto fdw_core for each, or together. I hope the separation helps the review of the patches. Contents of each patch are: fdw_core : DDL for FDW HANDLER and FOREIGN TABLE, ForeignScan file_fdw : FDW for file, as contrib (loadable) module pgsql_fdw: FDW for PG, as contrib module The fdw_core patch is still very large (it has 10336 lines), but about the half of the patch is unchanged lines. The brief of lines actually changed are: backend codes : 3000 regression tests : 500 sgml documents : 1500 Basic functionality is not changed, but when I separate file_fdw from core, I relaxed the requirement that FDW must have HANDLER to create FOREIGN TABLE. Now FDW is required to have valid Iterate function only when scanning (or EXPLAINing) the foreign table actually. It enables testing DDLs for FOREIGN TABLE with HANDLER-less-FDW. > (I wonder if it > would be more sensible to name all the FDWs as "fdw_foo" rather than > "foo_fdw", so that they alphabetize together, but I believe that Tom > has shot down similar suggestions in the past, so maybe it's not a > good idea after all.) I agree the naming you suggested, but it would need to change the name of existing function postgresql_fdw_validator to keep consistency. I couldn't find any discussion about the naming of fdw. Regards, -- Shigeru Hanada
Вложения
On 23.11.2010 12:09, Shigeru HANADA wrote: > On Sun, 21 Nov 2010 21:16:05 -0500 > Robert Haas<robertmhaas@gmail.com> wrote: > <snip> >> Ultimately, we probably want and need to get this patch down to chunks >> of less than 2000 lines each. But for starters, it looks quite simple >> to break this into three chunks: one for the PostgreSQL FDW, one for >> the CSV FDW, and one for the core functionality. I think that the CSV >> FDW, like the PG FDW, should be a loadable module. > > I've separated the patch into tree parts. They have codes, documents > and tests within, and file_fdw and pgsql_fdw can be applied onto > fdw_core for each, or together. I hope the separation helps the > review of the patches. The docs need some work. The CREATE FOREIGN TABLE reference page seems to be copy-pasted from CREATE TABLE, because it talks about constraints and WITH/WITHOUT OIDS which surely don't apply to foreign tables. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Tue, 23 Nov 2010 12:30:52 +0200 Heikki Linnakangas <heikki.linnakangas@enterprisedb.com> wrote: > The docs need some work. The CREATE FOREIGN TABLE reference page seems > to be copy-pasted from CREATE TABLE, because it talks about constraints > and WITH/WITHOUT OIDS which surely don't apply to foreign tables. Thanks for the comments. The page you pointed has been edited for foreign table. In current design, OID system column and CHECK constraints are supported. OID is supported to get oid from the source table (yes, it works only for postgresql_fdw but it seems useful to support). CHECK constraint is supported to enable constraint exclusion. In addition, TABLEOID is supported to show which table is the actual source. I agree that some kind of documents, such as "How to create new FDW", should be written. Regards, -- Shigeru Hanada
On 23.11.2010 14:22, Shigeru HANADA wrote: > On Tue, 23 Nov 2010 12:30:52 +0200 > Heikki Linnakangas<heikki.linnakangas@enterprisedb.com> wrote: >> The docs need some work. The CREATE FOREIGN TABLE reference page seems >> to be copy-pasted from CREATE TABLE, because it talks about constraints >> and WITH/WITHOUT OIDS which surely don't apply to foreign tables. > > Thanks for the comments. > > The page you pointed has been edited for foreign table. In current > design, OID system column and CHECK constraints are supported. Oh, ok. > OID is supported to get oid from the source table (yes, it works only > for postgresql_fdw but it seems useful to support). I don't think that's worthwhile. Oids on user tables is a legacy feature, not recommended for new applications. And if you have to access an existing table that uses oids, you can define a regular column for the oid: CREATE FOREIGN TABLE foreigntable (oid oid, data int4) SERVER myserver; > CHECK constraint > is supported to enable constraint exclusion. Hmm, my gut reaction is that that's a premature optimization. But what about DEFAULTs then, surely that doesn't make sense for a foreign table? > I agree that some kind of documents, such as "How to create new FDW", > should be written. A well-documented file FDW implementation goes a long way for that. But a chapter that explains SQL/MED, how to create foreign tables, servers, user mappings etc, and how they behave. That we need. -- Heikki Linnakangas EnterpriseDB http://www.enterprisedb.com
On Tue, 2010-11-23 at 20:18 +0200, Heikki Linnakangas wrote: > On 23.11.2010 14:22, Shigeru HANADA wrote: > > OID is supported to get oid from the source table (yes, it works only > > for postgresql_fdw but it seems useful to support). > > I don't think that's worthwhile. Oids on user tables is a legacy > feature, not recommended for new applications. Agreed. We should do everything we can to NOT encourage their use. JD -- PostgreSQL.org Major Contributor Command Prompt, Inc: http://www.commandprompt.com/ - 509.416.6579 Consulting, Training, Support, Custom Development, Engineering http://twitter.com/cmdpromptinc | http://identi.ca/commandprompt
On Tue, Nov 23, 2010 at 5:09 AM, Shigeru HANADA <hanada@metrosystems.co.jp> wrote: > I've separated the patch into tree parts. They have codes, documents > and tests within, and file_fdw and pgsql_fdw can be applied onto > fdw_core for each, or together. I hope the separation helps the > review of the patches. Contents of each patch are: > > fdw_core : DDL for FDW HANDLER and FOREIGN TABLE, ForeignScan > file_fdw : FDW for file, as contrib (loadable) module > pgsql_fdw: FDW for PG, as contrib module I think it would be useful to repost each of these on their own thread and adjust the CF app to have an entry referring to each thread. That way we can keep discussion and review for each one separate from the others. Note that this thread is already totally off-topic anyway - it started as a question about when SQL/MED would land and then it got patches posted to it. -- Robert Haas EnterpriseDB: http://www.enterprisedb.com The Enterprise PostgreSQL Company
On Tue, 23 Nov 2010 10:22:02 -0800 "Joshua D. Drake" <jd@commandprompt.com> wrote: > On Tue, 2010-11-23 at 20:18 +0200, Heikki Linnakangas wrote: > > On 23.11.2010 14:22, Shigeru HANADA wrote: > > > > OID is supported to get oid from the source table (yes, it works only > > > for postgresql_fdw but it seems useful to support). > > > > I don't think that's worthwhile. Oids on user tables is a legacy > > feature, not recommended for new applications. > > Agreed. We should do everything we can to NOT encourage their use. Agreed. I'll remove OIDs support and repost the patch in new thread. Regards, -- Shigeru Hanada