*** a/doc/src/sgml/mvcc.sgml
--- b/doc/src/sgml/mvcc.sgml
***************
*** 386,407 **** COMMIT;
behave the same as SELECT
in terms of searching for target rows: they will only find target rows
that were committed as of the transaction start time. However, such a
! target
! row might have already been updated (or deleted or locked) by
! another concurrent transaction by the time it is found. In this case, the
! serializable transaction will wait for the first updating transaction to commit or
! roll back (if it is still in progress). If the first updater rolls back,
! then its effects are negated and the serializable transaction can proceed
! with updating the originally found row. But if the first updater commits
! (and actually updated or deleted the row, not just locked it)
! then the serializable transaction will be rolled back with the message
ERROR: could not serialize access due to concurrent update
! because a serializable transaction cannot modify or lock rows changed by
! other transactions after the serializable transaction began.
--- 386,407 ----
behave the same as SELECT
in terms of searching for target rows: they will only find target rows
that were committed as of the transaction start time. However, such a
! target row might have already been subject to a concurrent
! UPDATE, DELETE, SELECT
! FOR UPDATE, or SELECT FOR SHARE. In this case,
! the serializable transaction will wait for the other transaction to commit
! or roll back (if it is still in progress). If it rolls back then its effects
! are negated and the serializable transaction can proceed with modifying
! or locking the originally found row. If it commits, and the two commands
! conflict according to ,
! the serializable transaction is rolled back with the message
ERROR: could not serialize access due to concurrent update
! since serializable transaction cannot simply proceed with the newer row
! version like read committed ones do.
***************
*** 418,423 **** ERROR: could not serialize access due to concurrent update
--- 418,463 ----
transactions will never have serialization conflicts.
+
+ Serialization Conflicts
+
+
+
+
+
+
+ Serializable Transaction
+ Concurrent Transaction
+
+
+ UPDATE, DELETE
+ SELECT FOR UPDATE
+ SELECT FOR SHARE
+
+
+
+
+ UPDATE, DELETE
+ X
+ X
+ X
+
+
+ SELECT FOR UPDATE
+ X
+ X
+ X
+
+
+ SELECT FOR SHARE
+ X
+
+
+
+
+
+
+
The Serializable mode provides a rigorous guarantee that each
transaction sees a wholly consistent view of the database. However,
***************
*** 921,926 **** SELECT SUM(value) FROM mytab WHERE class = 2;
--- 961,974 ----
+ Serializable transactions are affected by concurrent
+ SELECT FOR SHARE and SELECT FOR UPDATE
+ for longer than those locks are actually held, and may be aborted
+ when trying to obtain a conflicting lock. For details,
+ see
+
+
+
PostgreSQL doesn't remember any
information about modified rows in memory, so there is no limit on
the number of rows locked at one time. However, locking a row
*** a/src/backend/access/heap/heapam.c
--- b/src/backend/access/heap/heapam.c
***************
*** 83,88 **** static XLogRecPtr log_heap_update(Relation reln, Buffer oldbuf,
--- 83,89 ----
bool all_visible_cleared, bool new_all_visible_cleared);
static bool HeapSatisfiesHOTUpdate(Relation relation, Bitmapset *hot_attrs,
HeapTuple oldtup, HeapTuple newtup);
+ static bool HeapSatisfiesLockersVisible(HeapTupleHeader tuple, Snapshot snapshot);
/* ----------------------------------------------------------------
***************
*** 2033,2040 **** simple_heap_insert(Relation relation, HeapTuple tup)
* update_xmax - output parameter, used only for failure case (see below)
* cid - delete command ID (used for visibility test, and stored into
* cmax if successful)
- * crosscheck - if not InvalidSnapshot, also check tuple against this
* wait - true if should wait for any conflicting update to commit/abort
*
* Normal, successful return value is HeapTupleMayBeUpdated, which
* actually means we did delete it. Failure return codes are
--- 2034,2042 ----
* update_xmax - output parameter, used only for failure case (see below)
* cid - delete command ID (used for visibility test, and stored into
* cmax if successful)
* wait - true if should wait for any conflicting update to commit/abort
+ * lockcheck_snapshot - if not NULL, report the tuple as updated if it
+ * was locked by a transaction not visible under this snapshot
*
* Normal, successful return value is HeapTupleMayBeUpdated, which
* actually means we did delete it. Failure return codes are
***************
*** 2049,2055 **** simple_heap_insert(Relation relation, HeapTuple tup)
HTSU_Result
heap_delete(Relation relation, ItemPointer tid,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, Snapshot crosscheck, bool wait)
{
HTSU_Result result;
TransactionId xid = GetCurrentTransactionId();
--- 2051,2057 ----
HTSU_Result
heap_delete(Relation relation, ItemPointer tid,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, bool wait, Snapshot lockcheck_snapshot)
{
HTSU_Result result;
TransactionId xid = GetCurrentTransactionId();
***************
*** 2171,2189 **** l1:
result = HeapTupleUpdated;
}
! if (crosscheck != InvalidSnapshot && result == HeapTupleMayBeUpdated)
! {
! /* Perform additional check for transaction-snapshot mode RI updates */
! if (!HeapTupleSatisfiesVisibility(&tp, crosscheck, buffer))
! result = HeapTupleUpdated;
! }
if (result != HeapTupleMayBeUpdated)
{
Assert(result == HeapTupleSelfUpdated ||
result == HeapTupleUpdated ||
result == HeapTupleBeingUpdated);
! Assert(!(tp.t_data->t_infomask & HEAP_XMAX_INVALID));
*ctid = tp.t_data->t_ctid;
*update_xmax = HeapTupleHeaderGetXmax(tp.t_data);
UnlockReleaseBuffer(buffer);
--- 2173,2190 ----
result = HeapTupleUpdated;
}
! /* Verify visibility of locking transactions. */
! if ((result == HeapTupleMayBeUpdated) &&
! !HeapSatisfiesLockersVisible(tp.t_data, lockcheck_snapshot))
! result = HeapTupleUpdated;
if (result != HeapTupleMayBeUpdated)
{
Assert(result == HeapTupleSelfUpdated ||
result == HeapTupleUpdated ||
result == HeapTupleBeingUpdated);
! Assert(!(tp.t_data->t_infomask & HEAP_XMAX_INVALID) ||
! (tp.t_data->t_infomask & HEAP_IS_LOCKED));
*ctid = tp.t_data->t_ctid;
*update_xmax = HeapTupleHeaderGetXmax(tp.t_data);
UnlockReleaseBuffer(buffer);
***************
*** 2313,2320 **** simple_heap_delete(Relation relation, ItemPointer tid)
result = heap_delete(relation, tid,
&update_ctid, &update_xmax,
! GetCurrentCommandId(true), InvalidSnapshot,
! true /* wait for commit */ );
switch (result)
{
case HeapTupleSelfUpdated:
--- 2314,2322 ----
result = heap_delete(relation, tid,
&update_ctid, &update_xmax,
! GetCurrentCommandId(true),
! true /* wait for commit */ ,
! InvalidSnapshot);
switch (result)
{
case HeapTupleSelfUpdated:
***************
*** 2349,2355 **** simple_heap_delete(Relation relation, ItemPointer tid)
* update_xmax - output parameter, used only for failure case (see below)
* cid - update command ID (used for visibility test, and stored into
* cmax/cmin if successful)
! * crosscheck - if not InvalidSnapshot, also check old tuple against this
* wait - true if should wait for any conflicting update to commit/abort
*
* Normal, successful return value is HeapTupleMayBeUpdated, which
--- 2351,2358 ----
* update_xmax - output parameter, used only for failure case (see below)
* cid - update command ID (used for visibility test, and stored into
* cmax/cmin if successful)
! * lockcheck_snapshot - if not NULL, report the tuple as updated if it
! * was locked by a transaction not visible under this snapshot
* wait - true if should wait for any conflicting update to commit/abort
*
* Normal, successful return value is HeapTupleMayBeUpdated, which
***************
*** 2371,2377 **** simple_heap_delete(Relation relation, ItemPointer tid)
HTSU_Result
heap_update(Relation relation, ItemPointer otid, HeapTuple newtup,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, Snapshot crosscheck, bool wait)
{
HTSU_Result result;
TransactionId xid = GetCurrentTransactionId();
--- 2374,2380 ----
HTSU_Result
heap_update(Relation relation, ItemPointer otid, HeapTuple newtup,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, bool wait, Snapshot lockcheck_snapshot)
{
HTSU_Result result;
TransactionId xid = GetCurrentTransactionId();
***************
*** 2523,2541 **** l2:
result = HeapTupleUpdated;
}
! if (crosscheck != InvalidSnapshot && result == HeapTupleMayBeUpdated)
! {
! /* Perform additional check for transaction-snapshot mode RI updates */
! if (!HeapTupleSatisfiesVisibility(&oldtup, crosscheck, buffer))
! result = HeapTupleUpdated;
! }
if (result != HeapTupleMayBeUpdated)
{
Assert(result == HeapTupleSelfUpdated ||
result == HeapTupleUpdated ||
result == HeapTupleBeingUpdated);
! Assert(!(oldtup.t_data->t_infomask & HEAP_XMAX_INVALID));
*ctid = oldtup.t_data->t_ctid;
*update_xmax = HeapTupleHeaderGetXmax(oldtup.t_data);
UnlockReleaseBuffer(buffer);
--- 2526,2543 ----
result = HeapTupleUpdated;
}
! /* Verify visibility of locking transactions. */
! if ((result == HeapTupleMayBeUpdated) &&
! !HeapSatisfiesLockersVisible(oldtup.t_data, lockcheck_snapshot))
! result = HeapTupleUpdated;
if (result != HeapTupleMayBeUpdated)
{
Assert(result == HeapTupleSelfUpdated ||
result == HeapTupleUpdated ||
result == HeapTupleBeingUpdated);
! Assert(!(oldtup.t_data->t_infomask & HEAP_XMAX_INVALID) ||
! (oldtup.t_data->t_infomask & HEAP_IS_LOCKED));
*ctid = oldtup.t_data->t_ctid;
*update_xmax = HeapTupleHeaderGetXmax(oldtup.t_data);
UnlockReleaseBuffer(buffer);
***************
*** 2961,2966 **** HeapSatisfiesHOTUpdate(Relation relation, Bitmapset *hot_attrs,
--- 2963,3063 ----
return true;
}
+
+ /*
+ * Returns false if one of the tuple's lockers committed but aren't visible
+ * according to lockcheck_snapshot, excluding subtransactions of the current
+ * transaction. Assumes that all locking transaction either committed or
+ * aborted, but aren't still in progress.
+ */
+ static bool
+ HeapSatisfiesLockersVisible(HeapTupleHeader tuple,
+ Snapshot lockcheck_snapshot)
+ {
+ if (lockcheck_snapshot == InvalidSnapshot)
+ return true;
+
+ if (tuple->t_infomask & HEAP_IS_LOCKED)
+ {
+ /*
+ * If the tuple was locked, we now check whether the locking
+ * transaction(s) are visible under lockcheck_snapshot. If
+ * they aren't, we pretend that the tuple was updated.
+ */
+
+ if (tuple->t_infomask & HEAP_XMAX_IS_MULTI)
+ {
+ TransactionId* xids;
+ int xids_l = GetMultiXactIdMembers(HeapTupleHeaderGetXmax(tuple),
+ &xids);
+
+ if (xids_l < 1)
+ {
+ /*
+ * The multi xact either is too old to be inspected or
+ * doesn't contain members. The second case is probably
+ * impossible, but even if not it doesn't pose any problem.
+ *
+ * In the first case, we have to trust that all xids that
+ * were contained in the xact are in fact visible under
+ * lockcheck_snapshot. Currently this is always the case,
+ * since lockcheck_snapshot is always the transaction's
+ * snapshot, and we call MultiXactIdSetOldestVisible() before
+ * acquiring that snapshot.
+ */
+ return true;
+ }
+ else
+ {
+ int i;
+ for (i = 0; i < xids_l; ++i)
+ {
+ /* Ignore our own subtransactions */
+ if (TransactionIdIsCurrentTransactionId(xids[i]))
+ continue;
+
+ /*
+ * We expect to be called after the locking transactions'
+ * fates have been decided
+ */
+ Assert(!TransactionIdIsInProgress(xids[i]));
+
+ if (!TransactionIdDidAbort(xids[i]) &&
+ XidInMVCCSnapshot(xids[i], lockcheck_snapshot))
+ {
+ /* Non-aborted, invisible locker */
+ return false;
+ }
+ }
+ return true;
+ }
+ }
+ else
+ {
+ TransactionId xid = HeapTupleHeaderGetXmax(tuple);
+
+ /* Ignore our own subtransactions */
+ if (TransactionIdIsCurrentTransactionId(xid))
+ return true;
+
+ /*
+ * We expect to be called after the locking transactions' fates
+ * have been decided
+ */
+ Assert(!TransactionIdIsInProgress(xid));
+
+ /* Locker must either be visible or have aborted */
+ return TransactionIdDidAbort(xid) ||
+ !XidInMVCCSnapshot(xid, lockcheck_snapshot);
+ }
+ }
+ else
+ {
+ /* Tuple wasn't locked */
+ return true;
+ }
+ }
+
/*
* simple_heap_update - replace a tuple
*
***************
*** 2978,2985 **** simple_heap_update(Relation relation, ItemPointer otid, HeapTuple tup)
result = heap_update(relation, otid, tup,
&update_ctid, &update_xmax,
! GetCurrentCommandId(true), InvalidSnapshot,
! true /* wait for commit */ );
switch (result)
{
case HeapTupleSelfUpdated:
--- 3075,3082 ----
result = heap_update(relation, otid, tup,
&update_ctid, &update_xmax,
! GetCurrentCommandId(true),
! true /* wait for commit */, InvalidSnapshot);
switch (result)
{
case HeapTupleSelfUpdated:
***************
*** 3013,3018 **** simple_heap_update(Relation relation, ItemPointer otid, HeapTuple tup)
--- 3110,3118 ----
* tuple's cmax if lock is successful)
* mode: indicates if shared or exclusive tuple lock is desired
* nowait: if true, ereport rather than blocking if lock not available
+ * lockcheck_snapshot: if not NULL, report the tuple as updated if it
+ * was locked by a transaction not visible under
+ * this snapshot
*
* Output parameters:
* *tuple: all fields filled in
***************
*** 3066,3072 **** simple_heap_update(Relation relation, ItemPointer otid, HeapTuple tup)
HTSU_Result
heap_lock_tuple(Relation relation, HeapTuple tuple, Buffer *buffer,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, LockTupleMode mode, bool nowait)
{
HTSU_Result result;
ItemPointer tid = &(tuple->t_self);
--- 3166,3173 ----
HTSU_Result
heap_lock_tuple(Relation relation, HeapTuple tuple, Buffer *buffer,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, LockTupleMode mode, bool nowait,
! Snapshot lockcheck_snapshot)
{
HTSU_Result result;
ItemPointer tid = &(tuple->t_self);
***************
*** 3247,3256 **** l3:
result = HeapTupleUpdated;
}
if (result != HeapTupleMayBeUpdated)
{
Assert(result == HeapTupleSelfUpdated || result == HeapTupleUpdated);
! Assert(!(tuple->t_data->t_infomask & HEAP_XMAX_INVALID));
*ctid = tuple->t_data->t_ctid;
*update_xmax = HeapTupleHeaderGetXmax(tuple->t_data);
LockBuffer(*buffer, BUFFER_LOCK_UNLOCK);
--- 3348,3363 ----
result = HeapTupleUpdated;
}
+ /* Verify visibility of locking transactions */
+ if ((result == HeapTupleMayBeUpdated) &&
+ !HeapSatisfiesLockersVisible(tuple->t_data, lockcheck_snapshot))
+ result = HeapTupleUpdated;
+
if (result != HeapTupleMayBeUpdated)
{
Assert(result == HeapTupleSelfUpdated || result == HeapTupleUpdated);
! Assert(!(tuple->t_data->t_infomask & HEAP_XMAX_INVALID) ||
! (tuple->t_data->t_infomask & HEAP_IS_LOCKED));
*ctid = tuple->t_data->t_ctid;
*update_xmax = HeapTupleHeaderGetXmax(tuple->t_data);
LockBuffer(*buffer, BUFFER_LOCK_UNLOCK);
*** a/src/backend/access/transam/multixact.c
--- b/src/backend/access/transam/multixact.c
***************
*** 211,217 **** static MemoryContext MXactContext = NULL;
#endif
/* internal MultiXactId management */
- static void MultiXactIdSetOldestVisible(void);
static MultiXactId CreateMultiXactId(int nxids, TransactionId *xids);
static void RecordNewMultiXact(MultiXactId multi, MultiXactOffset offset,
int nxids, TransactionId *xids);
--- 211,216 ----
***************
*** 531,537 **** MultiXactIdSetOldestMember(void)
* there is no live transaction, now or later, that can be a member of any
* MultiXactId older than the OldestVisibleMXactId we compute here.
*/
! static void
MultiXactIdSetOldestVisible(void)
{
if (!MultiXactIdIsValid(OldestVisibleMXactId[MyBackendId]))
--- 530,536 ----
* there is no live transaction, now or later, that can be a member of any
* MultiXactId older than the OldestVisibleMXactId we compute here.
*/
! void
MultiXactIdSetOldestVisible(void)
{
if (!MultiXactIdIsValid(OldestVisibleMXactId[MyBackendId]))
*** a/src/backend/commands/copy.c
--- b/src/backend/commands/copy.c
***************
*** 1088,1094 **** DoCopy(const CopyStmt *stmt, const char *queryString)
/* Create a QueryDesc requesting no output */
cstate->queryDesc = CreateQueryDesc(plan, queryString,
GetActiveSnapshot(),
- InvalidSnapshot,
dest, NULL, 0);
/*
--- 1088,1093 ----
*** a/src/backend/commands/explain.c
--- b/src/backend/commands/explain.c
***************
*** 372,378 **** ExplainOnePlan(PlannedStmt *plannedstmt, ExplainState *es,
/* Create a QueryDesc requesting no output */
queryDesc = CreateQueryDesc(plannedstmt, queryString,
! GetActiveSnapshot(), InvalidSnapshot,
None_Receiver, params, instrument_option);
INSTR_TIME_SET_CURRENT(starttime);
--- 372,378 ----
/* Create a QueryDesc requesting no output */
queryDesc = CreateQueryDesc(plannedstmt, queryString,
! GetActiveSnapshot(),
None_Receiver, params, instrument_option);
INSTR_TIME_SET_CURRENT(starttime);
*** a/src/backend/commands/trigger.c
--- b/src/backend/commands/trigger.c
***************
*** 2367,2380 **** GetTupleForTrigger(EState *estate,
Assert(epqstate != NULL);
/*
! * lock tuple for update
*/
ltrmark:;
tuple.t_self = *tid;
test = heap_lock_tuple(relation, &tuple, &buffer,
&update_ctid, &update_xmax,
estate->es_output_cid,
! LockTupleExclusive, false);
switch (test)
{
case HeapTupleSelfUpdated:
--- 2367,2386 ----
Assert(epqstate != NULL);
/*
! * Lock tuple for update.
! *
! * Transaction snapshot mode transactions pass their snapshot as the
! * logcheck_snapshot. This lets heap_lock_tuple report concurrently
! * FOR SHARE or FOR UPDATE locked tuples as HeapTupleUpdated.
*/
ltrmark:;
tuple.t_self = *tid;
test = heap_lock_tuple(relation, &tuple, &buffer,
&update_ctid, &update_xmax,
estate->es_output_cid,
! LockTupleExclusive, false,
! IsolationUsesXactSnapshot() ? estate->es_snapshot :
! InvalidSnapshot);
switch (test)
{
case HeapTupleSelfUpdated:
*** a/src/backend/executor/README
--- b/src/backend/executor/README
***************
*** 139,157 **** be hidden inside function calls). This case has a flow of control like
(a separate FreeExprContext call is not necessary)
! EvalPlanQual (READ COMMITTED Update Checking)
! ---------------------------------------------
For simple SELECTs, the executor need only pay attention to tuples that are
valid according to the snapshot seen by the current transaction (ie, they
were inserted by a previously committed transaction, and not deleted by any
previously committed transaction). However, for UPDATE and DELETE it is not
cool to modify or delete a tuple that's been modified by an open or
! concurrently-committed transaction. If we are running in SERIALIZABLE
! isolation level then we just raise an error when this condition is seen to
! occur. In READ COMMITTED isolation level, we must work a lot harder.
! The basic idea in READ COMMITTED mode is to take the modified tuple
committed by the concurrent transaction (after waiting for it to commit,
if need be) and re-evaluate the query qualifications to see if it would
still meet the quals. If so, we regenerate the updated tuple (if we are
--- 139,158 ----
(a separate FreeExprContext call is not necessary)
! EvalPlanQual (Statement-Snapshot Update Checking)
! -------------------------------------------------
For simple SELECTs, the executor need only pay attention to tuples that are
valid according to the snapshot seen by the current transaction (ie, they
were inserted by a previously committed transaction, and not deleted by any
previously committed transaction). However, for UPDATE and DELETE it is not
cool to modify or delete a tuple that's been modified by an open or
! concurrently-committed transaction. If we are running in a transaction-
! snapshot isolation level then we just raise an error when this condition is
! seen to occur. In statement-snapshot isolation levels, we must work a lot
! harder.
! The basic idea in statement-snapshot modes is to take the modified tuple
committed by the concurrent transaction (after waiting for it to commit,
if need be) and re-evaluate the query qualifications to see if it would
still meet the quals. If so, we regenerate the updated tuple (if we are
***************
*** 195,197 **** is no explicit prohibition on SRFs in UPDATE, but the net effect will be
--- 196,227 ----
that only the first result row of an SRF counts, because all subsequent
rows will result in attempts to re-update an already updated target row.
This is historical behavior and seems not worth changing.)
+
+ Row Locks and Transaction-Snapshot Mode Transactions
+ ----------------------------------------------------
+
+ In a statement-snapshot mode, a transaction which encounters a locked row
+ during an UPDATE, DELETE, SELECT FOR UPDATE or SELECT FOR SHARE simply blocks
+ until the locking transaction commits or roll backs, and in the former case
+ then re-executes the statement using the new row version, as described above.
+
+ For transaction-snapshot mode transactions this is not satisfactory. The RI
+ triggers, for example, take a FOR SHARE lock on a parent row before allowing
+ a child row to be inserted and verify that deleting a parent row leaves no
+ orphaned children behind before allowing the delete to occur. From within
+ statement-snapshot mode transactions, blocking upon a delete or a parent row
+ until all lockers have finished is sufficient to guarantee that this check
+ finds any potential orphan, since the check will be executed with a up-to-
+ date snapshot to which the locking transaction's changes are visible. This,
+ however, is not true for transaction-snapshot mode transactions since these
+ will continue to use their old snapshot and hence miss newly inserted rows.
+
+ Transaction-snapshot mode transactions therefore treat a FOR SHARE or FOR
+ UPDATE lock on a tuple the same as an actual update during UPDATE and SELECT
+ FOR SHARE. They are thus aborted when trying to UPDATE or FOR UPDATE lock a
+ row that was FOR SHARE or FOR UPDATE locked by a concurrent transaction.
+
+ This is implemented by the lockcheck_snapshot parameter of heap_update,
+ heap_delete and heap_lock_tuple. If such a snapshot is provided to one of
+ these functions, they return HeapTupleUpdated if the tuple locked (but not
+ necessarily updated) by any transaction invisible to the snapshot.
*** a/src/backend/executor/execMain.c
--- b/src/backend/executor/execMain.c
***************
*** 183,189 **** standard_ExecutorStart(QueryDesc *queryDesc, int eflags)
* Copy other important information into the EState
*/
estate->es_snapshot = RegisterSnapshot(queryDesc->snapshot);
- estate->es_crosscheck_snapshot = RegisterSnapshot(queryDesc->crosscheck_snapshot);
estate->es_instrument = queryDesc->instrument_options;
/*
--- 183,188 ----
***************
*** 348,354 **** standard_ExecutorEnd(QueryDesc *queryDesc)
/* do away with our snapshots */
UnregisterSnapshot(estate->es_snapshot);
- UnregisterSnapshot(estate->es_crosscheck_snapshot);
/*
* Must switch out of context before destroying it
--- 347,352 ----
***************
*** 1533,1543 **** EvalPlanQualFetch(EState *estate, Relation relation, int lockmode,
/*
* This is a live tuple, so now try to lock it.
*/
test = heap_lock_tuple(relation, &tuple, &buffer,
&update_ctid, &update_xmax,
estate->es_output_cid,
! lockmode, false);
/* We now have two pins on the buffer, get rid of one */
ReleaseBuffer(buffer);
--- 1531,1549 ----
/*
* This is a live tuple, so now try to lock it.
+ *
+ * Transaction-snapshot mode transactions pass their snapshot as
+ * the logcheck_snapshot. This lets heap_lock_tuple report
+ * concurrently FOR SHARE or FOR UPDATE locked tuples as
+ * HeapTupleUpdated.
*/
+ Assert(!IsolationUsesXactSnapshot() || (estate->es_snapshot != InvalidSnapshot));
test = heap_lock_tuple(relation, &tuple, &buffer,
&update_ctid, &update_xmax,
estate->es_output_cid,
! lockmode, false,
! IsolationUsesXactSnapshot() ? estate->es_snapshot :
! InvalidSnapshot);
/* We now have two pins on the buffer, get rid of one */
ReleaseBuffer(buffer);
***************
*** 1906,1912 **** EvalPlanQualStart(EPQState *epqstate, EState *parentestate, Plan *planTree)
*/
estate->es_direction = ForwardScanDirection;
estate->es_snapshot = parentestate->es_snapshot;
- estate->es_crosscheck_snapshot = parentestate->es_crosscheck_snapshot;
estate->es_range_table = parentestate->es_range_table;
estate->es_plannedstmt = parentestate->es_plannedstmt;
estate->es_junkFilter = parentestate->es_junkFilter;
--- 1912,1917 ----
*** a/src/backend/executor/execUtils.c
--- b/src/backend/executor/execUtils.c
***************
*** 109,115 **** CreateExecutorState(void)
*/
estate->es_direction = ForwardScanDirection;
estate->es_snapshot = SnapshotNow;
- estate->es_crosscheck_snapshot = InvalidSnapshot; /* no crosscheck */
estate->es_range_table = NIL;
estate->es_plannedstmt = NULL;
--- 109,114 ----
*** a/src/backend/executor/functions.c
--- b/src/backend/executor/functions.c
***************
*** 415,421 **** postquel_start(execution_state *es, SQLFunctionCachePtr fcache)
if (IsA(es->stmt, PlannedStmt))
es->qd = CreateQueryDesc((PlannedStmt *) es->stmt,
fcache->src,
! snapshot, InvalidSnapshot,
dest,
fcache->paramLI, 0);
else
--- 415,421 ----
if (IsA(es->stmt, PlannedStmt))
es->qd = CreateQueryDesc((PlannedStmt *) es->stmt,
fcache->src,
! snapshot,
dest,
fcache->paramLI, 0);
else
*** a/src/backend/executor/nodeLockRows.c
--- b/src/backend/executor/nodeLockRows.c
***************
*** 71,76 **** lnext:
--- 71,77 ----
ItemPointerData update_ctid;
TransactionId update_xmax;
LockTupleMode lockmode;
+ Snapshot lockcheck_snapshot = InvalidSnapshot;
HTSU_Result test;
HeapTuple copyTuple;
***************
*** 110,123 **** lnext:
/* okay, try to lock the tuple */
if (erm->markType == ROW_MARK_EXCLUSIVE)
lockmode = LockTupleExclusive;
else
lockmode = LockTupleShared;
test = heap_lock_tuple(erm->relation, &tuple, &buffer,
&update_ctid, &update_xmax,
estate->es_output_cid,
! lockmode, erm->noWait);
ReleaseBuffer(buffer);
switch (test)
{
--- 111,136 ----
/* okay, try to lock the tuple */
if (erm->markType == ROW_MARK_EXCLUSIVE)
+ {
lockmode = LockTupleExclusive;
+
+ /*
+ * Transaction-snapshot mode transactions pass their snapshot as
+ * the logcheck_snapshot. This lets heap_lock_tuple report
+ * concurrently FOR SHARE or FOR UPDATE locked tuples as
+ * HeapTupleUpdated.
+ */
+ if (IsolationUsesXactSnapshot())
+ lockcheck_snapshot = estate->es_snapshot;
+ }
else
lockmode = LockTupleShared;
test = heap_lock_tuple(erm->relation, &tuple, &buffer,
&update_ctid, &update_xmax,
estate->es_output_cid,
! lockmode, erm->noWait,
! lockcheck_snapshot);
ReleaseBuffer(buffer);
switch (test)
{
*** a/src/backend/executor/nodeModifyTable.c
--- b/src/backend/executor/nodeModifyTable.c
***************
*** 307,323 **** ExecDelete(ItemPointer tupleid,
/*
* delete the tuple
*
! * Note: if es_crosscheck_snapshot isn't InvalidSnapshot, we check that
! * the row to be deleted is visible to that snapshot, and throw a can't-
! * serialize error if not. This is a special-case behavior needed for
! * referential integrity updates in transaction-snapshot mode transactions.
*/
ldelete:;
result = heap_delete(resultRelationDesc, tupleid,
&update_ctid, &update_xmax,
estate->es_output_cid,
! estate->es_crosscheck_snapshot,
! true /* wait for commit */ );
switch (result)
{
case HeapTupleSelfUpdated:
--- 307,323 ----
/*
* delete the tuple
*
! * Transaction-snapshot mode transactions pass their snapshot as the
! * logcheck_snapshot. This lets heap_lock_tuple report concurrently FOR
! * SHARE or FOR UPDATE locked tuples as HeapTupleUpdated.
*/
ldelete:;
result = heap_delete(resultRelationDesc, tupleid,
&update_ctid, &update_xmax,
estate->es_output_cid,
! true, /* wait for commit */
! IsolationUsesXactSnapshot() ? estate->es_snapshot :
! InvalidSnapshot);
switch (result)
{
case HeapTupleSelfUpdated:
***************
*** 496,511 **** lreplace:;
/*
* replace the heap tuple
*
! * Note: if es_crosscheck_snapshot isn't InvalidSnapshot, we check that
! * the row to be updated is visible to that snapshot, and throw a can't-
! * serialize error if not. This is a special-case behavior needed for
! * referential integrity updates in transaction-snapshot mode transactions.
*/
result = heap_update(resultRelationDesc, tupleid, tuple,
&update_ctid, &update_xmax,
estate->es_output_cid,
! estate->es_crosscheck_snapshot,
! true /* wait for commit */ );
switch (result)
{
case HeapTupleSelfUpdated:
--- 496,511 ----
/*
* replace the heap tuple
*
! * Transaction-snapshot mode transactions pass their snapshot as the
! * logcheck_snapshot. This lets heap_lock_tuple report concurrently FOR
! * SHARE or FOR UPDATE locked tuples as HeapTupleUpdated.
*/
result = heap_update(resultRelationDesc, tupleid, tuple,
&update_ctid, &update_xmax,
estate->es_output_cid,
! true, /* wait for commit */
! IsolationUsesXactSnapshot() ? estate->es_snapshot :
! InvalidSnapshot);
switch (result)
{
case HeapTupleSelfUpdated:
*** a/src/backend/executor/spi.c
--- b/src/backend/executor/spi.c
***************
*** 51,57 **** static void _SPI_prepare_plan(const char *src, SPIPlanPtr plan,
ParamListInfo boundParams);
static int _SPI_execute_plan(SPIPlanPtr plan, ParamListInfo paramLI,
! Snapshot snapshot, Snapshot crosscheck_snapshot,
bool read_only, bool fire_triggers, long tcount);
static ParamListInfo _SPI_convert_params(int nargs, Oid *argtypes,
--- 51,57 ----
ParamListInfo boundParams);
static int _SPI_execute_plan(SPIPlanPtr plan, ParamListInfo paramLI,
! Snapshot snapshot,
bool read_only, bool fire_triggers, long tcount);
static ParamListInfo _SPI_convert_params(int nargs, Oid *argtypes,
***************
*** 357,363 **** SPI_execute(const char *src, bool read_only, long tcount)
_SPI_prepare_plan(src, &plan, NULL);
res = _SPI_execute_plan(&plan, NULL,
! InvalidSnapshot, InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
--- 357,363 ----
_SPI_prepare_plan(src, &plan, NULL);
res = _SPI_execute_plan(&plan, NULL,
! InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
***************
*** 392,398 **** SPI_execute_plan(SPIPlanPtr plan, Datum *Values, const char *Nulls,
_SPI_convert_params(plan->nargs, plan->argtypes,
Values, Nulls,
0),
! InvalidSnapshot, InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
--- 392,398 ----
_SPI_convert_params(plan->nargs, plan->argtypes,
Values, Nulls,
0),
! InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
***************
*** 421,427 **** SPI_execute_plan_with_paramlist(SPIPlanPtr plan, ParamListInfo params,
return res;
res = _SPI_execute_plan(plan, params,
! InvalidSnapshot, InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
--- 421,427 ----
return res;
res = _SPI_execute_plan(plan, params,
! InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
***************
*** 444,450 **** SPI_execute_plan_with_paramlist(SPIPlanPtr plan, ParamListInfo params,
int
SPI_execute_snapshot(SPIPlanPtr plan,
Datum *Values, const char *Nulls,
! Snapshot snapshot, Snapshot crosscheck_snapshot,
bool read_only, bool fire_triggers, long tcount)
{
int res;
--- 444,450 ----
int
SPI_execute_snapshot(SPIPlanPtr plan,
Datum *Values, const char *Nulls,
! Snapshot snapshot,
bool read_only, bool fire_triggers, long tcount)
{
int res;
***************
*** 463,469 **** SPI_execute_snapshot(SPIPlanPtr plan,
_SPI_convert_params(plan->nargs, plan->argtypes,
Values, Nulls,
0),
! snapshot, crosscheck_snapshot,
read_only, fire_triggers, tcount);
_SPI_end_call(true);
--- 463,469 ----
_SPI_convert_params(plan->nargs, plan->argtypes,
Values, Nulls,
0),
! snapshot,
read_only, fire_triggers, tcount);
_SPI_end_call(true);
***************
*** 516,522 **** SPI_execute_with_args(const char *src,
/* We don't need to copy the plan since it will be thrown away anyway */
res = _SPI_execute_plan(&plan, paramLI,
! InvalidSnapshot, InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
--- 516,522 ----
/* We don't need to copy the plan since it will be thrown away anyway */
res = _SPI_execute_plan(&plan, paramLI,
! InvalidSnapshot,
read_only, true, tcount);
_SPI_end_call(true);
***************
*** 1752,1758 **** _SPI_prepare_plan(const char *src, SPIPlanPtr plan, ParamListInfo boundParams)
*
* snapshot: query snapshot to use, or InvalidSnapshot for the normal
* behavior of taking a new snapshot for each query.
- * crosscheck_snapshot: for RI use, all others pass InvalidSnapshot
* read_only: TRUE for read-only execution (no CommandCounterIncrement)
* fire_triggers: TRUE to fire AFTER triggers at end of query (normal case);
* FALSE means any AFTER triggers are postponed to end of outer query
--- 1752,1757 ----
***************
*** 1760,1766 **** _SPI_prepare_plan(const char *src, SPIPlanPtr plan, ParamListInfo boundParams)
*/
static int
_SPI_execute_plan(SPIPlanPtr plan, ParamListInfo paramLI,
! Snapshot snapshot, Snapshot crosscheck_snapshot,
bool read_only, bool fire_triggers, long tcount)
{
int my_res = 0;
--- 1759,1765 ----
*/
static int
_SPI_execute_plan(SPIPlanPtr plan, ParamListInfo paramLI,
! Snapshot snapshot,
bool read_only, bool fire_triggers, long tcount)
{
int my_res = 0;
***************
*** 1903,1909 **** _SPI_execute_plan(SPIPlanPtr plan, ParamListInfo paramLI,
qdesc = CreateQueryDesc((PlannedStmt *) stmt,
plansource->query_string,
! snap, crosscheck_snapshot,
dest,
paramLI, 0);
res = _SPI_pquery(qdesc, fire_triggers,
--- 1902,1908 ----
qdesc = CreateQueryDesc((PlannedStmt *) stmt,
plansource->query_string,
! snap,
dest,
paramLI, 0);
res = _SPI_pquery(qdesc, fire_triggers,
*** a/src/backend/tcop/pquery.c
--- b/src/backend/tcop/pquery.c
***************
*** 64,70 **** QueryDesc *
CreateQueryDesc(PlannedStmt *plannedstmt,
const char *sourceText,
Snapshot snapshot,
- Snapshot crosscheck_snapshot,
DestReceiver *dest,
ParamListInfo params,
int instrument_options)
--- 64,69 ----
***************
*** 76,83 **** CreateQueryDesc(PlannedStmt *plannedstmt,
qd->utilitystmt = plannedstmt->utilityStmt; /* in case DECLARE CURSOR */
qd->sourceText = sourceText; /* query text */
qd->snapshot = RegisterSnapshot(snapshot); /* snapshot */
- /* RI check snapshot */
- qd->crosscheck_snapshot = RegisterSnapshot(crosscheck_snapshot);
qd->dest = dest; /* output dest */
qd->params = params; /* parameter values passed into query */
qd->instrument_options = instrument_options; /* instrumentation
--- 75,80 ----
***************
*** 109,115 **** CreateUtilityQueryDesc(Node *utilitystmt,
qd->utilitystmt = utilitystmt; /* utility command */
qd->sourceText = sourceText; /* query text */
qd->snapshot = RegisterSnapshot(snapshot); /* snapshot */
- qd->crosscheck_snapshot = InvalidSnapshot; /* RI check snapshot */
qd->dest = dest; /* output dest */
qd->params = params; /* parameter values passed into query */
qd->instrument_options = false; /* uninteresting for utilities */
--- 106,111 ----
***************
*** 134,140 **** FreeQueryDesc(QueryDesc *qdesc)
/* forget our snapshots */
UnregisterSnapshot(qdesc->snapshot);
- UnregisterSnapshot(qdesc->crosscheck_snapshot);
/* Only the QueryDesc itself need be freed */
pfree(qdesc);
--- 130,135 ----
***************
*** 178,184 **** ProcessQuery(PlannedStmt *plan,
* Create the QueryDesc object
*/
queryDesc = CreateQueryDesc(plan, sourceText,
! GetActiveSnapshot(), InvalidSnapshot,
dest, params, 0);
/*
--- 173,179 ----
* Create the QueryDesc object
*/
queryDesc = CreateQueryDesc(plan, sourceText,
! GetActiveSnapshot(),
dest, params, 0);
/*
***************
*** 514,520 **** PortalStart(Portal portal, ParamListInfo params, Snapshot snapshot)
queryDesc = CreateQueryDesc((PlannedStmt *) linitial(portal->stmts),
portal->sourceText,
GetActiveSnapshot(),
- InvalidSnapshot,
None_Receiver,
params,
0);
--- 509,514 ----
*** a/src/backend/utils/adt/ri_triggers.c
--- b/src/backend/utils/adt/ri_triggers.c
***************
*** 230,236 **** static SPIPlanPtr ri_PlanCheck(const char *querystr, int nargs, Oid *argtypes,
static bool ri_PerformCheck(RI_QueryKey *qkey, SPIPlanPtr qplan,
Relation fk_rel, Relation pk_rel,
HeapTuple old_tuple, HeapTuple new_tuple,
- bool detectNewRows,
int expect_OK, const char *constrname);
static void ri_ExtractValues(RI_QueryKey *qkey, int key_idx,
Relation rel, HeapTuple tuple,
--- 230,235 ----
***************
*** 357,363 **** RI_FKey_check(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
NULL, NULL,
- false,
SPI_OK_SELECT,
NameStr(riinfo.conname));
--- 356,361 ----
***************
*** 500,506 **** RI_FKey_check(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
NULL, new_row,
- false,
SPI_OK_SELECT,
NameStr(riinfo.conname));
--- 498,503 ----
***************
*** 661,667 **** ri_Check_Pk_Match(Relation pk_rel, Relation fk_rel,
result = ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* treat like update */
SPI_OK_SELECT, NULL);
if (SPI_finish() != SPI_OK_FINISH)
--- 658,663 ----
***************
*** 818,824 **** RI_FKey_noaction_del(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_SELECT,
NameStr(riinfo.conname));
--- 814,819 ----
***************
*** 1006,1012 **** RI_FKey_noaction_upd(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_SELECT,
NameStr(riinfo.conname));
--- 1001,1006 ----
***************
*** 1168,1174 **** RI_FKey_cascade_del(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_DELETE,
NameStr(riinfo.conname));
--- 1162,1167 ----
***************
*** 1356,1362 **** RI_FKey_cascade_upd(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, new_row,
- true, /* must detect new rows */
SPI_OK_UPDATE,
NameStr(riinfo.conname));
--- 1349,1354 ----
***************
*** 1527,1533 **** RI_FKey_restrict_del(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_SELECT,
NameStr(riinfo.conname));
--- 1519,1524 ----
***************
*** 1710,1716 **** RI_FKey_restrict_upd(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_SELECT,
NameStr(riinfo.conname));
--- 1701,1706 ----
***************
*** 1881,1887 **** RI_FKey_setnull_del(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_UPDATE,
NameStr(riinfo.conname));
--- 1871,1876 ----
***************
*** 2097,2103 **** RI_FKey_setnull_upd(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_UPDATE,
NameStr(riinfo.conname));
--- 2086,2091 ----
***************
*** 2269,2275 **** RI_FKey_setdefault_del(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_UPDATE,
NameStr(riinfo.conname));
--- 2257,2262 ----
***************
*** 2472,2478 **** RI_FKey_setdefault_upd(PG_FUNCTION_ARGS)
ri_PerformCheck(&qkey, qplan,
fk_rel, pk_rel,
old_row, NULL,
- true, /* must detect new rows */
SPI_OK_UPDATE,
NameStr(riinfo.conname));
--- 2459,2464 ----
***************
*** 2792,2798 **** RI_Initial_Check(Trigger *trigger, Relation fk_rel, Relation pk_rel)
spi_result = SPI_execute_snapshot(qplan,
NULL, NULL,
GetLatestSnapshot(),
- InvalidSnapshot,
true, false, 1);
/* Check result */
--- 2778,2783 ----
***************
*** 3271,3284 **** static bool
ri_PerformCheck(RI_QueryKey *qkey, SPIPlanPtr qplan,
Relation fk_rel, Relation pk_rel,
HeapTuple old_tuple, HeapTuple new_tuple,
- bool detectNewRows,
int expect_OK, const char *constrname)
{
Relation query_rel,
source_rel;
int key_idx;
Snapshot test_snapshot;
- Snapshot crosscheck_snapshot;
int limit;
int spi_result;
Oid save_userid;
--- 3256,3267 ----
***************
*** 3330,3359 **** ri_PerformCheck(RI_QueryKey *qkey, SPIPlanPtr qplan,
}
/*
- * In READ COMMITTED mode, we just need to use an up-to-date regular
- * snapshot, and we will see all rows that could be interesting. But in
- * transaction-snapshot mode, we can't change the transaction snapshot.
- * If the caller passes detectNewRows == false then it's okay to do the query
- * with the transaction snapshot; otherwise we use a current snapshot, and
- * tell the executor to error out if it finds any rows under the current
- * snapshot that wouldn't be visible per the transaction snapshot. Note
- * that SPI_execute_snapshot will register the snapshots, so we don't need
- * to bother here.
- */
- if (IsolationUsesXactSnapshot() && detectNewRows)
- {
- CommandCounterIncrement(); /* be sure all my own work is visible */
- test_snapshot = GetLatestSnapshot();
- crosscheck_snapshot = GetTransactionSnapshot();
- }
- else
- {
- /* the default SPI behavior is okay */
- test_snapshot = InvalidSnapshot;
- crosscheck_snapshot = InvalidSnapshot;
- }
-
- /*
* If this is a select query (e.g., for a 'no action' or 'restrict'
* trigger), we only need to see if there is a single row in the table,
* matching the key. Otherwise, limit = 0 - because we want the query to
--- 3313,3318 ----
***************
*** 3369,3375 **** ri_PerformCheck(RI_QueryKey *qkey, SPIPlanPtr qplan,
/* Finally we can run the query. */
spi_result = SPI_execute_snapshot(qplan,
vals, nulls,
! test_snapshot, crosscheck_snapshot,
false, false, limit);
/* Restore UID and security context */
--- 3328,3334 ----
/* Finally we can run the query. */
spi_result = SPI_execute_snapshot(qplan,
vals, nulls,
! InvalidSnapshot,
false, false, limit);
/* Restore UID and security context */
*** a/src/backend/utils/time/snapmgr.c
--- b/src/backend/utils/time/snapmgr.c
***************
*** 27,32 ****
--- 27,33 ----
#include "access/transam.h"
#include "access/xact.h"
+ #include "access/multixact.h"
#include "storage/proc.h"
#include "storage/procarray.h"
#include "utils/memutils.h"
***************
*** 126,131 **** GetTransactionSnapshot(void)
--- 127,142 ----
{
Assert(RegisteredSnapshots == 0);
+ /*
+ * We must store the oldest visible multi xact *before* taking the
+ * transaction snapshot. Otherwise HeapSatisfiesLockersVisible in
+ * heapam.c will be in trouble, since it depends on being able to
+ * inspect all multi xact ids which might contain xids invisible to
+ * the transaction snapshot.
+ */
+ if (IsolationUsesXactSnapshot())
+ MultiXactIdSetOldestVisible();
+
CurrentSnapshot = GetSnapshotData(&CurrentSnapshotData);
FirstSnapshotSet = true;
*** a/src/backend/utils/time/tqual.c
--- b/src/backend/utils/time/tqual.c
***************
*** 72,80 **** SnapshotData SnapshotSelfData = {HeapTupleSatisfiesSelf};
SnapshotData SnapshotAnyData = {HeapTupleSatisfiesAny};
SnapshotData SnapshotToastData = {HeapTupleSatisfiesToast};
- /* local functions */
- static bool XidInMVCCSnapshot(TransactionId xid, Snapshot snapshot);
-
/*
* SetHintBits()
--- 72,77 ----
***************
*** 1253,1259 **** HeapTupleSatisfiesVacuum(HeapTupleHeader tuple, TransactionId OldestXmin,
* by this function. This is OK for current uses, because we actually only
* apply this for known-committed XIDs.
*/
! static bool
XidInMVCCSnapshot(TransactionId xid, Snapshot snapshot)
{
uint32 i;
--- 1250,1256 ----
* by this function. This is OK for current uses, because we actually only
* apply this for known-committed XIDs.
*/
! bool
XidInMVCCSnapshot(TransactionId xid, Snapshot snapshot)
{
uint32 i;
*** a/src/include/access/heapam.h
--- b/src/include/access/heapam.h
***************
*** 98,112 **** extern Oid heap_insert(Relation relation, HeapTuple tup, CommandId cid,
int options, BulkInsertState bistate);
extern HTSU_Result heap_delete(Relation relation, ItemPointer tid,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, Snapshot crosscheck, bool wait);
extern HTSU_Result heap_update(Relation relation, ItemPointer otid,
HeapTuple newtup,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, Snapshot crosscheck, bool wait);
extern HTSU_Result heap_lock_tuple(Relation relation, HeapTuple tuple,
Buffer *buffer, ItemPointer ctid,
TransactionId *update_xmax, CommandId cid,
! LockTupleMode mode, bool nowait);
extern void heap_inplace_update(Relation relation, HeapTuple tuple);
extern bool heap_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid,
Buffer buf);
--- 98,113 ----
int options, BulkInsertState bistate);
extern HTSU_Result heap_delete(Relation relation, ItemPointer tid,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, bool wait, Snapshot lockcheck_snapshot);
extern HTSU_Result heap_update(Relation relation, ItemPointer otid,
HeapTuple newtup,
ItemPointer ctid, TransactionId *update_xmax,
! CommandId cid, bool wait, Snapshot lockcheck_snapshot);
extern HTSU_Result heap_lock_tuple(Relation relation, HeapTuple tuple,
Buffer *buffer, ItemPointer ctid,
TransactionId *update_xmax, CommandId cid,
! LockTupleMode mode, bool nowait,
! Snapshot lockcheck_snapshot);
extern void heap_inplace_update(Relation relation, HeapTuple tuple);
extern bool heap_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid,
Buffer buf);
*** a/src/include/access/multixact.h
--- b/src/include/access/multixact.h
***************
*** 49,54 **** extern bool MultiXactIdIsCurrent(MultiXactId multi);
--- 49,55 ----
extern void MultiXactIdWait(MultiXactId multi);
extern bool ConditionalMultiXactIdWait(MultiXactId multi);
extern void MultiXactIdSetOldestMember(void);
+ extern void MultiXactIdSetOldestVisible(void);
extern int GetMultiXactIdMembers(MultiXactId multi, TransactionId **xids);
extern void AtEOXact_MultiXact(void);
*** a/src/include/executor/execdesc.h
--- b/src/include/executor/execdesc.h
***************
*** 39,45 **** typedef struct QueryDesc
Node *utilitystmt; /* utility statement, or null */
const char *sourceText; /* source text of the query */
Snapshot snapshot; /* snapshot to use for query */
- Snapshot crosscheck_snapshot; /* crosscheck for RI update/delete */
DestReceiver *dest; /* the destination for tuple output */
ParamListInfo params; /* param values being passed in */
int instrument_options; /* OR of InstrumentOption flags */
--- 39,44 ----
***************
*** 57,63 **** typedef struct QueryDesc
extern QueryDesc *CreateQueryDesc(PlannedStmt *plannedstmt,
const char *sourceText,
Snapshot snapshot,
- Snapshot crosscheck_snapshot,
DestReceiver *dest,
ParamListInfo params,
int instrument_options);
--- 56,61 ----
*** a/src/include/executor/spi.h
--- b/src/include/executor/spi.h
***************
*** 82,88 **** extern int SPI_execp(SPIPlanPtr plan, Datum *Values, const char *Nulls,
extern int SPI_execute_snapshot(SPIPlanPtr plan,
Datum *Values, const char *Nulls,
Snapshot snapshot,
- Snapshot crosscheck_snapshot,
bool read_only, bool fire_triggers, long tcount);
extern int SPI_execute_with_args(const char *src,
int nargs, Oid *argtypes,
--- 82,87 ----
*** a/src/include/nodes/execnodes.h
--- b/src/include/nodes/execnodes.h
***************
*** 337,343 **** typedef struct EState
/* Basic state for all query types: */
ScanDirection es_direction; /* current scan direction */
Snapshot es_snapshot; /* time qual to use */
- Snapshot es_crosscheck_snapshot; /* crosscheck time qual for RI */
List *es_range_table; /* List of RangeTblEntry */
PlannedStmt *es_plannedstmt; /* link to top of plan tree */
--- 337,342 ----
*** a/src/include/utils/tqual.h
--- b/src/include/utils/tqual.h
***************
*** 41,46 **** extern PGDLLIMPORT SnapshotData SnapshotToastData;
--- 41,48 ----
#define IsMVCCSnapshot(snapshot) \
((snapshot)->satisfies == HeapTupleSatisfiesMVCC)
+ bool XidInMVCCSnapshot(TransactionId xid, Snapshot snapshot);
+
/*
* HeapTupleSatisfiesVisibility
* True iff heap tuple satisfies a time qual.