Files
openGauss-server/src/gausskernel/storage/ipc/standby.cpp
Chunling Wang 028719bd3b issue#I5UEWO Cleanup sessions for one database which
is conflict with when recovery drop database
2022-10-09 16:54:11 +08:00

1293 lines
49 KiB
C++
Executable File

/* -------------------------------------------------------------------------
*
* standby.cpp
* Misc functions used in Hot Standby mode.
*
* All functions for handling RM_STANDBY_ID, which relate to
* AccessExclusiveLocks and starting snapshots for Hot Standby mode.
* Plus conflict recovery processing.
*
* Portions Copyright (c) 2020 Huawei Technologies Co.,Ltd.
* Portions Copyright (c) 1996-2012, PostgreSQL Global Development Group
* Portions Copyright (c) 1994, Regents of the University of California
*
* IDENTIFICATION
* src/gausskernel/storage/ipc/standby.cpp
*
* -------------------------------------------------------------------------
*/
#include "postgres.h"
#include "knl/knl_variable.h"
#include "access/transam.h"
#include "access/twophase.h"
#include "access/xact.h"
#include "access/xloginsert.h"
#include "access/csnlog.h"
#include "access/multi_redo_api.h"
#include "miscadmin.h"
#include "storage/buf/bufmgr.h"
#include "storage/lmgr.h"
#include "storage/proc.h"
#include "storage/procarray.h"
#include "storage/sinvaladt.h"
#include "storage/standby.h"
#include "utils/ps_status.h"
#include "utils/timestamp.h"
#include "utils/snapmgr.h"
#include "pgxc/poolutils.h"
#include "replication/walreceiver.h"
static void ResolveRecoveryConflictWithVirtualXIDs(VirtualTransactionId* waitlist, TransactionId* xminArray,
ProcSignalReason reason);
static void ResolveRecoveryConflictWithLock(Oid dbOid, Oid relOid);
static void LogAccessExclusiveLocks(int nlocks, xl_standby_lock* locks);
static void LogReleaseAccessExclusiveLocks(int nlocks, xl_standby_lock* locks);
static XLogRecPtr LogCurrentRunningXacts(RunningTransactions CurrRunningXacts);
static void RecordCommittingCsnInfo(TransactionId xid);
void InitRecoveryLockHash()
{
HASHCTL hash_ctl;
const long maxNum = 64;
/*
* Initialize the hash table for tracking the list of locks held by each
* transaction.
*/
errno_t rc = memset_s(&hash_ctl, sizeof(hash_ctl), 0, sizeof(hash_ctl));
securec_check_ss(rc, "\0", "\0");
hash_ctl.keysize = sizeof(TransactionId);
hash_ctl.entrysize = sizeof(RecoveryLockListsEntry);
t_thrd.storage_cxt.RecoveryLockList = hash_create("RecoveryLockLists", maxNum, &hash_ctl, HASH_ELEM | HASH_BLOBS);
}
/*
* InitRecoveryTransactionEnvironment
* Initialize tracking of in-progress transactions in master
*
* We need to issue shared invalidations and hold locks. Holding locks
* means others may want to wait on us, so we need to make a lock table
* vxact entry like a real transaction. We could create and delete
* lock table entries for each transaction but its simpler just to create
* one permanent entry and leave it there all the time. Locks are then
* acquired and released as needed. Yes, this means you can see the
* Startup process in pg_locks once we have run this.
*/
void InitRecoveryTransactionEnvironment(void)
{
VirtualTransactionId vxid;
InitRecoveryLockHash();
/*
* Initialize shared invalidation management for Startup process, being
* careful to register ourselves as a sendOnly process so we don't need to
* read messages, nor will we get signalled when the queue starts filling
* up.
*/
SharedInvalBackendInit(true, false);
/*
* Lock a virtual transaction id for Startup process.
*
* We need to do GetNextLocalTransactionId() because
* SharedInvalBackendInit() leaves localTransactionid invalid and the lock
* manager doesn't like that at all.
*
* Note that we don't need to run XactLockTableInsert() because nobody
* needs to wait on xids. That sounds a little strange, but table locks
* are held by vxids and row level locks are held by xids. All queries
* hold AccessShareLocks so never block while we write or lock new rows.
*/
vxid.backendId = t_thrd.proc_cxt.MyBackendId;
vxid.localTransactionId = GetNextLocalTransactionId();
VirtualXactLockTableInsert(vxid);
t_thrd.xlog_cxt.standbyState = STANDBY_INITIALIZED;
}
/*
* ShutdownRecoveryTransactionEnvironment
* Shut down transaction tracking
*
* Prepare to switch from hot standby mode to normal operation. Shut down
* recovery-time transaction tracking.
*/
void ShutdownRecoveryTransactionEnvironment(void)
{
/* Release all locks the tracked transactions were holding */
StandbyReleaseAllLocks();
hash_destroy(t_thrd.storage_cxt.RecoveryLockList);
t_thrd.storage_cxt.RecoveryLockList = NULL;
/* Cleanup our VirtualTransaction */
VirtualXactLockTableCleanup();
}
/*
* -----------------------------------------------------
* Standby wait timers and backend cancel logic
* -----------------------------------------------------
*/
/*
* Determine the cutoff time at which we want to start canceling conflicting
* transactions. Returns zero (a time safely in the past) if we are willing
* to wait forever.
*/
static TimestampTz GetStandbyLimitTime(TimestampTz startTime)
{
TimestampTz rtime = startTime;
bool fromStream = (t_thrd.xlog_cxt.XLogReceiptSource == XLOG_FROM_STREAM);
if (fromStream) {
if (u_sess->attr.attr_storage.max_standby_streaming_delay < 0)
return 0; /* wait forever */
return TimestampTzPlusMilliseconds(rtime, u_sess->attr.attr_storage.max_standby_streaming_delay);
} else {
if (u_sess->attr.attr_storage.max_standby_archive_delay < 0)
return 0; /* wait forever */
return TimestampTzPlusMilliseconds(rtime, u_sess->attr.attr_storage.max_standby_archive_delay);
}
}
#define STANDBY_INITIAL_WAIT_US 1000
/*
* Standby wait logic for ResolveRecoveryConflictWithVirtualXIDs.
* We wait here for a while then return. If we decide we can't wait any
* more then we return true, if we can wait some more return false.
*/
static bool WaitExceedsMaxStandbyDelay(TimestampTz startTime)
{
TimestampTz ltime = 0;
/* Are we past the limit time? */
ltime = GetStandbyLimitTime(startTime);
if (ltime && GetCurrentTimestamp() >= ltime)
return true;
/*
* Sleep a bit (this is essential to avoid busy-waiting).
*/
pg_usleep(t_thrd.storage_cxt.standbyWait_us);
/*
* Progressively increase the sleep times, but not to more than 1s, since
* pg_usleep isn't interruptable on some platforms.
*/
t_thrd.storage_cxt.standbyWait_us *= 2;
if (t_thrd.storage_cxt.standbyWait_us > 1000000)
t_thrd.storage_cxt.standbyWait_us = 1000000;
return false;
}
/*
* This is the main executioner for any query backend that conflicts with
* recovery processing. Judgement has already been passed on it within
* a specific rmgr. Here we just issue the orders to the procs. The procs
* then throw the required error as instructed.
*/
static void ResolveRecoveryConflictWithVirtualXIDs(VirtualTransactionId* waitlist, TransactionId* xminArray,
ProcSignalReason reason)
{
TimestampTz waitStart;
char* new_status = NULL;
bool waited = false;
/* Fast exit, to avoid a kernel call if there's no work to be done. */
if (!VirtualTransactionIdIsValid(*waitlist))
return;
waitStart = GetCurrentTimestamp();
while (VirtualTransactionIdIsValid(*waitlist)) {
/* reset standbyWait_us for each xact we wait for */
t_thrd.storage_cxt.standbyWait_us = STANDBY_INITIAL_WAIT_US;
bool notPrintNotified = true;
/* wait until the virtual xid is gone */
while (!VirtualXactLock(*waitlist, false)) {
PGPROC* proc = BackendIdGetProc((*waitlist).backendId);
PGXACT* pgxact = &g_instance.proc_base_all_xacts[proc->pgprocno];
if (xminArray != NULL && pgxact->xmin != *xminArray) {
ereport(WARNING, (errmsg("hotstandby:snapshot changed old xmin = %lu, new xmin = %lu",
*xminArray, pgxact->xmin)));
break;
}
/*
* Report via ps if we have been waiting for more than 500 msec
* (should that be configurable?)
*/
if (u_sess->attr.attr_common.update_process_title && new_status == NULL &&
TimestampDifferenceExceeds(waitStart, GetCurrentTimestamp(), 500)) {
const char* old_status = NULL;
int len;
errno_t rc;
const int strlen = 9;
old_status = get_ps_display(&len);
new_status = (char*)palloc(len + strlen);
if (len > 0) {
rc = memcpy_s(new_status, len, old_status, len);
securec_check(rc, "\0", "\0");
}
rc = strcpy_s(new_status + len, strlen, " waiting");
securec_check(rc, "\0", "\0");
set_ps_display(new_status, false);
new_status[len] = '\0'; /* truncate off " waiting" */
}
/* Is it time to kill it? */
if (WaitExceedsMaxStandbyDelay(waitStart)) {
ThreadId pid;
/*
* Now find out who to throw out of the balloon.
*/
Assert(VirtualTransactionIdIsValid(*waitlist));
pid = CancelVirtualTransaction(*waitlist, reason);
/*
* Wait a little bit for it to die so that we avoid flooding
* an unresponsive backend when system is heavily loaded.
*/
if (pid != 0)
pg_usleep(5000L);
if (notPrintNotified) {
notPrintNotified = false;
waited = true;
ereport(LOG, (errmsg("hotstandby:snapshot changed, wait pid %lu", pid)));
}
}
}
/* The virtual transaction is gone now, wait for the next one */
waitlist++;
if (xminArray != NULL)
xminArray++;
}
/* Reset ps display if we changed it */
if (new_status != NULL) {
set_ps_display(new_status, false);
pfree(new_status);
new_status = NULL;
}
if (waited) {
ereport(LOG, (errmsg("hotstandby:snapshot Conflict resolve ok, wait time:%d ms",
ComputeTimeStamp(waitStart))));
}
}
void ResolveRecoveryConflictWithSnapshot(TransactionId latestRemovedXid, const RelFileNode& node, XLogRecPtr lsn)
{
VirtualTransactionId* backends = NULL;
/*
* If we get passed InvalidTransactionId then we are a little surprised,
* but it is theoretically possible in normal running. It also happens
* when replaying already applied WAL records after a standby crash or
* restart. If latestRemovedXid is invalid then there is no conflict. That
* rule applies across all record types that suffer from this conflict.
*/
if (!TransactionIdIsValid(latestRemovedXid))
return;
CommitSeqNo limitXminCSN = InvalidCommitSeqNo;
if (IS_DISASTER_RECOVER_MODE) {
limitXminCSN = CSNLogGetDRCommitSeqNo(latestRemovedXid);
while (true) {
CommitSeqNo lastReplayedConflictCSN = (CommitSeqNo)pg_atomic_read_u64(
&(g_instance.comm_cxt.predo_cxt.last_replayed_conflict_csn));
if (limitXminCSN <= lastReplayedConflictCSN) {
break;
}
if (pg_atomic_compare_exchange_u64(&(g_instance.comm_cxt.predo_cxt.last_replayed_conflict_csn),
&lastReplayedConflictCSN, limitXminCSN)) {
break;
}
}
}
ProcArrayStruct* arrayP = g_instance.proc_array_idx;
if (t_thrd.storage_cxt.xminArray == NULL) {
t_thrd.storage_cxt.xminArray = (TransactionId*)MemoryContextAlloc(
THREAD_GET_MEM_CXT_GROUP(MEMORY_CONTEXT_STORAGE),
sizeof(TransactionId) * (unsigned int)(arrayP->maxProcs + 1));
if (t_thrd.storage_cxt.xminArray == NULL)
ereport(ERROR, (errcode(ERRCODE_OUT_OF_MEMORY), errmsg("out of memory")));
}
backends = GetConflictingVirtualXIDs(latestRemovedXid, node.dbNode, lsn, limitXminCSN,
t_thrd.storage_cxt.xminArray);
ResolveRecoveryConflictWithVirtualXIDs(backends, t_thrd.storage_cxt.xminArray, PROCSIG_RECOVERY_CONFLICT_SNAPSHOT);
}
void ResolveRecoveryConflictWithSnapshotOid(TransactionId latestRemovedXid, Oid dbid, XLogRecPtr lsn)
{
VirtualTransactionId* backends = NULL;
/*
* If we get passed InvalidTransactionId then we are a little surprised,
* but it is theoretically possible in normal running. It also happens
* when replaying already applied WAL records after a standby crash or
* restart. If latestRemovedXid is invalid then there is no conflict. That
* rule applies across all record types that suffer from this conflict.
*/
if (!TransactionIdIsValid(latestRemovedXid)) {
return;
}
ProcArrayStruct* arrayP = g_instance.proc_array_idx;
if (t_thrd.storage_cxt.xminArray == NULL) {
t_thrd.storage_cxt.xminArray = (TransactionId*)MemoryContextAlloc(
THREAD_GET_MEM_CXT_GROUP(MEMORY_CONTEXT_STORAGE),
sizeof(TransactionId) * (unsigned int)(arrayP->maxProcs + 1));
if (t_thrd.storage_cxt.xminArray == NULL)
ereport(ERROR, (errcode(ERRCODE_OUT_OF_MEMORY), errmsg("out of memory")));
}
backends = GetConflictingVirtualXIDs(latestRemovedXid, dbid, lsn, InvalidCommitSeqNo, t_thrd.storage_cxt.xminArray);
ResolveRecoveryConflictWithVirtualXIDs(backends, t_thrd.storage_cxt.xminArray, PROCSIG_RECOVERY_CONFLICT_SNAPSHOT);
}
void ResolveRecoveryConflictWithTablespace(Oid tsid)
{
VirtualTransactionId* temp_file_users = NULL;
/*
* Standby users may be currently using this tablespace for their
* temporary files. We only care about current users because
* temp_tablespace parameter will just ignore tablespaces that no longer
* exist.
*
* Ask everybody to cancel their queries immediately so we can ensure no
* temp files remain and we can remove the tablespace. Nuke the entire
* site from orbit, it's the only way to be sure.
*
* XXX: We could work out the pids of active backends using this
* tablespace by examining the temp filenames in the directory. We would
* then convert the pids into VirtualXIDs before attempting to cancel
* them.
*
* We don't wait for commit because drop tablespace is non-transactional.
*/
temp_file_users = GetConflictingVirtualXIDs(InvalidTransactionId, InvalidOid);
ResolveRecoveryConflictWithVirtualXIDs(temp_file_users, NULL, PROCSIG_RECOVERY_CONFLICT_TABLESPACE);
}
void ResolveRecoveryConflictWithDatabase(Oid dbid)
{
/*
* We don't do ResolveRecoveryConflictWithVirtualXIDs() here since that
* only waits for transactions and completely idle sessions would block
* us. This is rare enough that we do this as simply as possible: no wait,
* just force them off immediately.
*
* No locking is required here because we already acquired
* AccessExclusiveLock. Anybody trying to connect while we do this will
* block during InitPostgres() and then disconnect when they see the
* database has been removed.
*/
if (ENABLE_THREAD_POOL) {
ThreadPoolSessControl *sess_ctrl = g_threadPoolControler->GetSessionCtrl();
while (sess_ctrl->CountDBSessionsNotCleaned(dbid, InvalidOid) > 0) {
sess_ctrl->CleanDBSessions(dbid, InvalidOid);
pg_usleep(10000);
}
} else {
while (CountDBBackends(dbid) > 0) {
CancelDBBackends(dbid, PROCSIG_RECOVERY_CONFLICT_DATABASE, true);
/*
* Wait awhile for them to die so that we avoid flooding an
* unresponsive backend when system is heavily loaded.
*/
pg_usleep(10000);
}
}
}
static void ResolveRecoveryConflictWithLock(Oid dbOid, Oid relOid)
{
VirtualTransactionId* backends = NULL;
bool lock_acquired = false;
int num_attempts = 0;
LOCKTAG locktag;
SET_LOCKTAG_RELATION(locktag, dbOid, relOid);
/*
* If blowing away everybody with conflicting locks doesn't work, after
* the first two attempts then we just start blowing everybody away until
* it does work. We do this because its likely that we either have too
* many locks and we just can't get one at all, or that there are many
* people crowding for the same table. Recovery must win; the end
* justifies the means.
*/
while (!lock_acquired) {
if (++num_attempts < 3)
backends = GetLockConflicts(&locktag, AccessExclusiveLock);
else
backends = GetConflictingVirtualXIDs(InvalidTransactionId, InvalidOid);
ResolveRecoveryConflictWithVirtualXIDs(backends, NULL, PROCSIG_RECOVERY_CONFLICT_LOCK);
if (LockAcquireExtended(&locktag, AccessExclusiveLock, true, true, false) != LOCKACQUIRE_NOT_AVAIL)
lock_acquired = true;
}
}
/*
* ResolveRecoveryConflictWithBufferPin is called from LockBufferForCleanup()
* to resolve conflicts with other backends holding buffer pins.
*
* We either resolve conflicts immediately or set a SIGALRM to wake us at
* the limit of our patience. The sleep in LockBufferForCleanup() is
* performed here, for code clarity.
*
* Resolve conflicts by sending a PROCSIG signal to all backends to check if
* they hold one of the buffer pins that is blocking Startup process. If so,
* backends will take an appropriate error action, ERROR or FATAL.
*
* We also must check for deadlocks. Deadlocks occur because if queries
* wait on a lock, that must be behind an AccessExclusiveLock, which can only
* be cleared if the Startup process replays a transaction completion record.
* If Startup process is also waiting then that is a deadlock. The deadlock
* can occur if the query is waiting and then the Startup sleeps, or if
* Startup is sleeping and the query waits on a lock. We protect against
* only the former sequence here, the latter sequence is checked prior to
* the query sleeping, in CheckRecoveryConflictDeadlock().
*
* Deadlocks are extremely rare, and relatively expensive to check for,
* so we don't do a deadlock check right away ... only if we have had to wait
* at least deadlock_timeout. Most of the logic about that is in proc.c.
*/
void ResolveRecoveryConflictWithBufferPin(void)
{
bool sig_alarm_enabled = false;
TimestampTz ltime;
TimestampTz now;
Assert(InHotStandby);
now = GetCurrentTimestamp();
ltime = GetStandbyLimitTime(now);
if (!ltime) {
/*
* We're willing to wait forever for conflicts, so set timeout for
* deadlock check (only)
*/
if (enable_standby_sig_alarm(now, now, true)) {
sig_alarm_enabled = true;
ereport(LOG, (errmsg("wait forever for process wakeup")));
} else {
ereport(FATAL, (errmsg("could not set timer for process wakeup")));
}
} else if (now >= ltime) {
/*
* We're already behind, so clear a path as quickly as possible.
*/
SendRecoveryConflictWithBufferPin(PROCSIG_RECOVERY_CONFLICT_BUFFERPIN);
} else {
/*
* Wake up at ltime, and check for deadlocks as well if we will be
* waiting longer than deadlock_timeout
*/
if (enable_standby_sig_alarm(now, ltime, false)) {
sig_alarm_enabled = true;
ereport(LOG, (errmsg("wait sometime for process to wakeup")));
} else {
ereport(FATAL, (errmsg("could not set timer for process wakeup")));
}
}
/* Wait to be signaled by UnpinBuffer() */
ProcWaitForSignal();
if (sig_alarm_enabled) {
if (!disable_standby_sig_alarm()) {
ereport(FATAL, (errmsg("could not disable timer for process wakeup")));
}
}
ereport(LOG, (errmsg("buffer pin confilict resolve ok, proc time %d ms", ComputeTimeStamp(now))));
}
void SendRecoveryConflictWithBufferPin(ProcSignalReason reason)
{
Assert(reason == PROCSIG_RECOVERY_CONFLICT_BUFFERPIN || reason == PROCSIG_RECOVERY_CONFLICT_STARTUP_DEADLOCK);
/*
* We send signal to all backends to ask them if they are holding the
* buffer pin which is delaying the Startup process. We must not set the
* conflict flag yet, since most backends will be innocent. Let the
* SIGUSR1 handling in each backend decide their own fate.
*/
CancelDBBackends(InvalidOid, reason, false);
}
/*
* In Hot Standby perform early deadlock detection. We abort the lock
* wait if we are about to sleep while holding the buffer pin that Startup
* process is waiting for.
*
* Note: this code is pessimistic, because there is no way for it to
* determine whether an actual deadlock condition is present: the lock we
* need to wait for might be unrelated to any held by the Startup process.
* Sooner or later, this mechanism should get ripped out in favor of somehow
* accounting for buffer locks in DeadLockCheck(). However, errors here
* seem to be very low-probability in practice, so for now it's not worth
* the trouble.
*/
void CheckRecoveryConflictDeadlock(void)
{
Assert(!t_thrd.xlog_cxt.InRecovery); /* do not call in Startup process */
if (!HoldingBufferPinThatDelaysRecovery())
return;
/*
* Error message should match ProcessInterrupts() but we avoid calling
* that because we aren't handling an interrupt at this point. Note that
* we only cancel the current transaction here, so if we are in a
* subtransaction and the pin is held by a parent, then the Startup
* process will continue to wait even though we have avoided deadlock.
*/
ereport(ERROR,
(errcode(ERRCODE_T_R_DEADLOCK_DETECTED),
errmsg("canceling statement due to conflict with recovery"),
errdetail("User transaction caused buffer deadlock with recovery.")));
}
/*
* Locking in Recovery Mode
*
* All locks are held by the Startup process using a single virtual
* transaction. This implementation is both simpler and in some senses,
* more correct. The locks held mean "some original transaction held
* this lock, so query access is not allowed at this time". So the Startup
* process is the proxy by which the original locks are implemented.
*
* We only keep track of AccessExclusiveLocks, which are only ever held by
* one transaction on one relation, and don't worry about lock queuing.
*
* We keep a single dynamically expandible list of locks in local memory,
* RelationLockList, so we can keep track of the various entries made by
* the Startup process's virtual xid in the shared lock table.
*
* We record the lock against the top-level xid, rather than individual
* subtransaction xids. This means AccessExclusiveLocks held by aborted
* subtransactions are not released as early as possible on standbys.
*
* List elements use type xl_rel_lock, since the WAL record type exactly
* matches the information that we need to keep track of.
*
* We use session locks rather than normal locks so we don't need
* ResourceOwners.
*/
void StandbyAcquireAccessExclusiveLock(TransactionId xid, Oid dbOid, Oid relOid)
{
LOCKTAG locktag;
/* dbOid is InvalidOid when we are locking a shared relation. */
Assert(OidIsValid(relOid));
#ifdef ENABLE_MULTIPLE_NODES
/*
* We skip to acquire standby lock in hot standby mode only for user relation
* instead of skip all. That's necessary 'cause for pgxc, we should not allow user
* query in hot standby. But we have to give the privilege to CM Agent which only
* loads several specific sys relations.
* When recovering vacuum full redo records, those sys relations may be changed by
* redo. We should keep the standby lock to resolve conflicts in redo and standby query.
*
* We checked out that in data replication in standby, we should hold relation lock
* while we are trying to receive data page to buffer(without relation lock, there may be
* coredump in smgr buffer operation in drop/vacuum/truncate scene). That relation
* lock conflicts with hot standby lock, so skip hot standby lock if we are a user relation.
*
* During inplace upgrade, we don't acquire access exclusive lock for system catalogs to
* avoid dead lock.
*/
if (relOid >= FirstNormalObjectId) {
return;
}
#endif
/* Already processed? */
if (!TransactionIdIsValid(xid) || TransactionIdDidCommit(xid) || TransactionIdDidAbort(xid))
return;
ereport(trace_recovery(DEBUG4), (errmsg("adding recovery lock: db %u rel %u", dbOid, relOid)));
bool found = false;
/* Create a new list for this xid, if we don't have one already. */
RecoveryLockListsEntry *entry = (RecoveryLockListsEntry *)hash_search(t_thrd.storage_cxt.RecoveryLockList, &xid,
HASH_ENTER, &found);
if (!found) {
entry->xid = xid;
entry->locks = NIL;
}
xl_standby_lock* newlock = (xl_standby_lock*)palloc(sizeof(xl_standby_lock));
newlock->xid = xid;
newlock->dbOid = dbOid;
newlock->relOid = relOid;
entry->locks = lappend(entry->locks, newlock);
SET_LOCKTAG_RELATION(locktag, newlock->dbOid, newlock->relOid);
if (LockAcquireExtended(&locktag, AccessExclusiveLock, true, true, false) == LOCKACQUIRE_NOT_AVAIL)
ResolveRecoveryConflictWithLock(newlock->dbOid, newlock->relOid);
}
static void StandbyReleaseLockList(List *locks)
{
while (locks) {
xl_standby_lock *lock = (xl_standby_lock *) linitial(locks);
LOCKTAG locktag;
ereport(trace_recovery(DEBUG4), (errmsg("releasing recovery lock: xid %lu db %u rel %u", lock->xid,
lock->dbOid, lock->relOid)));
SET_LOCKTAG_RELATION(locktag, lock->dbOid, lock->relOid);
if (!LockRelease(&locktag, AccessExclusiveLock, true)) {
ereport(LOG, (errmsg("RecoveryLockLists contains entry for lock no longer recorded by lock manager: "
"xid %lu database %u relation %u", lock->xid, lock->dbOid, lock->relOid)));
Assert(false);
}
pfree(lock);
locks = list_delete_first(locks);
}
}
static void StandbyReleaseLocks(TransactionId xid)
{
RecoveryLockListsEntry *entry = NULL;
if (TransactionIdIsValid(xid)) {
entry = (RecoveryLockListsEntry *)hash_search(t_thrd.storage_cxt.RecoveryLockList, &xid, HASH_FIND, NULL);
if (entry != NULL) {
StandbyReleaseLockList(entry->locks);
hash_search(t_thrd.storage_cxt.RecoveryLockList, entry, HASH_REMOVE, NULL);
}
} else {
StandbyReleaseAllLocks();
}
}
/*
* Release locks for a transaction tree, starting at xid down, from
* RecoveryLockList.
*
* Called during WAL replay of COMMIT/ROLLBACK when in hot standby mode,
* to remove any AccessExclusiveLocks requested by a transaction.
*/
void StandbyReleaseLockTree(TransactionId xid, int nsubxids, TransactionId* subxids)
{
int i;
StandbyReleaseLocks(xid);
for (i = 0; i < nsubxids; i++)
StandbyReleaseLocks(subxids[i]);
}
/*
* Called at end of recovery and when we see a shutdown checkpoint.
*/
void StandbyReleaseAllLocks(void)
{
if (t_thrd.storage_cxt.RecoveryLockList == NULL) {
return;
}
HASH_SEQ_STATUS status;
RecoveryLockListsEntry *entry = NULL;
ereport(trace_recovery(DEBUG2), (errmsg("release all standby locks")));
hash_seq_init(&status, t_thrd.storage_cxt.RecoveryLockList);
while ((entry = (RecoveryLockListsEntry *)hash_seq_search(&status))) {
StandbyReleaseLockList(entry->locks);
hash_search(t_thrd.storage_cxt.RecoveryLockList, entry, HASH_REMOVE, NULL);
}
}
/*
* StandbyReleaseOldLocks
* Release standby locks held by top-level XIDs that aren't running,
* as long as they're not prepared transactions.
*/
void StandbyReleaseOldLocks(TransactionId oldxid)
{
if (t_thrd.storage_cxt.RecoveryLockList == NULL) {
return;
}
HASH_SEQ_STATUS status;
RecoveryLockListsEntry *entry = NULL;
hash_seq_init(&status, t_thrd.storage_cxt.RecoveryLockList);
while ((entry = (RecoveryLockListsEntry *)hash_seq_search(&status))) {
Assert(TransactionIdIsValid(entry->xid));
/* Skip if prepared transaction. */
if (StandbyTransactionIdIsPrepared(entry->xid)) {
continue;
}
/* Skip if >= oldxid. */
if (!TransactionIdPrecedes(entry->xid, oldxid)) {
continue;
}
/* Remove all locks and hash table entry. */
StandbyReleaseLockList(entry->locks);
hash_search(t_thrd.storage_cxt.RecoveryLockList, entry, HASH_REMOVE, NULL);
}
}
/*
* For hot standby, maintain a list for csn commting status compensation
* when primary dn restart after crash.
*/
void DealCSNLogForHotStby(XLogReaderState* record, uint8 info)
{
/* XLOG_STANDBY_CSN_COMMITTING and XLOG_STANDBY_CSN_ABORTED need to operate on both primary and standby. */
if (info == XLOG_STANDBY_CSN_COMMITTING) {
uint64* id = ((uint64 *)XLogRecGetData(record));
uint32 newVersionMinSize = 3;
if (XLogRecGetDataLen(record) >= (newVersionMinSize * sizeof(uint64))) {
uint64 childrenxidnum = id[2];
Assert(XLogRecGetDataLen(record) >= ((newVersionMinSize + childrenxidnum) * sizeof(uint64)));
XactLockTableInsert(id[0]);
CSNLogSetCommitSeqNo(id[0], (int)childrenxidnum, &id[3], (COMMITSEQNO_COMMIT_INPROGRESS | id[1]));
RecordCommittingCsnInfo(id[0]);
} else {
XactLockTableInsert(id[0]);
CSNLogSetCommitSeqNo(id[0], 0, 0, (COMMITSEQNO_COMMIT_INPROGRESS | id[1]));
RecordCommittingCsnInfo(id[0]);
}
return;
} else {
uint64* id = ((uint64 *)XLogRecGetData(record));
#ifndef ENABLE_MULTIPLE_NODES
CSNLogSetCommitSeqNo(id[0], 0, NULL, COMMITSEQNO_ABORTED);
XactLockTableDelete(id[0]);
(void)RemoveCommittedCsnInfo(id[0]);
#else
if (id[0] == InvalidTransactionId) {
RemoveAllCommittedCsnInfo();
} else {
Assert(t_thrd.proc->workingVersionNum < DISASTER_READ_VERSION_NUM);
CSNLogSetCommitSeqNo(id[0], 0, NULL, COMMITSEQNO_ABORTED);
XactLockTableDelete(id[0]);
(void)RemoveCommittedCsnInfo(id[0]);
}
#endif
return;
}
}
/*
* --------------------------------------------------------------------
* Recovery handling for Rmgr RM_STANDBY_ID
*
* These record types will only be created if XLogStandbyInfoActive()
* --------------------------------------------------------------------
*/
void standby_redo(XLogReaderState* record)
{
uint8 info = XLogRecGetInfo(record) & ~XLR_INFO_MASK;
/* Backup blocks are not used in standby records */
Assert(!XLogRecHasAnyBlockRefs(record));
if (info == XLOG_STANDBY_CSN_COMMITTING || info == XLOG_STANDBY_CSN_ABORTED) {
DealCSNLogForHotStby(record, info);
return;
}
/* Do nothing if we're not in hot standby mode */
if (t_thrd.xlog_cxt.standbyState == STANDBY_DISABLED)
return;
if (info == XLOG_STANDBY_LOCK) {
xl_standby_locks* xlrec = (xl_standby_locks*)XLogRecGetData(record);
int i;
for (i = 0; i < xlrec->nlocks; i++)
StandbyAcquireAccessExclusiveLock(xlrec->locks[i].xid, xlrec->locks[i].dbOid, xlrec->locks[i].relOid);
} else if (info == XLOG_RUNNING_XACTS) {
RunningTransactionsData running;
xl_running_xacts* xlrec = (xl_running_xacts*)XLogRecGetData(record);
running.nextXid = xlrec->nextXid;
running.oldestRunningXid = xlrec->oldestRunningXid;
running.latestCompletedXid = xlrec->latestCompletedXid;
ProcArrayApplyRecoveryInfo(&running);
} else if (info == XLOG_STANDBY_CSN) {
TransactionId new_global_xmin = *((TransactionId*)XLogRecGetData(record));
if (!TransactionIdIsValid(t_thrd.xact_cxt.ShmemVariableCache->standbyXmin)) {
t_thrd.xact_cxt.ShmemVariableCache->standbyXmin = new_global_xmin;
}
} else if (info == XLOG_STANDBY_UNLOCK) {
xl_standby_locks* xlrec = (xl_standby_locks*)XLogRecGetData(record);
int i;
for (i = 0; i < xlrec->nlocks; i++)
StandbyReleaseLocks(xlrec->locks[i].xid);
} else
ereport(PANIC, (errmsg("standby_redo: unknown op code %hhu", info)));
}
/*
* Record an enhanced snapshot of running transactions into WAL.
*
* The definitions of RunningTransactionsData and xl_xact_running_xacts are
* similar. We keep them separate because xl_xact_running_xacts is a
* contiguous chunk of memory and never exists fully until it is assembled in
* WAL. The inserted records are marked as not being important for durability,
* to avoid triggering superfluous checkpoint / archiving activity.
*/
static XLogRecPtr LogCurrentRunningXacts(RunningTransactions CurrRunningXacts)
{
xl_running_xacts xlrec;
XLogRecPtr recptr;
TransactionId recentXmin = CurrRunningXacts->globalXmin;
XLogBeginInsert();
XLogRegisterData((char*)(&recentXmin), sizeof(TransactionId));
recptr = XLogInsert(RM_STANDBY_ID, XLOG_STANDBY_CSN);
xlrec.xcnt = CurrRunningXacts->xcnt;
xlrec.subxcnt = CurrRunningXacts->subxcnt;
xlrec.subxid_overflow = CurrRunningXacts->subxid_overflow;
xlrec.nextXid = CurrRunningXacts->nextXid;
xlrec.oldestRunningXid = CurrRunningXacts->oldestRunningXid;
xlrec.latestCompletedXid = CurrRunningXacts->latestCompletedXid;
/* Header */
XLogBeginInsert();
XLogRegisterData((char*)(&xlrec), MinSizeOfXactRunningXacts);
/* array of TransactionIds */
if (xlrec.xcnt > 0)
XLogRegisterData((char*)CurrRunningXacts->xids, (xlrec.xcnt + xlrec.subxcnt) * sizeof(TransactionId));
recptr = XLogInsert(RM_STANDBY_ID, XLOG_RUNNING_XACTS);
if (CurrRunningXacts->subxid_overflow)
ereport(trace_recovery(DEBUG2),
(errmsg("snapshot of %d running transactions overflowed (lsn %X/%X oldest xid %lu latest complete %lu next "
"xid %lu)",
CurrRunningXacts->xcnt, (uint32)(recptr >> 32),
(uint32)recptr, CurrRunningXacts->oldestRunningXid,
CurrRunningXacts->latestCompletedXid,
CurrRunningXacts->nextXid)));
else
ereport(trace_recovery(DEBUG2),
(errmsg(
"snapshot of %d+%d running transaction ids (lsn %X/%X oldest xid %lu latest complete %lu next xid %lu)",
CurrRunningXacts->xcnt, CurrRunningXacts->subxcnt,
(uint32)(recptr >> 32), (uint32)recptr,
CurrRunningXacts->oldestRunningXid,
CurrRunningXacts->latestCompletedXid,
CurrRunningXacts->nextXid)));
/*
* Ensure running_xacts information is synced to disk not too far in the
* future. We don't want to stall anything though (i.e. use XLogFlush()),
* so we let the wal writer do it during normal operation.
* XLogSetAsyncXactLSN() conveniently will mark the LSN as to-be-synced
* and nudge the WALWriter into action if sleeping. Check
* XLogBackgroundFlush() for details why a record might not be flushed
* without it.
*/
XLogSetAsyncXactLSN(recptr);
return recptr;
}
/*
* Log details of the current snapshot to WAL. This allows the snapshot state
* to be reconstructed on the standby and for logical decoding.
*
* This is used for Hot Standby as follows:
*
* We can move directly to STANDBY_SNAPSHOT_READY at startup if we
* start from a shutdown checkpoint because we know nothing was running
* at that time and our recovery snapshot is known empty. In the more
* typical case of an online checkpoint we need to jump through a few
* hoops to get a correct recovery snapshot and this requires a two or
* sometimes a three stage process.
*
* The initial snapshot must contain all running xids and all current
* AccessExclusiveLocks at a point in time on the standby. Assembling
* that information while the server is running requires many and
* various LWLocks, so we choose to derive that information piece by
* piece and then re-assemble that info on the standby. When that
* information is fully assembled we move to STANDBY_SNAPSHOT_READY.
*
* Since locking on the primary when we derive the information is not
* strict, we note that there is a time window between the derivation and
* writing to WAL of the derived information. That allows race conditions
* that we must resolve, since xids and locks may enter or leave the
* snapshot during that window. This creates the issue that an xid or
* lock may start *after* the snapshot has been derived yet *before* the
* snapshot is logged in the running xacts WAL record. We resolve this by
* starting to accumulate changes at a point just prior to when we derive
* the snapshot on the primary, then ignore duplicates when we later apply
* the snapshot from the running xacts record. This is implemented during
* CreateCheckpoint() where we use the logical checkpoint location as
* our starting point and then write the running xacts record immediately
* before writing the main checkpoint WAL record. Since we always start
* up from a checkpoint and are immediately at our starting point, we
* unconditionally move to STANDBY_INITIALIZED. After this point we
* must do 4 things:
* * move shared nextXid forwards as we see new xids
* * extend the clog and subtrans with each new xid
* * keep track of uncommitted known assigned xids
* * keep track of uncommitted AccessExclusiveLocks
*
* When we see a commit/abort we must remove known assigned xids and locks
* from the completing transaction. Attempted removals that cannot locate
* an entry are expected and must not cause an error when we are in state
* STANDBY_INITIALIZED.
* This is implemented in StandbyReleaseLocks() and KnownAssignedXidsRemove().
*
* Later, when we apply the running xact data we must be careful to ignore
* transactions already committed, since those commits raced ahead when
* making WAL entries.
*
* The loose timing also means that locks may be recorded that have a
* zero xid, since xids are removed from procs before locks are removed.
* So we must prune the lock list down to ensure we hold locks only for
* currently running xids, performed by StandbyReleaseOldLocks().
* Zero xids should no longer be possible, but we may be replaying WAL
* from a time when they were possible.
*
* For logical decoding only the running xacts information is needed;
* there's no need to look at the locking information, but it's logged anyway,
* as there's no independent knob to just enable logical decoding. For
* details of how this is used, check snapbuild.c's introductory comment.
*
*
* Returns the RecPtr of the last inserted record.
*/
XLogRecPtr LogStandbySnapshot(void)
{
xl_standby_lock* locks = NULL;
XLogRecPtr recptr;
RunningTransactions running;
int nlocks;
Assert(XLogStandbyInfoActive());
/*
* Get details of any AccessExclusiveLocks being held at the moment.
*/
locks = GetRunningTransactionLocks(&nlocks);
if (nlocks > 0)
LogAccessExclusiveLocks(nlocks, locks);
pfree(locks);
locks = NULL;
/*
* Log details of all in-progress transactions. This should be the last
* record we write, because standby will open up when it sees this.
*/
running = GetRunningTransactionData();
/*
* GetRunningTransactionData() acquired ProcArrayLock, we must release it.
* For Hot Standby this can be done before inserting the WAL record
* because ProcArrayApplyRecoveryInfo() rechecks the commit status using
* the clog. For logical decoding, though, the lock can't be released
* early because the clog might be "in the future" from the POV of the
* historic snapshot. This would allow for situations where we're waiting
* for the end of a transaction listed in the xl_running_xacts record
* which, according to the WAL, has committed before the xl_running_xacts
* record. Fortunately this routine isn't executed frequently, and it's
* only a shared lock.
*/
if (g_instance.attr.attr_storage.wal_level < WAL_LEVEL_LOGICAL)
LWLockRelease(ProcArrayLock);
recptr = LogCurrentRunningXacts(running);
/* Release lock if we kept it longer ... */
if (g_instance.attr.attr_storage.wal_level >= WAL_LEVEL_LOGICAL)
LWLockRelease(ProcArrayLock);
/* GetRunningTransactionData() acquired XidGenLock, we must release it */
LWLockRelease(XidGenLock);
return recptr;
}
/*
* Wholesale logging of AccessExclusiveLocks. Other lock types need not be
* logged, as described in backend/storage/lmgr/README.
*/
static void LogAccessExclusiveLocks(int nlocks, xl_standby_lock* locks)
{
xl_standby_locks xlrec;
xlrec.nlocks = nlocks;
XLogBeginInsert();
XLogRegisterData((char*)&xlrec, MinSizeOfXactStandbyLocks);
XLogRegisterData((char*)locks, nlocks * sizeof(xl_standby_lock));
(void)XLogInsert(RM_STANDBY_ID, XLOG_STANDBY_LOCK);
}
/*
* Individual logging of AccessExclusiveLocks for use during LockAcquire()
*/
void LogAccessExclusiveLock(Oid dbOid, Oid relOid)
{
xl_standby_lock xlrec;
xlrec.xid = GetTopTransactionId();
/*
* Decode the locktag back to the original values, to avoid sending lots
* of empty bytes with every message. See lock.h to check how a locktag
* is defined for LOCKTAG_RELATION
*/
xlrec.dbOid = dbOid;
xlrec.relOid = relOid;
LogAccessExclusiveLocks(1, &xlrec);
}
/*
* Prepare to log an AccessExclusiveLock, for use during LockAcquire()
*/
void LogAccessExclusiveLockPrepare(void)
{
/*
* Ensure that a TransactionId has been assigned to this transaction, for
* two reasons, both related to lock release on the standby. First, we
* must assign an xid so that RecordTransactionCommit() and
* RecordTransactionAbort() do not optimise away the transaction
* completion record which recovery relies upon to release locks. It's a
* hack, but for a corner case not worth adding code for into the main
* commit path. Second, we must assign an xid before the lock is recorded
* in shared memory, otherwise a concurrently executing
* GetRunningTransactionLocks() might see a lock associated with an
* InvalidTransactionId which we later assert cannot happen.
*/
(void)GetTopTransactionId();
}
static void LogReleaseAccessExclusiveLocks(int nlocks, xl_standby_lock* locks)
{
xl_standby_locks xlrec;
xlrec.nlocks = nlocks;
XLogBeginInsert();
XLogRegisterData((char*)&xlrec, MinSizeOfXactStandbyLocks);
XLogRegisterData((char*)locks, nlocks * sizeof(xl_standby_lock));
(void)XLogInsert(RM_STANDBY_ID, XLOG_STANDBY_UNLOCK);
}
void LogReleaseAccessExclusiveLock(TransactionId xid, Oid dbOid, Oid relOid)
{
xl_standby_lock xlrec;
xlrec.xid = xid;
/*
* Decode the locktag back to the original values, to avoid sending lots
* of empty bytes with every message. See lock.h to check how a locktag
* is defined for LOCKTAG_RELATION
*/
xlrec.dbOid = dbOid;
xlrec.relOid = relOid;
LogReleaseAccessExclusiveLocks(1, &xlrec);
}
void StandbyXlogStartup(void)
{
t_thrd.xlog_cxt.committing_csn_list = NIL;
}
void *XLogReleaseAndGetCommittingCsnList()
{
ListCell* l = NULL;
foreach (l, t_thrd.xlog_cxt.committing_csn_list) {
TransactionId* action = (TransactionId*)lfirst(l);
if (log_min_messages <= DEBUG4) {
ereport(LOG, (errmsg("XLogReleaseAndGetCommittingCsnList: action xid:%lu", *action)));
}
CSNLogSetCommitSeqNo(*action, 0, NULL, COMMITSEQNO_ABORTED);
XactLockTableDelete(*action);
}
List* committingCsnList = t_thrd.xlog_cxt.committing_csn_list;
#ifdef ENABLE_MULTIPLE_NODES
list_free(t_thrd.xlog_cxt.committing_csn_list);
committingCsnList = NIL;
#endif
t_thrd.xlog_cxt.committing_csn_list = NIL;
return committingCsnList;
}
void CleanUpMakeCommitAbort(List* committingCsnList)
{
ListCell* l = NULL;
foreach (l, committingCsnList) {
TransactionId* action = (TransactionId*)lfirst(l);
if (log_min_messages <= DEBUG4) {
ereport(LOG, (errmsg("CleanUpMakeCommitAbort: action xid:%lu", *action)));
}
XLogBeginInsert();
XLogRegisterData((char *)action, sizeof(TransactionId));
XLogInsert(RM_STANDBY_ID, XLOG_STANDBY_CSN_ABORTED);
}
MemoryContext oldCtx = NULL;
if (IsExtremeRtoRunning()) {
oldCtx = MemoryContextSwitchTo(g_instance.comm_cxt.predo_cxt.parallelRedoCtx);
}
list_free(committingCsnList);
if (IsExtremeRtoRunning()) {
(void)MemoryContextSwitchTo(oldCtx);
}
}
void StandbyXlogCleanup(void)
{
ListCell* l = NULL;
if (log_min_messages <= DEBUG4) {
ereport(LOG,
(errmsg(
"StandbyXlogCleanup: committing_csn_list %p",
t_thrd.xlog_cxt.committing_csn_list)));
}
foreach (l, t_thrd.xlog_cxt.committing_csn_list) {
TransactionId* action = (TransactionId*)lfirst(l);
if (log_min_messages <= DEBUG4) {
ereport(LOG,
(errmsg("StandbyXlogCleanup: action xid:%lu", *action)));
}
CSNLogSetCommitSeqNo(*action, 0, NULL, COMMITSEQNO_ABORTED);
XactLockTableDelete(*action);
#ifndef ENABLE_MULTIPLE_NODES
XLogBeginInsert();
XLogRegisterData((char *)action, sizeof(TransactionId));
XLogInsert(RM_STANDBY_ID, XLOG_STANDBY_CSN_ABORTED);
#endif
}
#ifdef ENABLE_MULTIPLE_NODES
TransactionId clean_xid = InvalidTransactionId;
if (t_thrd.role == STARTUP && !IS_PGXC_COORDINATOR && t_thrd.proc->workingVersionNum >= DISASTER_READ_VERSION_NUM) {
if (log_min_messages <= DEBUG4) {
ereport(LOG, (errmsg("StandbyXlogCleanup: insert clean xlog")));
}
XLogBeginInsert();
XLogRegisterData((char*)(&clean_xid), sizeof(TransactionId));
XLogInsert(RM_STANDBY_ID, XLOG_STANDBY_CSN_ABORTED);
}
#endif
MemoryContext oldCtx = NULL;
if (IsExtremeRtoRunning()) {
oldCtx = MemoryContextSwitchTo(g_instance.comm_cxt.predo_cxt.parallelRedoCtx);
}
list_free(t_thrd.xlog_cxt.committing_csn_list);
if (IsExtremeRtoRunning()) {
(void)MemoryContextSwitchTo(oldCtx);
}
t_thrd.xlog_cxt.committing_csn_list = NIL;
}
bool StandbySafeRestartpoint(void)
{
#ifdef ENABLE_MULTIPLE_NODES
ListCell* l = NULL;
if (t_thrd.xlog_cxt.committing_csn_list && IS_DISASTER_RECOVER_MODE) {
foreach (l, t_thrd.xlog_cxt.committing_csn_list) {
TransactionId* action = (TransactionId*)lfirst(l);
if (log_min_messages <= DEBUG4) {
ereport(LOG,
(errmsg("StandbySafeRestartpoint: action xid:%lu", *action)));
}
}
return false;
}
#else
if (t_thrd.xlog_cxt.committing_csn_list) {
return false;
}
#endif
return true;
}
static void RecordCommittingCsnInfo(TransactionId xid)
{
MemoryContext oldCtx = NULL;
if (IsExtremeRtoRunning()) {
oldCtx = MemoryContextSwitchTo(g_instance.comm_cxt.predo_cxt.parallelRedoCtx);
}
TransactionId* action = (TransactionId*)palloc(sizeof(TransactionId));
if (log_min_messages <= DEBUG4) {
ereport(LOG,
(errmsg("RecordCommittingCsnInfo: xid:%lu, committing_csn_list:%p",
xid,
t_thrd.xlog_cxt.committing_csn_list)));
}
*action = xid;
t_thrd.xlog_cxt.committing_csn_list = lappend(t_thrd.xlog_cxt.committing_csn_list, action);
if (IsExtremeRtoRunning()) {
(void)MemoryContextSwitchTo(oldCtx);
}
}
bool RemoveCommittedCsnInfo(TransactionId xid)
{
ListCell* l = NULL;
foreach (l, t_thrd.xlog_cxt.committing_csn_list) {
TransactionId* action = (TransactionId*)lfirst(l);
if (*action == xid) {
if (log_min_messages <= DEBUG4) {
ereport(LOG,
(errmsg("RemoveCommittedCsnInfo successfully: input xid:%lu, action xid:%lu",
xid, *action)));
}
MemoryContext oldCtx = NULL;
if (IsExtremeRtoRunning()) {
oldCtx = MemoryContextSwitchTo(g_instance.comm_cxt.predo_cxt.parallelRedoCtx);
}
t_thrd.xlog_cxt.committing_csn_list = list_delete_ptr(t_thrd.xlog_cxt.committing_csn_list, action);
pfree(action);
if (IsExtremeRtoRunning()) {
(void)MemoryContextSwitchTo(oldCtx);
}
return true; /* need not look further */
}
}
return false;
}
void RemoveAllCommittedCsnInfo()
{
ListCell* l = NULL;
foreach (l, t_thrd.xlog_cxt.committing_csn_list) {
TransactionId* action = (TransactionId*)lfirst(l);
if (log_min_messages <= DEBUG4) {
ereport(LOG,
(errmsg("RemoveAllCommittedCsnInfo successfully: action xid:%lu", *action)));
}
CSNLogSetCommitSeqNo(*action, 0, NULL, COMMITSEQNO_ABORTED);
XactLockTableDelete(*action);
}
list_free_deep(t_thrd.xlog_cxt.committing_csn_list);
t_thrd.xlog_cxt.committing_csn_list = NIL;
}