2986 lines
109 KiB
C++
2986 lines
109 KiB
C++
/* -------------------------------------------------------------------------
|
|
*
|
|
* plancache.c
|
|
* Plan cache management.
|
|
*
|
|
* The plan cache manager has two principal responsibilities: deciding when
|
|
* to use a generic plan versus a custom (parameter-value-specific) plan,
|
|
* and tracking whether cached plans need to be invalidated because of schema
|
|
* changes in the objects they depend on.
|
|
*
|
|
* The logic for choosing generic or custom plans is in ChooseCustomPlan,
|
|
* which see for comments.
|
|
*
|
|
* Cache invalidation is driven off sinval events. Any CachedPlanSource
|
|
* that matches the event is marked invalid, as is its generic CachedPlan
|
|
* if it has one. When (and if) the next demand for a cached plan occurs,
|
|
* parse analysis and rewrite is repeated to build a new valid query tree,
|
|
* and then planning is performed as normal. We also force re-analysis and
|
|
* re-planning if the active search_path is different from the previous time.
|
|
*
|
|
* Note that if the sinval was a result of user DDL actions, parse analysis
|
|
* could throw an error, for example if a column referenced by the query is
|
|
* no longer present. Another possibility is for the query's output tupdesc
|
|
* to change (for instance "SELECT *" might expand differently than before).
|
|
* The creator of a cached plan can specify whether it is allowable for the
|
|
* query to change output tupdesc on replan --- if so, it's up to the
|
|
* caller to notice changes and cope with them.
|
|
*
|
|
* Currently, we track exactly the dependencies of plans on relations and
|
|
* user-defined functions. On relcache invalidation events or pg_proc
|
|
* syscache invalidation events, we invalidate just those plans that depend
|
|
* on the particular object being modified. (Note: this scheme assumes
|
|
* that any table modification that requires replanning will generate a
|
|
* relcache inval event.) We also watch for inval events on certain other
|
|
* system catalogs, such as pg_namespace; but for them, our response is
|
|
* just to invalidate all plans. We expect updates on those catalogs to
|
|
* be infrequent enough that more-detailed tracking is not worth the effort.
|
|
*
|
|
*
|
|
* Portions Copyright (c) 1996-2012, PostgreSQL Global Development Group
|
|
* Portions Copyright (c) 1994, Regents of the University of California
|
|
* Portions Copyright (c) 2021, openGauss Contributors
|
|
*
|
|
* IDENTIFICATION
|
|
* src/backend/utils/cache/plancache.c
|
|
*
|
|
* -------------------------------------------------------------------------
|
|
*/
|
|
#include "postgres.h"
|
|
#include "knl/knl_variable.h"
|
|
|
|
#include <limits.h>
|
|
|
|
#include "access/transam.h"
|
|
#include "catalog/namespace.h"
|
|
#include "executor/executor.h"
|
|
#include "executor/lightProxy.h"
|
|
#include "executor/spi.h"
|
|
#include "executor/spi_priv.h"
|
|
#include "nodes/nodeFuncs.h"
|
|
#include "opfusion/opfusion.h"
|
|
#include "optimizer/planmain.h"
|
|
#include "optimizer/prep.h"
|
|
#include "optimizer/planner.h"
|
|
#include "optimizer/bucketpruning.h"
|
|
#include "parser/analyze.h"
|
|
#include "parser/parsetree.h"
|
|
#include "storage/lmgr.h"
|
|
#include "parser/parse_hint.h"
|
|
#include "tcop/pquery.h"
|
|
#include "tcop/utility.h"
|
|
#include "utils/hotkey.h"
|
|
#include "utils/inval.h"
|
|
#include "utils/memutils.h"
|
|
#include "utils/snapmgr.h"
|
|
#include "utils/syscache.h"
|
|
#include "utils/globalplancache.h"
|
|
#include "instruments/instr_unique_sql.h"
|
|
#include "instruments/instr_slow_query.h"
|
|
#include "commands/sqladvisor.h"
|
|
|
|
#ifdef ENABLE_MOT
|
|
#include "storage/mot/jit_exec.h"
|
|
#endif
|
|
|
|
#ifdef PGXC
|
|
#include "commands/prepare.h"
|
|
#include "pgxc/execRemote.h"
|
|
#include "pgxc/pgxc.h"
|
|
|
|
static void drop_datanode_statements(Plan* plannode);
|
|
#endif
|
|
|
|
/*
|
|
* We must skip "overhead" operations that involve database access when the
|
|
* cached plan's subject statement is a transaction control command.
|
|
*/
|
|
#define IsTransactionStmtPlan(plansource) \
|
|
((plansource)->raw_parse_tree && IsA((plansource)->raw_parse_tree, TransactionStmt))
|
|
|
|
static void ReleaseGenericPlan(CachedPlanSource* plansource);
|
|
static bool CheckCachedPlan(CachedPlanSource* plansource);
|
|
static CachedPlan* BuildCachedPlan(CachedPlanSource* plansource, List* qlist, ParamListInfo boundParams,
|
|
bool isBuildingCustomPlan);
|
|
static void ReportReasonForPlanChoose(PlanChooseReason reason);
|
|
static bool ChooseCustomPlan(CachedPlanSource* plansource, ParamListInfo boundParams);
|
|
static bool IsForceCustomplan(CachedPlanSource *plansource);
|
|
static bool IsDeleteLimit(CachedPlanSource* plansource, ParamListInfo boundParams);
|
|
static double cached_plan_cost(CachedPlan* plan);
|
|
static void ScanQueryForLocks(Query* parsetree, bool acquire);
|
|
static bool ScanQueryWalker(Node* node, bool* acquire);
|
|
static TupleDesc PlanCacheComputeResultDesc(List* stmt_list);
|
|
#ifdef ENABLE_MULTIPLE_NODES
|
|
static void GPCCheckStreamPlan(CachedPlanSource *plansource, List* plist);
|
|
static bool check_stream_plan(Plan* plan);
|
|
#endif
|
|
static bool is_upsert_query_with_update_param(Node* raw_parse_tree);
|
|
static void GPCFillPlanCache(CachedPlanSource* plansource, bool isBuildingCustomPlan);
|
|
|
|
bool IsStreamSupport()
|
|
{
|
|
#ifdef ENABLE_MULTIPLE_NODES
|
|
return u_sess->attr.attr_sql.enable_stream_operator;
|
|
#else
|
|
return u_sess->opt_cxt.query_dop > 1;
|
|
#endif
|
|
}
|
|
|
|
/*
|
|
* InitPlanCache: initialize module during InitPostgres.
|
|
*
|
|
* All we need to do is hook into inval.c's callback lists.
|
|
*/
|
|
void InitPlanCache(void)
|
|
{
|
|
CacheRegisterSessionRelcacheCallback(PlanCacheRelCallback, (Datum)0);
|
|
CacheRegisterSessionPartcacheCallback(PlanCacheRelCallback, (Datum)0);
|
|
CacheRegisterSessionSyscacheCallback(PROCOID, PlanCacheFuncCallback, (Datum)0);
|
|
CacheRegisterSessionSyscacheCallback(NAMESPACEOID, PlanCacheSysCallback, (Datum)0);
|
|
CacheRegisterSessionSyscacheCallback(OPEROID, PlanCacheSysCallback, (Datum)0);
|
|
CacheRegisterSessionSyscacheCallback(AMOPOPID, PlanCacheSysCallback, (Datum)0);
|
|
}
|
|
|
|
/* ddl no need to global it */
|
|
static bool IsSupportGPCStmt(const Node* node)
|
|
{
|
|
bool isSupportGPC = false;
|
|
switch (nodeTag(node)) {
|
|
case T_InsertStmt:
|
|
isSupportGPC = !has_no_gpc_hint(((InsertStmt*)node)->hintState);
|
|
break;
|
|
case T_DeleteStmt:
|
|
isSupportGPC = !has_no_gpc_hint(((DeleteStmt*)node)->hintState);
|
|
break;
|
|
case T_UpdateStmt:
|
|
isSupportGPC = !has_no_gpc_hint(((UpdateStmt*)node)->hintState);
|
|
break;
|
|
case T_MergeStmt:
|
|
isSupportGPC = !has_no_gpc_hint(((MergeStmt*)node)->hintState);
|
|
break;
|
|
case T_SelectStmt:
|
|
isSupportGPC = !has_no_gpc_hint(((SelectStmt*)node)->hintState);
|
|
break;
|
|
default:
|
|
isSupportGPC = false;
|
|
break;
|
|
}
|
|
return isSupportGPC;
|
|
}
|
|
|
|
/*
|
|
* CreateCachedPlan: initially create a plan cache entry.
|
|
*
|
|
* Creation of a cached plan is divided into two steps, CreateCachedPlan and
|
|
* CompleteCachedPlan. CreateCachedPlan should be called after running the
|
|
* query through raw_parser, but before doing parse analysis and rewrite;
|
|
* CompleteCachedPlan is called after that. The reason for this arrangement
|
|
* is that it can save one round of copying of the raw parse tree, since
|
|
* the parser will normally scribble on the raw parse tree. Callers would
|
|
* otherwise need to make an extra copy of the parse tree to ensure they
|
|
* still had a clean copy to present at plan cache creation time.
|
|
*
|
|
* All arguments presented to CreateCachedPlan are copied into a memory
|
|
* context created as a child of the call-time CurrentMemoryContext, which
|
|
* should be a reasonably short-lived working context that will go away in
|
|
* event of an error. This ensures that the cached plan data structure will
|
|
* likewise disappear if an error occurs before we have fully constructed it.
|
|
* Once constructed, the cached plan can be made longer-lived, if needed,
|
|
* by calling SaveCachedPlan.
|
|
*
|
|
* raw_parse_tree: output of raw_parser()
|
|
* query_string: original query text
|
|
* commandTag: compile-time-constant tag for query, or NULL if empty query
|
|
*/
|
|
CachedPlanSource* CreateCachedPlan(Node* raw_parse_tree, const char* query_string,
|
|
#ifdef PGXC
|
|
const char* stmt_name,
|
|
#endif
|
|
const char* commandTag, bool enable_spi_gpc)
|
|
{
|
|
CachedPlanSource* plansource = NULL;
|
|
MemoryContext source_context;
|
|
MemoryContext oldcxt;
|
|
|
|
Assert(query_string != NULL); /* required as of 8.4 */
|
|
bool isSupportGPC = raw_parse_tree && IsSupportGPCStmt(raw_parse_tree);
|
|
bool enable_pbe_gpc = ENABLE_GPC && stmt_name != NULL && stmt_name[0] != '\0' && isSupportGPC;
|
|
|
|
if(!enable_pbe_gpc && !(ENABLE_CN_GPC && enable_spi_gpc)) {
|
|
/*
|
|
* Make a dedicated memory context for the CachedPlanSource and its
|
|
* permanent subsidiary data. It's probably not going to be large, but
|
|
* just in case, use the default maxsize parameter. Initially it's a
|
|
* child of the caller's context (which we assume to be transient), so
|
|
* that it will be cleaned up on error.
|
|
*/
|
|
source_context = AllocSetContextCreate(CurrentMemoryContext,
|
|
"CachedPlanSource",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE);
|
|
} else {
|
|
char *contextName = NULL;
|
|
if (enable_pbe_gpc) {
|
|
contextName = "GPCCachedPlanSource";
|
|
}
|
|
if (ENABLE_CN_GPC && enable_spi_gpc) {
|
|
contextName = "SPI_GPCCachedPlanSource";
|
|
}
|
|
source_context = AllocSetContextCreate(GLOBAL_PLANCACHE_MEMCONTEXT,
|
|
contextName,
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE,
|
|
SHARED_CONTEXT);
|
|
ResourceOwnerEnlargeGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner);
|
|
ResourceOwnerRememberGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, source_context);
|
|
}
|
|
|
|
/*
|
|
* Create and fill the CachedPlanSource struct within the new context.
|
|
* Most fields are just left empty for the moment.
|
|
*/
|
|
oldcxt = MemoryContextSwitchTo(source_context);
|
|
|
|
plansource = (CachedPlanSource*)palloc0(sizeof(CachedPlanSource));
|
|
plansource->magic = CACHEDPLANSOURCE_MAGIC;
|
|
plansource->raw_parse_tree = (Node*)copyObject(raw_parse_tree);
|
|
plansource->query_string = pstrdup(query_string);
|
|
plansource->commandTag = commandTag;
|
|
plansource->param_types = NULL;
|
|
plansource->num_params = 0;
|
|
plansource->parserSetup = NULL;
|
|
plansource->parserSetupArg = NULL;
|
|
plansource->cursor_options = 0;
|
|
plansource->rewriteRoleId = InvalidOid;
|
|
plansource->dependsOnRole = false;
|
|
plansource->fixed_result = false;
|
|
plansource->resultDesc = NULL;
|
|
plansource->search_path = NULL;
|
|
plansource->context = source_context;
|
|
#ifdef PGXC
|
|
plansource->stmt_name = (stmt_name ? pstrdup(stmt_name) : NULL);
|
|
plansource->stream_enabled = IsStreamSupport();
|
|
plansource->cplan = NULL;
|
|
plansource->single_exec_node = NULL;
|
|
plansource->is_read_only = false;
|
|
plansource->lightProxyObj = NULL;
|
|
/* Initialize gplan_is_fqs is true, and set to false when find it's not fqs */
|
|
plansource->gplan_is_fqs = true;
|
|
#endif
|
|
|
|
plansource->query_list = NIL;
|
|
plansource->relationOids = NIL;
|
|
plansource->invalItems = NIL;
|
|
plansource->query_context = NULL;
|
|
plansource->gplan = NULL;
|
|
plansource->is_oneshot = false;
|
|
plansource->is_complete = false;
|
|
plansource->is_saved = false;
|
|
plansource->is_valid = false;
|
|
plansource->generation = 0;
|
|
plansource->next_saved = NULL;
|
|
plansource->generic_cost = -1;
|
|
plansource->total_custom_cost = 0;
|
|
plansource->num_custom_plans = 0;
|
|
plansource->opFusionObj = NULL;
|
|
plansource->is_checked_opfusion = false;
|
|
plansource->is_support_gplan = false;
|
|
plansource->spi_signature = {(uint32)-1, 0, (uint32)-1, -1};
|
|
|
|
#ifdef ENABLE_MOT
|
|
plansource->storageEngineType = SE_TYPE_UNSPECIFIED;
|
|
plansource->mot_jit_context = NULL;
|
|
#endif
|
|
|
|
if (enable_pbe_gpc) {
|
|
plansource->gpc.status.ShareInit();
|
|
} else if (ENABLE_CN_GPC && enable_spi_gpc) {
|
|
Assert(u_sess->SPI_cxt._current->plan_id >= 0 && u_sess->SPI_cxt._current->visit_id >= 0);
|
|
plansource->gpc.status.ShareInit();
|
|
plansource->spi_signature.spi_key = u_sess->SPI_cxt._current->spi_hash_key;
|
|
plansource->spi_signature.func_oid = u_sess->SPI_cxt._current->func_oid;
|
|
plansource->spi_signature.spi_id = u_sess->SPI_cxt._current->visit_id;
|
|
plansource->spi_signature.plansource_id = u_sess->SPI_cxt._current->plan_id;
|
|
}
|
|
|
|
MemoryContextSwitchTo(oldcxt);
|
|
GPC_LOG("create plancache", plansource, plansource->stmt_name);
|
|
|
|
return plansource;
|
|
}
|
|
|
|
/*
|
|
* CreateOneShotCachedPlan: initially create a one-shot plan cache entry.
|
|
*
|
|
* This variant of CreateCachedPlan creates a plan cache entry that is meant
|
|
* to be used only once. No data copying occurs: all data structures remain
|
|
* in the caller's memory context (which typically should get cleared after
|
|
* completing execution). The CachedPlanSource struct itself is also created
|
|
* in that context.
|
|
*
|
|
* A one-shot plan cannot be saved or copied, since we make no effort to
|
|
* preserve the raw parse tree unmodified. There is also no support for
|
|
* invalidation, so plan use must be completed in the current transaction,
|
|
* and DDL that might invalidate the querytree_list must be avoided as well.
|
|
*
|
|
* raw_parse_tree: output of raw_parser()
|
|
* query_string: original query text
|
|
* commandTag: compile-time-constant tag for query, or NULL if empty query
|
|
*/
|
|
CachedPlanSource* CreateOneShotCachedPlan(Node* raw_parse_tree, const char* query_string, const char* commandTag)
|
|
{
|
|
CachedPlanSource* plansource = NULL;
|
|
|
|
Assert(query_string != NULL); /* required as of 8.4 */
|
|
|
|
/*
|
|
* Create and fill the CachedPlanSource struct within the caller's memory
|
|
* context. Most fields are just left empty for the moment.
|
|
*/
|
|
plansource = (CachedPlanSource*)palloc0(sizeof(CachedPlanSource));
|
|
plansource->magic = CACHEDPLANSOURCE_MAGIC;
|
|
plansource->raw_parse_tree = raw_parse_tree;
|
|
plansource->query_string = query_string;
|
|
plansource->commandTag = commandTag;
|
|
plansource->param_types = NULL;
|
|
plansource->num_params = 0;
|
|
plansource->parserSetup = NULL;
|
|
plansource->parserSetupArg = NULL;
|
|
plansource->cursor_options = 0;
|
|
plansource->rewriteRoleId = InvalidOid;
|
|
plansource->dependsOnRole = false;
|
|
plansource->fixed_result = false;
|
|
plansource->resultDesc = NULL;
|
|
plansource->search_path = NULL;
|
|
plansource->context = CurrentMemoryContext;
|
|
plansource->query_list = NIL;
|
|
plansource->relationOids = NIL;
|
|
plansource->invalItems = NIL;
|
|
plansource->query_context = NULL;
|
|
plansource->gplan = NULL;
|
|
plansource->is_oneshot = true;
|
|
plansource->is_complete = false;
|
|
plansource->is_saved = false;
|
|
plansource->is_valid = false;
|
|
plansource->generation = 0;
|
|
plansource->next_saved = NULL;
|
|
plansource->generic_cost = -1;
|
|
plansource->total_custom_cost = 0;
|
|
plansource->num_custom_plans = 0;
|
|
plansource->spi_signature = {(uint32)-1, 0, (uint32)-1, -1};
|
|
|
|
#ifdef ENABLE_MOT
|
|
plansource->storageEngineType = SE_TYPE_UNSPECIFIED;
|
|
plansource->mot_jit_context = NULL;
|
|
#endif
|
|
|
|
#ifdef PGXC
|
|
plansource->stream_enabled = IsStreamSupport();
|
|
plansource->cplan = NULL;
|
|
plansource->single_exec_node = NULL;
|
|
plansource->is_read_only = false;
|
|
plansource->lightProxyObj = NULL;
|
|
#endif
|
|
|
|
return plansource;
|
|
}
|
|
|
|
/*
|
|
* CompleteCachedPlan: second step of creating a plan cache entry.
|
|
*
|
|
* Pass in the analyzed-and-rewritten form of the query, as well as the
|
|
* required subsidiary data about parameters and such. All passed values will
|
|
* be copied into the CachedPlanSource's memory, except as specified below.
|
|
* After this is called, GetCachedPlan can be called to obtain a plan, and
|
|
* optionally the CachedPlanSource can be saved using SaveCachedPlan.
|
|
*
|
|
* If querytree_context is not NULL, the querytree_list must be stored in that
|
|
* context (but the other parameters need not be). The querytree_list is not
|
|
* copied, rather the given context is kept as the initial query_context of
|
|
* the CachedPlanSource. (It should have been created as a child of the
|
|
* caller's working memory context, but it will now be reparented to belong
|
|
* to the CachedPlanSource.) The querytree_context is normally the context in
|
|
* which the caller did raw parsing and parse analysis. This approach saves
|
|
* one tree copying step compared to passing NULL, but leaves lots of extra
|
|
* cruft in the query_context, namely whatever extraneous stuff parse analysis
|
|
* created, as well as whatever went unused from the raw parse tree. Using
|
|
* this option is a space-for-time tradeoff that is appropriate if the
|
|
* CachedPlanSource is not expected to survive long.
|
|
*
|
|
* plancache.c cannot know how to copy the data referenced by parserSetupArg,
|
|
* and it would often be inappropriate to do so anyway. When using that
|
|
* option, it is caller's responsibility that the referenced data remains
|
|
* valid for as long as the CachedPlanSource exists.
|
|
*
|
|
* If the CachedPlanSource is a "oneshot" plan, then no querytree copying
|
|
* occurs at all, and querytree_context is ignored; it is caller's
|
|
* responsibility that the passed querytree_list is sufficiently long-lived.
|
|
*
|
|
* plansource: structure returned by CreateCachedPlan
|
|
* querytree_list: analyzed-and-rewritten form of query (list of Query nodes)
|
|
* querytree_context: memory context containing querytree_list,
|
|
* or NULL to copy querytree_list into a fresh context
|
|
* param_types: array of fixed parameter type OIDs, or NULL if none
|
|
* num_params: number of fixed parameters
|
|
* parserSetup: alternate method for handling query parameters
|
|
* parserSetupArg: data to pass to parserSetup
|
|
* cursor_options: options bitmask to pass to planner
|
|
* fixed_result: TRUE to disallow future changes in query's result tupdesc
|
|
*/
|
|
void CompleteCachedPlan(CachedPlanSource* plansource, List* querytree_list, MemoryContext querytree_context,
|
|
Oid* param_types, const char* paramModes, int num_params, ParserSetupHook parserSetup, void* parserSetupArg,
|
|
int cursor_options, bool fixed_result, const char* stmt_name, ExecNodes* single_exec_node, bool is_read_only)
|
|
{
|
|
MemoryContext source_context = plansource->context;
|
|
MemoryContext oldcxt = CurrentMemoryContext;
|
|
|
|
/* Assert caller is doing things in a sane order */
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(!plansource->is_complete);
|
|
|
|
/*
|
|
* If caller supplied a querytree_context, reparent it underneath the
|
|
* CachedPlanSource's context; otherwise, create a suitable context and
|
|
* copy the querytree_list into it. But no data copying should be done
|
|
* for one-shot plans; for those, assume the passed querytree_list is
|
|
* sufficiently long-lived.
|
|
*/
|
|
if (plansource->is_oneshot) {
|
|
querytree_context = CurrentMemoryContext;
|
|
Assert (plansource->gpc.status.IsPrivatePlan());
|
|
} else if (querytree_context != NULL) {
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
MemoryContextSetParent(querytree_context, source_context);
|
|
}
|
|
MemoryContextSwitchTo(querytree_context);
|
|
} else {
|
|
if (!plansource->gpc.status.IsPrivatePlan()) {
|
|
querytree_context = AllocSetContextCreate(source_context,
|
|
"GPCCachedPlanQuery",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE,
|
|
SHARED_CONTEXT);
|
|
} else {
|
|
/* Again, it's a good bet the querytree_context can be small */
|
|
querytree_context = AllocSetContextCreate(source_context,
|
|
"CachedPlanQuery",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE);
|
|
}
|
|
MemoryContextSwitchTo(querytree_context);
|
|
querytree_list = (List*)copyObject(querytree_list);
|
|
}
|
|
|
|
/*
|
|
* Use the planner machinery to extract dependencies. Data is saved in
|
|
* query_context. (We assume that not a lot of extra cruft is created by
|
|
* this call.) We can skip this for one-shot plans, and transaction
|
|
* control commands have no such dependencies anyway.
|
|
*/
|
|
if (!plansource->is_oneshot && !IsTransactionStmtPlan(plansource)) {
|
|
extract_query_dependencies((Node*)querytree_list,
|
|
&plansource->relationOids,
|
|
&plansource->invalItems,
|
|
&plansource->dependsOnRole,
|
|
&plansource->force_custom_plan);
|
|
|
|
/*
|
|
* Also save the current search_path in the query_context. (This
|
|
* should not generate much extra cruft either, since almost certainly
|
|
* the path is already valid.) Again, we don't really need this for
|
|
* one-shot plans; and we *must* skip this for transaction control
|
|
* commands, because this could result in catalog accesses.
|
|
*/
|
|
plansource->search_path = GetOverrideSearchPath(source_context);
|
|
}
|
|
|
|
/* Update RLS info as well. */
|
|
plansource->rewriteRoleId = GetUserId();
|
|
plansource->query_context = querytree_context;
|
|
plansource->query_list = querytree_list;
|
|
|
|
/*
|
|
* Save the final parameter types (or other parameter specification data)
|
|
* into the source_context, as well as our other parameters. Also save
|
|
* the result tuple descriptor.
|
|
*/
|
|
MemoryContextSwitchTo(source_context);
|
|
|
|
errno_t rc;
|
|
if (num_params > 0) {
|
|
plansource->param_types = (Oid*)palloc(num_params * sizeof(Oid));
|
|
rc = memcpy_s(plansource->param_types, num_params * sizeof(Oid), param_types, num_params * sizeof(Oid));
|
|
securec_check(rc, "", "");
|
|
} else
|
|
plansource->param_types = NULL;
|
|
|
|
if (num_params > 0 && paramModes != NULL) {
|
|
plansource->param_modes = (char*)palloc(num_params * sizeof(char));
|
|
rc = memcpy_s(plansource->param_modes, num_params * sizeof(char), paramModes, num_params * sizeof(char));
|
|
securec_check(rc, "", "");
|
|
} else {
|
|
plansource->param_modes = NULL;
|
|
}
|
|
|
|
plansource->num_params = num_params;
|
|
plansource->parserSetup = parserSetup;
|
|
plansource->parserSetupArg = parserSetupArg;
|
|
plansource->cursor_options = cursor_options;
|
|
plansource->fixed_result = fixed_result;
|
|
#ifdef PGXC
|
|
plansource->stmt_name = (*stmt_name ? pstrdup(stmt_name) : NULL);
|
|
#endif
|
|
plansource->resultDesc = PlanCacheComputeResultDesc(querytree_list);
|
|
MemoryContextSwitchTo(querytree_context);
|
|
plansource->single_exec_node = (ExecNodes*)copyObject(single_exec_node);
|
|
|
|
MemoryContextSwitchTo(oldcxt);
|
|
|
|
plansource->is_complete = true;
|
|
plansource->is_valid = true;
|
|
plansource->is_read_only = is_read_only;
|
|
}
|
|
|
|
/*
|
|
* SaveCachedPlan: save a cached plan permanently
|
|
*
|
|
* This function moves the cached plan underneath u_sess->cache_mem_cxt (making
|
|
* it live for the life of the backend, unless explicitly dropped), and adds
|
|
* it to the list of cached plans that are checked for invalidation when an
|
|
* sinval event occurs.
|
|
*
|
|
* This is guaranteed not to throw error, except for the caller-error case
|
|
* of trying to save a one-shot plan. Callers typically depend on that
|
|
* since this is called just before or just after adding a pointer to the
|
|
* CachedPlanSource to some permanent data structure of their own. Up until
|
|
* this is done, a CachedPlanSource is just transient data that will go away
|
|
* automatically on transaction abort.
|
|
*/
|
|
void SaveCachedPlan(CachedPlanSource* plansource)
|
|
{
|
|
/* Assert caller is doing things in a sane order */
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(plansource->is_complete);
|
|
Assert(!plansource->is_saved);
|
|
Assert(plansource->gpc.status.InShareTable() == false);
|
|
/* This seems worth a real test, though */
|
|
if (plansource->is_oneshot)
|
|
ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot save one-shot cached plan")));
|
|
|
|
/*
|
|
* In typical use, this function would be called before generating any
|
|
* plans from the CachedPlanSource. If there is a generic plan, moving it
|
|
* into u_sess->cache_mem_cxt would be pretty risky since it's unclear
|
|
* whether the caller has taken suitable care with making references
|
|
* long-lived. Best thing to do seems to be to discard the plan.
|
|
*/
|
|
ReleaseGenericPlan(plansource);
|
|
|
|
/*
|
|
* Reparent the source memory context under u_sess->cache_mem_cxt so that it
|
|
* will live indefinitely. The query_context follows along since it's
|
|
* already a child of the other one.
|
|
*/
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
MemoryContextSetParent(plansource->context, u_sess->cache_mem_cxt);
|
|
}
|
|
|
|
START_CRIT_SECTION();
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, plansource->context);
|
|
/*
|
|
* Add the entry to the session's global list of cached plans.
|
|
*/
|
|
plansource->next_saved = u_sess->pcache_cxt.first_saved_plan;
|
|
u_sess->pcache_cxt.first_saved_plan = plansource;
|
|
|
|
plansource->is_saved = true;
|
|
END_CRIT_SECTION();
|
|
|
|
}
|
|
|
|
/*
|
|
* DropCachedPlan: destroy a cached plan.
|
|
*
|
|
* Actually this only destroys the CachedPlanSource: any referenced CachedPlan
|
|
* is released, but not destroyed until its refcount goes to zero. That
|
|
* handles the situation where DropCachedPlan is called while the plan is
|
|
* still in use.
|
|
*/
|
|
void DropCachedPlan(CachedPlanSource* plansource)
|
|
{
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
if (ENABLE_GPC && plansource->gpc.status.InShareTable())
|
|
elog(PANIC, "should not drop shared plan");
|
|
|
|
/* If it's been saved, remove it from the list */
|
|
if (plansource->is_saved) {
|
|
if (u_sess->pcache_cxt.first_saved_plan == plansource)
|
|
u_sess->pcache_cxt.first_saved_plan = plansource->next_saved;
|
|
else {
|
|
CachedPlanSource* psrc = NULL;
|
|
|
|
for (psrc = u_sess->pcache_cxt.first_saved_plan; psrc; psrc = psrc->next_saved) {
|
|
if (psrc->next_saved == plansource) {
|
|
psrc->next_saved = plansource->next_saved;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (ENABLE_CN_GPC) {
|
|
if (u_sess->pcache_cxt.ungpc_saved_plan == plansource) {
|
|
u_sess->pcache_cxt.ungpc_saved_plan = plansource->next_saved;
|
|
} else {
|
|
CachedPlanSource* psrc = NULL;
|
|
for (psrc = u_sess->pcache_cxt.ungpc_saved_plan; psrc; psrc = psrc->next_saved) {
|
|
if (psrc->next_saved == plansource) {
|
|
psrc->next_saved = plansource->next_saved;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if (ENABLE_GPC)
|
|
GPC_LOG("BEFORE DROP CACHE PLAN", plansource, plansource->stmt_name);
|
|
plansource->is_saved = false;
|
|
}
|
|
plansource->next_saved = NULL;
|
|
DropCachedPlanInternal(plansource);
|
|
|
|
/* Mark it no longer valid */
|
|
plansource->magic = 0;
|
|
|
|
if (ENABLE_DN_GPC)
|
|
GPC_LOG("DROP CACHE PLAN", plansource, 0);
|
|
if (ENABLE_CN_GPC)
|
|
CN_GPC_LOG("DROP CACHE PLAN", plansource, 0);
|
|
|
|
/*
|
|
* Remove the CachedPlanSource and all subsidiary data (including the
|
|
* query_context if any). But if it's a one-shot we can't free anything.
|
|
*/
|
|
if (!plansource->is_oneshot)
|
|
MemoryContextDelete(plansource->context);
|
|
|
|
}
|
|
|
|
/*
|
|
* ReleaseGenericPlan: release a CachedPlanSource's generic plan, if any.
|
|
*/
|
|
static void ReleaseGenericPlan(CachedPlanSource* plansource)
|
|
{
|
|
/* Be paranoid about the possibility that ReleaseCachedPlan fails */
|
|
if (plansource->gplan || plansource->cplan) {
|
|
CachedPlan* plan = NULL;
|
|
|
|
/* custom plan and generic plan should not both exists */
|
|
Assert(plansource->gplan == NULL || plansource->cplan == NULL);
|
|
|
|
if (plansource->gplan)
|
|
plan = plansource->gplan;
|
|
else
|
|
plan = plansource->cplan;
|
|
|
|
#ifdef PGXC
|
|
/* Drop this plan on remote nodes */
|
|
if (plan != NULL && !plan->isShared() && !u_sess->pcache_cxt.gpc_in_try_store) {
|
|
ListCell* lc = NULL;
|
|
|
|
/* Close any active planned Datanode statements */
|
|
foreach (lc, plan->stmt_list) {
|
|
Node* node = (Node*)lfirst(lc);
|
|
|
|
if (IsA(node, PlannedStmt)) {
|
|
PlannedStmt* ps = (PlannedStmt*)node;
|
|
drop_datanode_statements(ps->planTree);
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
Assert(plan->magic == CACHEDPLAN_MAGIC);
|
|
plansource->gplan = NULL;
|
|
plansource->cplan = NULL;
|
|
ReleaseCachedPlan(plan, false);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* RevalidateCachedQuery: ensure validity of analyzed-and-rewritten query tree.
|
|
*
|
|
* What we do here is re-acquire locks and redo parse analysis if necessary.
|
|
* On return, the query_list is valid and we have sufficient locks to begin
|
|
* planning.
|
|
*
|
|
* If any parse analysis activity is required, the caller's memory context is
|
|
* used for that work.
|
|
*
|
|
* The result value is the transient analyzed-and-rewritten query tree if we
|
|
* had to do re-analysis, and NIL otherwise. (This is returned just to save
|
|
* a tree copying step in a subsequent BuildCachedPlan call.)
|
|
*/
|
|
List* RevalidateCachedQuery(CachedPlanSource* plansource, bool has_lp)
|
|
{
|
|
bool snapshot_set = false;
|
|
Node* rawtree = NULL;
|
|
List* tlist = NIL; /* transient query-tree list */
|
|
List* qlist = NIL; /* permanent query-tree list */
|
|
TupleDesc resultDesc;
|
|
MemoryContext querytree_context;
|
|
MemoryContext oldcxt;
|
|
bool need_reset_singlenode = false;
|
|
/*
|
|
* For one-shot plans, we do not support revalidation checking; it's
|
|
* assumed the query is parsed, planned, and executed in one transaction,
|
|
* so that no lock re-acquisition is necessary.
|
|
*/
|
|
if (plansource->is_oneshot || IsTransactionStmtPlan(plansource)) {
|
|
Assert(plansource->is_valid);
|
|
return NIL;
|
|
}
|
|
/* if is shared plan, we should acquire plan lock before check recreate plan */
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
Assert(plansource->is_valid);
|
|
return NIL;
|
|
}
|
|
/*
|
|
* If there were no parsetrees, we don't need to check the the plan is whether invalid or not cause we do nothing but call
|
|
* NullCommand() in the execute stage.
|
|
*/
|
|
if (plansource->raw_parse_tree == NULL) {
|
|
plansource->is_valid = true;
|
|
return NIL;
|
|
}
|
|
|
|
/*
|
|
* If the query is currently valid, we should have a saved search_path ---
|
|
* check to see if that matches the current environment. If not, we want
|
|
* to force replan.
|
|
*/
|
|
if (plansource->is_valid) {
|
|
Assert(plansource->search_path != NULL);
|
|
if (!OverrideSearchPathMatchesCurrent(plansource->search_path)) {
|
|
/* Invalidate the querytree and generic plan */
|
|
plansource->is_valid = false;
|
|
if (plansource->gplan)
|
|
plansource->gplan->is_valid = false;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If the query rewrite phase had a possible RLS dependency, we must redo
|
|
* it if either the role setting has changed.
|
|
*/
|
|
if (plansource->is_valid && plansource->dependsOnRole && (plansource->rewriteRoleId != GetUserId()))
|
|
plansource->is_valid = false;
|
|
|
|
/*
|
|
* If the query is currently valid, acquire locks on the referenced
|
|
* objects; then check again. We need to do it this way to cover the race
|
|
* condition that an invalidation message arrives before we get the locks.
|
|
*/
|
|
if (plansource->is_valid) {
|
|
AcquirePlannerLocks(plansource->query_list, true);
|
|
|
|
/*
|
|
* By now, if any invalidation has happened, the inval callback
|
|
* functions will have marked the query invalid.
|
|
*/
|
|
if (plansource->is_valid) {
|
|
/* Successfully revalidated and locked the query. */
|
|
return NIL;
|
|
}
|
|
|
|
/* Ooops, the race case happened. Release useless locks. */
|
|
AcquirePlannerLocks(plansource->query_list, false);
|
|
}
|
|
|
|
Assert(!plansource->gpc.status.InShareTable());
|
|
|
|
/*
|
|
* Discard the no-longer-useful query tree. (Note: we don't want to do
|
|
* this any earlier, else we'd not have been able to release locks
|
|
* correctly in the race condition case.)
|
|
*/
|
|
plansource->is_valid = false;
|
|
plansource->query_list = NIL;
|
|
plansource->relationOids = NIL;
|
|
plansource->invalItems = NIL;
|
|
plansource->search_path = NULL;
|
|
if (plansource->single_exec_node) {
|
|
need_reset_singlenode = true;
|
|
}
|
|
/*
|
|
* Free the query_context. We don't really expect MemoryContextDelete to
|
|
* fail, but just in case, make sure the CachedPlanSource is left in a
|
|
* reasonably sane state. (The generic plan won't get unlinked yet, but
|
|
* that's acceptable.)
|
|
*/
|
|
if (plansource->query_context) {
|
|
MemoryContext qcxt = plansource->query_context;
|
|
|
|
plansource->query_context = NULL;
|
|
MemoryContextDelete(qcxt);
|
|
}
|
|
|
|
/*
|
|
* Now re-do parse analysis and rewrite. This not incidentally acquires
|
|
* the locks we need to do planning safely.
|
|
*/
|
|
Assert(plansource->is_complete);
|
|
|
|
/*
|
|
* If a snapshot is already set (the normal case), we can just use that
|
|
* for parsing/planning. But if it isn't, install one. Note: no point in
|
|
* checking whether parse analysis requires a snapshot; utility commands
|
|
* don't have invalidatable plans, so we'd not get here for such a
|
|
* command.
|
|
*/
|
|
snapshot_set = false;
|
|
if (!ActiveSnapshotSet()) {
|
|
PushActiveSnapshot(GetTransactionSnapshot(GTM_LITE_MODE));
|
|
snapshot_set = true;
|
|
}
|
|
|
|
/*
|
|
* Query Tree is about to be rebuilt, reset is_top_unique_sql, otherwise, unique sql id
|
|
* can not be set.
|
|
*/
|
|
if (IS_UNIQUE_SQL_TRACK_TOP)
|
|
SetIsTopUniqueSQL(false);
|
|
|
|
/*
|
|
* Run parse analysis and rule rewriting. The parser tends to scribble on
|
|
* its input, so we must copy the raw parse tree to prevent corruption of
|
|
* the cache.
|
|
*/
|
|
rawtree = (Node*)copyObject(plansource->raw_parse_tree);
|
|
if (plansource->parserSetup != NULL)
|
|
tlist = pg_analyze_and_rewrite_params(
|
|
rawtree, plansource->query_string, plansource->parserSetup, plansource->parserSetupArg);
|
|
else
|
|
tlist =
|
|
pg_analyze_and_rewrite(rawtree, plansource->query_string, plansource->param_types, plansource->num_params);
|
|
|
|
/* Release snapshot if we got one */
|
|
if (snapshot_set)
|
|
PopActiveSnapshot();
|
|
|
|
/*
|
|
* Check or update the result tupdesc. XXX should we use a weaker
|
|
* condition than equalTupleDescs() here?
|
|
*
|
|
* We assume the parameter types didn't change from the first time, so no
|
|
* need to update that.
|
|
*/
|
|
resultDesc = PlanCacheComputeResultDesc(tlist);
|
|
if (resultDesc == NULL && plansource->resultDesc == NULL) {
|
|
/* OK, doesn't return tuples */
|
|
} else if (resultDesc == NULL || plansource->resultDesc == NULL ||
|
|
!equalTupleDescs(resultDesc, plansource->resultDesc)) {
|
|
/* can we give a better error message? */
|
|
if (plansource->fixed_result)
|
|
ereport(ERROR, (errcode(ERRCODE_INVALID_CACHE_PLAN), errmsg("cached plan must not change result type")));
|
|
oldcxt = MemoryContextSwitchTo(plansource->context);
|
|
if (resultDesc)
|
|
resultDesc = CreateTupleDescCopy(resultDesc);
|
|
if (plansource->resultDesc)
|
|
FreeTupleDesc(plansource->resultDesc);
|
|
plansource->resultDesc = resultDesc;
|
|
MemoryContextSwitchTo(oldcxt);
|
|
}
|
|
|
|
if (!plansource->gpc.status.IsPrivatePlan()) {
|
|
querytree_context = AllocSetContextCreate(plansource->context,
|
|
"GPCCachedPlanQuery",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE,
|
|
SHARED_CONTEXT);
|
|
} else {
|
|
/*
|
|
* Allocate new query_context and copy the completed querytree into it.
|
|
* It's transient until we complete the copying and dependency extraction.
|
|
*/
|
|
querytree_context = AllocSetContextCreate(u_sess->top_mem_cxt,
|
|
"CachedPlanQuery",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE);
|
|
}
|
|
|
|
oldcxt = MemoryContextSwitchTo(querytree_context);
|
|
|
|
qlist = (List*)copyObject(tlist);
|
|
|
|
/*
|
|
* Use the planner machinery to extract dependencies. Data is saved in
|
|
* query_context. (We assume that not a lot of extra cruft is created by
|
|
* this call.)
|
|
*/
|
|
extract_query_dependencies((Node*)qlist,
|
|
&plansource->relationOids,
|
|
&plansource->invalItems,
|
|
&plansource->dependsOnRole,
|
|
&plansource->force_custom_plan);
|
|
|
|
/* Update RLS info as well. */
|
|
plansource->rewriteRoleId = GetUserId();
|
|
|
|
/*
|
|
* Also save the current search_path in the query_context. (This should
|
|
* not generate much extra cruft either, since almost certainly the path
|
|
* is already valid.)
|
|
*/
|
|
plansource->search_path = GetOverrideSearchPath(querytree_context);
|
|
|
|
MemoryContextSwitchTo(oldcxt);
|
|
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
/* Now reparent the finished query_context and save the links */
|
|
MemoryContextSetParent(querytree_context, plansource->context);
|
|
}
|
|
|
|
plansource->query_context = querytree_context;
|
|
plansource->query_list = qlist;
|
|
|
|
/* Update ExecNodes for Light CN */
|
|
if (need_reset_singlenode || has_lp) {
|
|
ExecNodes* single_exec_node = NULL;
|
|
|
|
/* should be only one query */
|
|
if (list_length(qlist) == 1) {
|
|
Query* query = (Query*)linitial(qlist);
|
|
single_exec_node = lightProxy::checkLightQuery(query);
|
|
|
|
/* only deal with single node */
|
|
if (single_exec_node != NULL && list_length(single_exec_node->nodeList) +
|
|
list_length(single_exec_node->primarynodelist) > 1) {
|
|
FreeExecNodes(&single_exec_node);
|
|
}
|
|
CleanHotkeyCandidates(true);
|
|
}
|
|
|
|
oldcxt = MemoryContextSwitchTo(querytree_context);
|
|
|
|
/* copy first in case memory error occurs */
|
|
ExecNodes* tmp_en1 = (ExecNodes*)copyObject(single_exec_node);
|
|
plansource->single_exec_node = tmp_en1;
|
|
|
|
MemoryContextSwitchTo(oldcxt);
|
|
}
|
|
|
|
/* clean lightProxyObj if exists */
|
|
if (plansource->lightProxyObj != NULL) {
|
|
lightProxy* lp = (lightProxy*)plansource->lightProxyObj;
|
|
lightProxy::tearDown(lp);
|
|
plansource->lightProxyObj = NULL;
|
|
}
|
|
|
|
/* clean opFuisonObj if exists */
|
|
if (plansource->opFusionObj != NULL) {
|
|
OpFusion::tearDown((OpFusion*)plansource->opFusionObj);
|
|
plansource->opFusionObj = NULL;
|
|
}
|
|
|
|
#ifdef ENABLE_MOT
|
|
/* clean JIT context if exists */
|
|
if (plansource->mot_jit_context) {
|
|
JitExec::DestroyJitContext(plansource->mot_jit_context);
|
|
plansource->mot_jit_context = NULL;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Note: we do not reset generic_cost or total_custom_cost, although we
|
|
* could choose to do so. If the DDL or statistics change that prompted
|
|
* the invalidation meant a significant change in the cost estimates, it
|
|
* would be better to reset those variables and start fresh; but often it
|
|
* doesn't, and we're better retaining our hard-won knowledge about the
|
|
* relative costs.
|
|
*/
|
|
|
|
plansource->is_valid = true;
|
|
plansource->gpc.status.SetStatus(GPC_VALID);
|
|
|
|
/* Return transient copy of querytrees for possible use in planning */
|
|
return tlist;
|
|
}
|
|
|
|
/*
|
|
* CheckCachedPlan: see if the CachedPlanSource's generic plan is valid.
|
|
*
|
|
* Caller must have already called RevalidateCachedQuery to verify that the
|
|
* querytree is up to date.
|
|
*
|
|
* On a "true" return, we have acquired the locks needed to run the plan.
|
|
* (We must do this for the "true" result to be race-condition-free.)
|
|
*/
|
|
static bool CheckCachedPlan(CachedPlanSource* plansource)
|
|
{
|
|
CachedPlan* plan = plansource->gplan;
|
|
|
|
/* Assert that caller checked the querytree */
|
|
Assert(plansource->is_valid);
|
|
|
|
/* If there's no generic plan, just say "false" */
|
|
if (plan == NULL) {
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
elog(PANIC, "CheckCachedPlan no gplan for sharedplan %s", plansource->stmt_name);
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/* If stream_operator alreadly change, need build plan again.*/
|
|
if ((!plansource->gpc.status.InShareTable()) && plansource->stream_enabled != IsStreamSupport()) {
|
|
return false;
|
|
}
|
|
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
return true;
|
|
}
|
|
|
|
Assert(plan->magic == CACHEDPLAN_MAGIC);
|
|
/* Generic plans are never one-shot */
|
|
Assert(!plan->is_oneshot);
|
|
|
|
/* If plan isn't valid for current role, we can't use it. */
|
|
if (plan->is_valid && plan->dependsOnRole && plan->planRoleId != GetUserId())
|
|
plan->is_valid = false;
|
|
|
|
/*
|
|
* If it appears valid, acquire locks and recheck; this is much the same
|
|
* logic as in RevalidateCachedQuery, but for a plan.
|
|
*/
|
|
if (plan->is_valid) {
|
|
/*
|
|
* Plan must have positive refcount because it is referenced by
|
|
* plansource; so no need to fear it disappears under us here.
|
|
*/
|
|
Assert(plan->refcount > 0);
|
|
|
|
AcquireExecutorLocks(plan->stmt_list, true);
|
|
|
|
/*
|
|
* If plan was transient, check to see if TransactionXmin has
|
|
* advanced, and if so invalidate it.
|
|
*/
|
|
if (plan->is_valid && TransactionIdIsValid(plan->saved_xmin) &&
|
|
!TransactionIdEquals(plan->saved_xmin, u_sess->utils_cxt.TransactionXmin))
|
|
plan->is_valid = false;
|
|
|
|
/*
|
|
* By now, if any invalidation has happened, the inval callback
|
|
* functions will have marked the plan invalid.
|
|
*/
|
|
if (plan->is_valid) {
|
|
/* Successfully revalidated and locked the query. */
|
|
return true;
|
|
}
|
|
|
|
/* Ooops, the race case happened. Release useless locks. */
|
|
AcquireExecutorLocks(plan->stmt_list, false);
|
|
}
|
|
Assert(!plansource->gpc.status.InShareTable());
|
|
/*
|
|
* Plan has been invalidated, so unlink it from the parent and release it.
|
|
*/
|
|
ReleaseGenericPlan(plansource);
|
|
|
|
return false;
|
|
}
|
|
|
|
static inline void ResetStream(bool outer_is_stream, bool outer_is_stream_support)
|
|
{
|
|
if (IS_PGXC_COORDINATOR) {
|
|
u_sess->opt_cxt.is_stream = outer_is_stream;
|
|
u_sess->opt_cxt.is_stream_support = outer_is_stream_support;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* BuildCachedPlan: construct a new CachedPlan from a CachedPlanSource.
|
|
*
|
|
* qlist should be the result value from a previous RevalidateCachedQuery,
|
|
* or it can be set to NIL if we need to re-copy the plansource's query_list.
|
|
*
|
|
* To build a generic, parameter-value-independent plan, pass NULL for
|
|
* boundParams. To build a custom plan, pass the actual parameter values via
|
|
* boundParams. For best effect, the PARAM_FLAG_CONST flag should be set on
|
|
* each parameter value; otherwise the planner will treat the value as a
|
|
* hint rather than a hard constant.
|
|
*
|
|
* Planning work is done in the caller's memory context. The finished plan
|
|
* is in a child memory context, which typically should get reparented
|
|
* (unless this is a one-shot plan, in which case we don't copy the plan).
|
|
*/
|
|
static CachedPlan* BuildCachedPlan(CachedPlanSource* plansource, List* qlist, ParamListInfo boundParams,
|
|
bool isBuildingCustomPlan)
|
|
{
|
|
CachedPlan* plan = NULL;
|
|
List* plist = NIL;
|
|
bool snapshot_set = false;
|
|
bool spi_pushed = false;
|
|
MemoryContext plan_context;
|
|
MemoryContext oldcxt = CurrentMemoryContext;
|
|
ListCell* lc = NULL;
|
|
bool is_transient = false;
|
|
PLpgSQL_execstate *saved_estate = plpgsql_estate;
|
|
bool outer_is_stream = false;
|
|
bool outer_is_stream_support = false;
|
|
|
|
/*
|
|
* NOTE: GetCachedPlan should have called RevalidateCachedQuery first, so
|
|
* we ought to be holding sufficient locks to prevent any invalidation.
|
|
* However, if we're building a custom plan after having built and
|
|
* rejected a generic plan, it's possible to reach here with is_valid
|
|
* false due to an invalidation while making the generic plan. In theory
|
|
* the invalidation must be a false positive, perhaps a consequence of an
|
|
* sinval reset event or the CLOBBER_CACHE_ALWAYS debug code.
|
|
*
|
|
* We should not call RevalidateCachedQuery here for the above case, for
|
|
* something has already been done when building a generic plan, for example
|
|
* QueryRewriteCTAS has already created the table, and if we call
|
|
* RevalidateCachedQuery again it will report an error 'table already exists'.
|
|
* Also as the comments above, there is also no need to call it.
|
|
*/
|
|
|
|
/*
|
|
* If we don't already have a copy of the querytree list that can be
|
|
* scribbled on by the planner, make one. For a one-shot plan, we assume
|
|
* it's okay to scribble on the original query_list.
|
|
*/
|
|
if (qlist == NIL) {
|
|
if (!plansource->is_oneshot)
|
|
qlist = (List*)copyObject(plansource->query_list);
|
|
else
|
|
qlist = plansource->query_list;
|
|
}
|
|
|
|
/*
|
|
* If a snapshot is already set (the normal case), we can just use that
|
|
* for planning. But if it isn't, and we need one, install one.
|
|
*/
|
|
snapshot_set = false;
|
|
if (!ActiveSnapshotSet() &&
|
|
#ifdef ENABLE_MOT
|
|
!(plansource->storageEngineType == SE_TYPE_MOT) &&
|
|
#endif
|
|
analyze_requires_snapshot(plansource->raw_parse_tree)) {
|
|
PushActiveSnapshot(GetTransactionSnapshot(GTM_LITE_MODE));
|
|
snapshot_set = true;
|
|
}
|
|
|
|
/*
|
|
* The planner may try to call SPI-using functions, which causes a problem
|
|
* if we're already inside one. Rather than expect all SPI-using code to
|
|
* do SPI_push whenever a replan could happen, it seems best to take care
|
|
* of the case here.
|
|
*/
|
|
SPI_STACK_LOG("push cond", NULL, NULL);
|
|
spi_pushed = SPI_push_conditional();
|
|
u_sess->pcache_cxt.query_has_params = (plansource->num_params > 0);
|
|
/*
|
|
* Generate the plan and we temporarily close enable_trigger_shipping as
|
|
* we don't support cached shipping plan for trigger.
|
|
*/
|
|
PG_TRY();
|
|
{
|
|
/* Save stream supported info since it will be reset when generate the plan. */
|
|
outer_is_stream = u_sess->opt_cxt.is_stream;
|
|
outer_is_stream_support = u_sess->opt_cxt.is_stream_support;
|
|
/* opengaussdb:jdbc to create vecplan */
|
|
set_default_stream();
|
|
plist = pg_plan_queries(qlist, plansource->cursor_options, boundParams);
|
|
}
|
|
PG_CATCH();
|
|
{
|
|
ResetStream(outer_is_stream, outer_is_stream_support);
|
|
PG_RE_THROW();
|
|
}
|
|
PG_END_TRY();
|
|
u_sess->pcache_cxt.query_has_params = false;
|
|
|
|
ResetStream(outer_is_stream, outer_is_stream_support);
|
|
|
|
/* Clean up SPI state */
|
|
SPI_STACK_LOG("pop cond", NULL, NULL);
|
|
SPI_pop_conditional(spi_pushed);
|
|
|
|
/* Release snapshot if we got one */
|
|
if (snapshot_set)
|
|
PopActiveSnapshot();
|
|
|
|
/*
|
|
* Normally we make a dedicated memory context for the CachedPlan and its
|
|
* subsidiary data. (It's probably not going to be large, but just in
|
|
* case, use the default maxsize parameter. It's transient for the
|
|
* moment.) But for a one-shot plan, we just leave it in the caller's
|
|
* memory context.
|
|
*/
|
|
if (!plansource->is_oneshot) {
|
|
if (!plansource->gpc.status.IsPrivatePlan()) {
|
|
plan_context = AllocSetContextCreate(GLOBAL_PLANCACHE_MEMCONTEXT,
|
|
"GPCCachedPlan",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE,
|
|
SHARED_CONTEXT);
|
|
} else {
|
|
plan_context = AllocSetContextCreate(u_sess->cache_mem_cxt,
|
|
"CachedPlan",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE);
|
|
}
|
|
|
|
/* We must track shared memory context for handling exception */
|
|
ResourceOwnerEnlargeGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner);
|
|
ResourceOwnerRememberGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, plan_context);
|
|
|
|
/*
|
|
* Copy plan into the new context.
|
|
*/
|
|
MemoryContextSwitchTo(plan_context);
|
|
|
|
plist = (List*)copyObject(plist);
|
|
} else
|
|
plan_context = CurrentMemoryContext;
|
|
|
|
#ifdef PGXC
|
|
/*
|
|
* If this plansource belongs to a named prepared statement, store the stmt
|
|
* name for the Datanode queries.
|
|
*/
|
|
bool is_named_prepare = (IS_PGXC_COORDINATOR && !IsConnFromCoord() && plansource->stmt_name);
|
|
int stmt_num = 0;
|
|
if (is_named_prepare) {
|
|
int n;
|
|
|
|
/*
|
|
* Scan the plans and set the statement field for all found RemoteQuery
|
|
* nodes so they use Datanode statements
|
|
*/
|
|
n = 0;
|
|
foreach (lc, plist) {
|
|
Node* st = NULL;
|
|
PlannedStmt* ps = NULL;
|
|
|
|
st = (Node*)lfirst(lc);
|
|
|
|
if (IsA(st, PlannedStmt)) {
|
|
ps = (PlannedStmt*)st;
|
|
n = SetRemoteStatementName(ps->planTree, plansource->stmt_name, plansource->num_params,
|
|
plansource->param_types, n, isBuildingCustomPlan);
|
|
}
|
|
}
|
|
stmt_num = n;
|
|
}
|
|
#endif
|
|
|
|
/*
|
|
* Create and fill the CachedPlan struct within the new context.
|
|
*/
|
|
plan = (CachedPlan*)palloc(sizeof(CachedPlan));
|
|
plan->magic = CACHEDPLAN_MAGIC;
|
|
plan->stmt_list = plist;
|
|
plan->dn_stmt_num = stmt_num;
|
|
|
|
/*
|
|
* CachedPlan is dependent on role either if RLS affected the rewrite
|
|
* phase or if a role dependency was injected during planning. And it's
|
|
* transient if any plan is marked so.
|
|
*/
|
|
plan->planRoleId = GetUserId();
|
|
plan->dependsOnRole = plansource->dependsOnRole;
|
|
|
|
foreach (lc, plist) {
|
|
PlannedStmt* plannedstmt = (PlannedStmt*)lfirst(lc);
|
|
|
|
if (!IsA(plannedstmt, PlannedStmt))
|
|
continue; /* Ignore utility statements */
|
|
|
|
if (plannedstmt->transientPlan)
|
|
is_transient = true;
|
|
|
|
if (plannedstmt->dependsOnRole)
|
|
plan->dependsOnRole = true;
|
|
}
|
|
|
|
if (is_transient) {
|
|
Assert(TransactionIdIsNormal(u_sess->utils_cxt.TransactionXmin));
|
|
plan->saved_xmin = u_sess->utils_cxt.TransactionXmin;
|
|
} else
|
|
plan->saved_xmin = InvalidTransactionId;
|
|
plan->refcount = 0;
|
|
plan->global_refcount = 0;
|
|
plan->context = plan_context;
|
|
plan->is_oneshot = plansource->is_oneshot;
|
|
plan->is_saved = false;
|
|
plan->is_valid = true;
|
|
|
|
/* assign generation number to new plan */
|
|
plan->generation = ++(plansource->generation);
|
|
|
|
MemoryContextSwitchTo(oldcxt);
|
|
|
|
/* Set plan real u_sess->attr.attr_sql.enable_stream_operator.*/
|
|
plansource->stream_enabled = IsStreamSupport();
|
|
//in shared hash table, we can not share the plan, we should throw error before this logic.
|
|
|
|
plan->is_share = false;
|
|
|
|
if (ENABLE_GPC) {
|
|
#ifdef ENABLE_MULTIPLE_NODES
|
|
GPCCheckStreamPlan(plansource, plist);
|
|
#endif
|
|
GPCFillPlanCache(plansource, isBuildingCustomPlan);
|
|
}
|
|
|
|
plpgsql_estate = saved_estate;
|
|
|
|
return plan;
|
|
}
|
|
|
|
static void GPCFillPlanCache(CachedPlanSource* plansource, bool isBuildingCustomPlan)
|
|
{
|
|
/* set flag is_support_gplan for plan not shared */
|
|
if (!plansource->gpc.status.InShareTable()) {
|
|
plansource->is_support_gplan = !isBuildingCustomPlan;
|
|
if (ENABLE_CN_GPC)
|
|
GPCReGplan(plansource);
|
|
else if (ENABLE_DN_GPC && plansource->is_support_gplan && plansource->gpc.status.InPrepareStmt()) {
|
|
/* add into first_saved_plan if not in */
|
|
plansource->next_saved = u_sess->pcache_cxt.first_saved_plan;
|
|
u_sess->pcache_cxt.first_saved_plan = plansource;
|
|
plansource->is_saved = true;
|
|
plansource->gpc.status.SetLoc(GPC_SHARE_IN_LOCAL_SAVE_PLAN_LIST);
|
|
}
|
|
}
|
|
if (plansource->gpc.status.IsSharePlan() && !isBuildingCustomPlan) {
|
|
pfree_ext(plansource->gpc.key);
|
|
MemoryContext oldcxt = MemoryContextSwitchTo(plansource->context);
|
|
plansource->gpc.key = (GPCKey*)palloc0(sizeof(GPCKey));
|
|
plansource->gpc.key->query_string = plansource->query_string;
|
|
plansource->gpc.key->query_length = strlen(plansource->query_string);
|
|
plansource->gpc.key->spi_signature = plansource->spi_signature;
|
|
GlobalPlanCache::EnvFill(&plansource->gpc.key->env, plansource->dependsOnRole);
|
|
plansource->gpc.key->env.search_path = plansource->search_path;
|
|
plansource->gpc.key->env.param_types = plansource->param_types;
|
|
plansource->gpc.key->env.num_params = plansource->num_params;
|
|
(void)MemoryContextSwitchTo(oldcxt);
|
|
}
|
|
}
|
|
|
|
#ifdef ENABLE_MULTIPLE_NODES
|
|
/* If is stream plan, do not share it.
|
|
* We need different plannodeid for dn's stream consumer. */
|
|
static void GPCCheckStreamPlan(CachedPlanSource *plansource, List* plist)
|
|
{
|
|
if (!ENABLE_CN_GPC)
|
|
return;
|
|
|
|
if (!plansource->gpc.status.IsSharePlan())
|
|
return;
|
|
|
|
ListCell* lc = NULL;
|
|
bool is_stream_plan = false;
|
|
foreach (lc, plist) {
|
|
PlannedStmt* plannedstmt = (PlannedStmt*)lfirst(lc);
|
|
if (IsA(plannedstmt, PlannedStmt)) {
|
|
if (check_stream_plan(plannedstmt->planTree)) {
|
|
is_stream_plan = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (is_stream_plan)
|
|
plansource->gpc.status.SetKind(GPC_UNSHARED);
|
|
}
|
|
|
|
static bool check_stream_plan(Plan* plan)
|
|
{
|
|
if (plan == NULL)
|
|
return false;
|
|
|
|
if (IsA(plan, RemoteQuery) || IsA(plan, VecRemoteQuery)) {
|
|
RemoteQuery* remote_query = (RemoteQuery*)plan;
|
|
if (remote_query->is_simple)
|
|
return true;
|
|
}
|
|
return check_stream_plan(plan->lefttree) || check_stream_plan(plan->righttree);
|
|
}
|
|
|
|
#endif
|
|
|
|
static bool IsDeleteLimit(CachedPlanSource* plansource, ParamListInfo boundParams)
|
|
{
|
|
bool flag = false;
|
|
if (IS_PGXC_COORDINATOR &&
|
|
plansource->raw_parse_tree != NULL &&
|
|
IsA(plansource->raw_parse_tree, DeleteStmt) &&
|
|
((DeleteStmt *)(plansource->raw_parse_tree))->limitClause != NULL &&
|
|
boundParams != NULL) {
|
|
flag = true;
|
|
}
|
|
return flag;
|
|
}
|
|
|
|
|
|
static bool IsForceCustomplan(CachedPlanSource *plansource) {
|
|
bool flag = false;
|
|
if (plansource->force_custom_plan) {
|
|
if (IS_PGXC_DATANODE)
|
|
flag = true;
|
|
else if (IS_PGXC_COORDINATOR && !plansource->gplan_is_fqs)
|
|
flag = true;
|
|
}
|
|
return flag;
|
|
}
|
|
|
|
static bool contain_ParamRef_walker(Node* node, void* context)
|
|
{
|
|
if (node == NULL) {
|
|
return false;
|
|
}
|
|
|
|
if (IsA(node, ParamRef)) {
|
|
return true;
|
|
}
|
|
|
|
return raw_expression_tree_walker(node, (bool (*)())contain_ParamRef_walker, (void*)context);
|
|
}
|
|
|
|
static bool contain_ParamRef(Node* clause)
|
|
{
|
|
return contain_ParamRef_walker(clause, NULL);
|
|
}
|
|
|
|
static bool is_upsert_query_with_update_param(Node* raw_parse_tree)
|
|
{
|
|
if (raw_parse_tree != NULL && IsA(raw_parse_tree, InsertStmt)) {
|
|
InsertStmt* stmt = (InsertStmt*)raw_parse_tree;
|
|
if (stmt->upsertClause != NULL) {
|
|
if (contain_ParamRef((Node*)(stmt->upsertClause->targetList))) {
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
static bool choose_by_hint(const CachedPlanSource* plansource, bool* choose_cplan)
|
|
{
|
|
if (list_length(plansource->query_list) != 1) {
|
|
return false;
|
|
}
|
|
Query *parse = (Query*)linitial(plansource->query_list);
|
|
HintState *hint = parse->hintState;
|
|
if (hint != NULL && hint->cache_plan_hint != NIL) {
|
|
PlanCacheHint* pchint = (PlanCacheHint*)llast(hint->cache_plan_hint);
|
|
if (pchint == NULL) {
|
|
return false;
|
|
}
|
|
pchint->base.state = HINT_STATE_USED;
|
|
*choose_cplan = pchint->chooseCustomPlan;
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/* report the reason why we choose this plan, corresponding to enum PlanChooseReason */
|
|
static void ReportReasonForPlanChoose(PlanChooseReason reason)
|
|
{
|
|
char* msg[MAX_PLANCHOOSEREASON] =
|
|
{"GPlan, reason: shared plancache need choose gplan.",
|
|
"CPlan, reason: Upsert with update query can't choose gplan.",
|
|
#ifdef ENABLE_MOT
|
|
"GPlan, reason: Don't choose custom plan if using pbe optimization and MOT engine.",
|
|
#endif
|
|
"CPlan, reason: For PBE, such as col=$1+$2, generate cplan.",
|
|
"CPlan, reason: One-shot plans will always be considered custom.",
|
|
"GPlan, reason: No parameters.",
|
|
"GPlan, reason: Transaction control statements.",
|
|
"GPlan, reason: Caller wants to force the decision.",
|
|
"CPlan, reason: Caller wants to force the decision.",
|
|
"CPlan, reason: Contains cstore table.",
|
|
"Choose by hint.",
|
|
"GPlan, reason: Using pbe optimization.",
|
|
"GPlan, reason: Settings force the decision.",
|
|
"CPlan, reason: Settings force the decision.",
|
|
"CPlan, reason: First 5 times using CPlan.",
|
|
"GPlan, reason: GPlan's cost is less than 10% more expensive than average custom plan.",
|
|
"CPlan, reason: Default choosing."};
|
|
int elevel = (log_min_messages <= DEBUG2 && module_logging_is_on(MOD_OPT_CHOICE)) ? NOTICE : DEBUG2;
|
|
ereport(elevel, (errmodule(MOD_OPT_CHOICE), errmsg("[Choosing C/G Plan]: %s", msg[reason])));
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* ChooseCustomPlan: choose whether to use custom or generic plan
|
|
*
|
|
* This defines the policy followed by GetCachedPlan.
|
|
*/
|
|
static bool ChooseCustomPlan(CachedPlanSource* plansource, ParamListInfo boundParams)
|
|
{
|
|
double avg_custom_cost;
|
|
bool ret = false;
|
|
|
|
/*
|
|
* Note: shared plancache need choose gplan, and shared plancache already has shared gplan.
|
|
* DO NOT create cachedplan for shared plancache. Only create cachedplan for plancache not in GPC.
|
|
*/
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
ReportReasonForPlanChoose(SHARED_PLANCACHE);
|
|
return false;
|
|
}
|
|
|
|
/* upsert with update query can't choose gplan */
|
|
if (is_upsert_query_with_update_param(plansource->raw_parse_tree)) {
|
|
ReportReasonForPlanChoose(UPSERT_UPDATE_QUERY);
|
|
return true;
|
|
}
|
|
|
|
#ifdef ENABLE_MOT
|
|
/* Don't choose custom plan if using pbe optimization and MOT engine. */
|
|
if (u_sess->attr.attr_sql.enable_pbe_optimization && IsMOTEngineUsed()) {
|
|
ReportReasonForPlanChoose(PBE_OPT_AND_MOT_ENGINE);
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
/* For PBE, such as col=$1+$2, generate cplan. */
|
|
if (IsDeleteLimit(plansource, boundParams) == true) {
|
|
ReportReasonForPlanChoose(PARAM_EXPR);
|
|
return true;
|
|
}
|
|
|
|
/* One-shot plans will always be considered custom */
|
|
if (plansource->is_oneshot) {
|
|
ReportReasonForPlanChoose(ONE_SHOT_PLAN);
|
|
return true;
|
|
}
|
|
|
|
/* Otherwise, never any point in a custom plan if there's no parameters */
|
|
if (boundParams == NULL) {
|
|
ReportReasonForPlanChoose(NO_BOUND_PARAM);
|
|
return false;
|
|
}
|
|
|
|
/* ... nor for transaction control statements */
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
ReportReasonForPlanChoose(TRANSACTION_STAT);
|
|
return false;
|
|
}
|
|
|
|
/* See if caller wants to force the decision */
|
|
if (plansource->cursor_options & CURSOR_OPT_GENERIC_PLAN) {
|
|
ReportReasonForPlanChoose(CALLER_FORCE_GPLAN);
|
|
return false;
|
|
}
|
|
if (plansource->cursor_options & CURSOR_OPT_CUSTOM_PLAN) {
|
|
ReportReasonForPlanChoose(CALLER_FORCE_CPLAN);
|
|
return true;
|
|
}
|
|
|
|
/* If we contains cstore table, always custom except fqs on CN */
|
|
if (IsForceCustomplan(plansource) == true) {
|
|
ReportReasonForPlanChoose(CSTORE_TABLE);
|
|
return true;
|
|
}
|
|
|
|
/* Make decision according to hint */
|
|
if (choose_by_hint(plansource, &ret)) {
|
|
ReportReasonForPlanChoose(CHOOSE_BY_HINT);
|
|
return ret;
|
|
}
|
|
|
|
/* Let settings force the decision */
|
|
if (unlikely(PLAN_CACHE_MODE_AUTO != u_sess->attr.attr_sql.g_planCacheMode)) {
|
|
if (PLAN_CACHE_MODE_FORCE_GENERIC_PLAN == u_sess->attr.attr_sql.g_planCacheMode) {
|
|
ReportReasonForPlanChoose(SETTING_FORCE_GPLAN);
|
|
return false;
|
|
}
|
|
|
|
if (PLAN_CACHE_MODE_FORCE_CUSTOM_PLAN == u_sess->attr.attr_sql.g_planCacheMode) {
|
|
ReportReasonForPlanChoose(SETTING_FORCE_CPLAN);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
/* Don't choose custom plan if using pbe optimization */
|
|
bool isPbeAndFqs = u_sess->attr.attr_sql.enable_pbe_optimization && plansource->gplan_is_fqs;
|
|
if (isPbeAndFqs) {
|
|
ReportReasonForPlanChoose(PBE_OPTIMIZATION);
|
|
return false;
|
|
}
|
|
|
|
/* Generate custom plans until we have done at least 5 (arbitrary) */
|
|
if (plansource->num_custom_plans < 5) {
|
|
ReportReasonForPlanChoose(TRY_CPLAN);
|
|
return true;
|
|
}
|
|
|
|
avg_custom_cost = plansource->total_custom_cost / plansource->num_custom_plans;
|
|
|
|
/*
|
|
* Prefer generic plan if it's less than 10% more expensive than average
|
|
* custom plan. This threshold is a bit arbitrary; it'd be better if we
|
|
* had some means of comparing planning time to the estimated runtime cost
|
|
* differential.
|
|
*
|
|
* Note that if generic_cost is -1 (indicating we've not yet determined
|
|
* the generic plan cost), we'll always prefer generic at this point.
|
|
*
|
|
* If the cost of generic plan and custom plan are both 0, in case of
|
|
* FQS/remotelimit, we prefer generic plan.
|
|
*/
|
|
if (plansource->generic_cost <= avg_custom_cost * 1.1) {
|
|
ReportReasonForPlanChoose(COST_PREFER_GPLAN);
|
|
return false;
|
|
}
|
|
|
|
ReportReasonForPlanChoose(DEFAULT_CHOOSE);
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* cached_plan_cost: calculate estimated cost of a plan
|
|
*/
|
|
static double cached_plan_cost(CachedPlan* plan)
|
|
{
|
|
double result = 0;
|
|
ListCell* lc = NULL;
|
|
|
|
foreach (lc, plan->stmt_list) {
|
|
PlannedStmt* plannedstmt = (PlannedStmt*)lfirst(lc);
|
|
|
|
if (!IsA(plannedstmt, PlannedStmt))
|
|
continue; /* Ignore utility statements */
|
|
|
|
result += plannedstmt->planTree->total_cost;
|
|
}
|
|
|
|
return result;
|
|
}
|
|
|
|
/*
|
|
* GetCachedPlan: get a cached plan from a CachedPlanSource.
|
|
*
|
|
* This function hides the logic that decides whether to use a generic
|
|
* plan or a custom plan for the given parameters: the caller does not know
|
|
* which it will get.
|
|
*
|
|
* On return, the plan is valid and we have sufficient locks to begin
|
|
* execution.
|
|
*
|
|
* On return, the refcount of the plan has been incremented; a later
|
|
* ReleaseCachedPlan() call is expected. The refcount has been reported
|
|
* to the CurrentResourceOwner if useResOwner is true (note that that must
|
|
* only be true if it's a "saved" CachedPlanSource).
|
|
*
|
|
* Note: if any replanning activity is required, the caller's memory context
|
|
* is used for that work.
|
|
*/
|
|
CachedPlan* GetCachedPlan(CachedPlanSource* plansource, ParamListInfo boundParams, bool useResOwner)
|
|
{
|
|
CachedPlan* plan = NULL;
|
|
List* qlist = NIL;
|
|
bool customplan = false;
|
|
|
|
/* Assert caller is doing things in a sane order */
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(plansource->is_complete);
|
|
/* This seems worth a real test, though */
|
|
if (useResOwner && !plansource->is_saved)
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot apply ResourceOwner to non-saved cached plan")));
|
|
|
|
/* Make sure the querytree list is valid and we have parse-time locks */
|
|
qlist = RevalidateCachedQuery(plansource);
|
|
|
|
/* Decide whether to use a custom plan */
|
|
customplan = ChooseCustomPlan(plansource, boundParams);
|
|
|
|
if (!customplan) {
|
|
if (CheckCachedPlan(plansource)) {
|
|
/* We want a generic plan, and we already have a valid one */
|
|
plan = plansource->gplan;
|
|
Assert(plan->magic == CACHEDPLAN_MAGIC);
|
|
|
|
/* Update soft parse counter for Unique SQL */
|
|
UniqueSQLStatCountSoftParse(1);
|
|
} else {
|
|
/* Whenever plan is rebuild, we need to drop the old one */
|
|
ReleaseGenericPlan(plansource);
|
|
/* Build a new generic plan */
|
|
plan = BuildCachedPlan(plansource, qlist, NULL, customplan);
|
|
Assert(!plan->isShared());
|
|
|
|
|
|
/* Link the new generic plan into the plansource */
|
|
plansource->gplan = plan;
|
|
plan->refcount++;
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, plan->context);
|
|
/* Immediately reparent into appropriate context */
|
|
if (plansource->is_saved) {
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
/* saved plans all live under CacheMemoryContext */
|
|
MemoryContextSetParent(plan->context, u_sess->cache_mem_cxt);
|
|
}
|
|
plan->is_saved = true;
|
|
} else {
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
/* otherwise, it should be a sibling of the plansource */
|
|
MemoryContextSetParent(plan->context,
|
|
MemoryContextGetParent(plansource->context));
|
|
}
|
|
}
|
|
/* Update generic_cost whenever we make a new generic plan */
|
|
plansource->generic_cost = cached_plan_cost(plan);
|
|
|
|
/*
|
|
* Judge if gplan is single-node fqs, if so, we can use it when
|
|
* enable_pbe_optimization is on
|
|
*/
|
|
if (IS_PGXC_COORDINATOR && u_sess->attr.attr_sql.enable_pbe_optimization) {
|
|
List* stmt_list = plan->stmt_list;
|
|
bool plan_is_fqs = false;
|
|
if (list_length(stmt_list) == 1) {
|
|
Node* pstmt = (Node*)linitial(stmt_list);
|
|
if (IsA(pstmt, PlannedStmt)) {
|
|
Plan* topPlan = ((PlannedStmt*)pstmt)->planTree;
|
|
if (IsA(topPlan, RemoteQuery)) {
|
|
RemoteQuery* rq = (RemoteQuery*)topPlan;
|
|
if (rq->exec_nodes &&
|
|
((rq->exec_nodes->nodeList == NULL && rq->exec_nodes->en_expr != NULL) ||
|
|
list_length(rq->exec_nodes->nodeList) == 1))
|
|
plan_is_fqs = true;
|
|
}
|
|
}
|
|
}
|
|
plansource->gplan_is_fqs = plan_is_fqs;
|
|
ereport(
|
|
DEBUG2, (errmodule(MOD_OPT), errmsg("Custom plan is used for \"%s\"", plansource->query_string)));
|
|
}
|
|
|
|
/*
|
|
* If, based on the now-known value of generic_cost, we'd not have
|
|
* chosen to use a generic plan, then forget it and make a custom
|
|
* plan. This is a bit of a wart but is necessary to avoid a
|
|
* glitch in behavior when the custom plans are consistently big
|
|
* winners; at some point we'll experiment with a generic plan and
|
|
* find it's a loser, but we don't want to actually execute that
|
|
* plan.
|
|
*/
|
|
customplan = ChooseCustomPlan(plansource, boundParams);
|
|
|
|
/*
|
|
* If we choose to plan again, we need to re-copy the query_list,
|
|
* since the planner probably scribbled on it. We can force
|
|
* BuildCachedPlan to do that by passing NIL.
|
|
*/
|
|
qlist = NIL;
|
|
}
|
|
}
|
|
|
|
/* In function BuildCachedPlan, we deparse query to obtain sql_statement,
|
|
* If we have generic plan, Sql_statement will have format parameter.
|
|
* We can not send this statement to DN directly. So we should replace format
|
|
* parameter with actual values. But, it will block pbe performance and
|
|
* have side effects, such as unexpected query rewrite. So,we suggest that
|
|
* u_sess->attr.attr_common.max_datanode_for_plan is 0 when testing performance.
|
|
*/
|
|
if (u_sess->attr.attr_common.max_datanode_for_plan > 0 && !customplan && boundParams != NULL &&
|
|
boundParams->params_need_process == true && IS_PGXC_COORDINATOR && !IsConnFromCoord()) {
|
|
if (!plansource->gpc.status.InShareTable()) {
|
|
/* we just replace params with virtual values and do not use plan. */
|
|
CachedPlan* tmp_cplan = BuildCachedPlan(plansource, qlist, boundParams, customplan);
|
|
/* Mark it no longer valid */
|
|
tmp_cplan->magic = 0;
|
|
/* One-shot plans do not own their context, so we can't free them */
|
|
if (!tmp_cplan->is_oneshot) {
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, tmp_cplan->context);
|
|
MemoryContextDelete(tmp_cplan->context);
|
|
}
|
|
} else {
|
|
/* for shared plan , just copy a plansource for explain with values */
|
|
CachedPlanSource* tmp_psrc = CopyCachedPlan(plansource, false);
|
|
CachedPlan* tmp_cplan = BuildCachedPlan(tmp_psrc, NIL, boundParams, customplan);
|
|
tmp_cplan->magic = 0;
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, tmp_cplan->context);
|
|
MemoryContextDelete(tmp_cplan->context);
|
|
tmp_psrc->magic = 0;
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, tmp_psrc->context);
|
|
MemoryContextDelete(tmp_psrc->context);
|
|
}
|
|
}
|
|
|
|
if (customplan) {
|
|
/* Whenever plan is rebuild, we need to drop the old one */
|
|
ReleaseGenericPlan(plansource);
|
|
|
|
/* Build a custom plan */
|
|
plan = BuildCachedPlan(plansource, qlist, boundParams, customplan);
|
|
/* Link the new custome plan into the plansource */
|
|
plansource->cplan = plan;
|
|
plan->refcount++;
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, plan->context);
|
|
|
|
if (plansource->is_saved) {
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
/* saved plans all live under CacheMemoryContext */
|
|
MemoryContextSetParent(plan->context, u_sess->cache_mem_cxt);
|
|
}
|
|
plan->is_saved = true;
|
|
} else {
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
/* otherwise, it should be a sibling of the plansource */
|
|
MemoryContextSetParent(plan->context,
|
|
MemoryContextGetParent(plansource->context));
|
|
}
|
|
}
|
|
|
|
/* Accumulate total costs of custom plans, but 'ware overflow */
|
|
if (plansource->num_custom_plans < INT_MAX) {
|
|
plansource->total_custom_cost += cached_plan_cost(plan);
|
|
plansource->num_custom_plans++;
|
|
}
|
|
|
|
ereport(DEBUG2, (errmodule(MOD_OPT), errmsg("Custom plan is used for \"%s\"", plansource->query_string)));
|
|
} else {
|
|
ereport(DEBUG2, (errmodule(MOD_OPT), errmsg("Generic plan is used for \"%s\"", plansource->query_string)));
|
|
}
|
|
|
|
/*
|
|
* Validate the plansource again
|
|
* The plansource may be invalidated by ResetPlanCache when handling invalid
|
|
* messages in BuildCachedPlan.
|
|
*/
|
|
plansource->is_valid = true;
|
|
|
|
/* Flag the plan as in use by caller */
|
|
if (useResOwner)
|
|
ResourceOwnerEnlargePlanCacheRefs(t_thrd.utils_cxt.CurrentResourceOwner);
|
|
plan->refcount++;
|
|
if (useResOwner)
|
|
ResourceOwnerRememberPlanCacheRef(t_thrd.utils_cxt.CurrentResourceOwner, plan);
|
|
|
|
/*
|
|
* Saved plans should be under u_sess->cache_mem_cxt so they will not go away
|
|
* until their reference count goes to zero. In the generic-plan cases we
|
|
* already took care of that, but for a custom plan, do it as soon as we
|
|
* have created a reference-counted link.
|
|
*/
|
|
if (customplan && plansource->is_saved) {
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
MemoryContextSetParent(plan->context, u_sess->cache_mem_cxt);
|
|
}
|
|
plan->is_saved = true;
|
|
}
|
|
|
|
if (!customplan) {
|
|
setCachedPlanBucketId(plan, boundParams);
|
|
}
|
|
|
|
#ifdef ENABLE_MOT
|
|
/* set plan storageEngineType */
|
|
plan->storageEngineType = plansource->storageEngineType;
|
|
plan->mot_jit_context = plansource->mot_jit_context;
|
|
#endif
|
|
|
|
ListCell *lc;
|
|
foreach (lc, plan->stmt_list) {
|
|
PlannedStmt* plannedstmt = (PlannedStmt*)lfirst(lc);
|
|
|
|
if (!IsA(plannedstmt, PlannedStmt))
|
|
continue; /* Ignore utility statements */
|
|
|
|
check_gtm_free_plan(plannedstmt, u_sess->attr.attr_sql.explain_allow_multinode ? WARNING : ERROR);
|
|
}
|
|
|
|
return plan;
|
|
}
|
|
|
|
/*
|
|
* Find and release all Datanode statements referenced by the plan node and subnodes
|
|
*/
|
|
#ifdef PGXC
|
|
static void drop_datanode_statements(Plan* plannode)
|
|
{
|
|
if (IsA(plannode, RemoteQuery)) {
|
|
RemoteQuery* step = (RemoteQuery*)plannode;
|
|
|
|
if (step->statement)
|
|
DropDatanodeStatement(step->statement);
|
|
} else if (IsA(plannode, ModifyTable)) {
|
|
ModifyTable* mt_plan = (ModifyTable*)plannode;
|
|
/* For ModifyTable plan recurse into each of the plans underneath */
|
|
ListCell* l = NULL;
|
|
foreach (l, mt_plan->plans) {
|
|
Plan* plan = (Plan*)lfirst(l);
|
|
drop_datanode_statements(plan);
|
|
}
|
|
}
|
|
|
|
if (innerPlan(plannode))
|
|
drop_datanode_statements(innerPlan(plannode));
|
|
|
|
if (outerPlan(plannode))
|
|
drop_datanode_statements(outerPlan(plannode));
|
|
}
|
|
#endif
|
|
|
|
void ReleaseSharedCachedPlan(CachedPlan* plan, bool useResOwner)
|
|
{
|
|
if (useResOwner) {
|
|
Assert(plan->is_saved);
|
|
ResourceOwnerForgetPlanCacheRef(t_thrd.utils_cxt.CurrentResourceOwner, plan);
|
|
}
|
|
Assert(plan->global_refcount > 0);
|
|
/* we only delete the plan's context when global plancache is off or the plancache is private */
|
|
if (pg_atomic_sub_fetch_u32((volatile uint32*)&plan->global_refcount, 1) == 0) {
|
|
/* TopTransactionResourceOwner is NULL when thread exit */
|
|
if (t_thrd.utils_cxt.TopTransactionResourceOwner)
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, plan->context);
|
|
/* Mark it no longer valid */
|
|
plan->magic = 0;
|
|
MemoryContextUnSeal(plan->context);
|
|
MemoryContextDelete(plan->context);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* ReleaseCachedPlan: release active use of a cached plan.
|
|
*
|
|
* This decrements the reference count, and frees the plan if the count
|
|
* has thereby gone to zero. If useResOwner is true, it is assumed that
|
|
* the reference count is managed by the CurrentResourceOwner.
|
|
*
|
|
* Note: useResOwner = false is used for releasing references that are in
|
|
* persistent data structures, such as the parent CachedPlanSource or a
|
|
* Portal. Transient references should be protected by a resource owner.
|
|
*/
|
|
void ReleaseCachedPlan(CachedPlan* plan, bool useResOwner)
|
|
{
|
|
Assert(plan != NULL);
|
|
Assert(plan->magic == CACHEDPLAN_MAGIC);
|
|
if (plan->isShared()) {
|
|
ReleaseSharedCachedPlan(plan, useResOwner);
|
|
return;
|
|
}
|
|
if (useResOwner) {
|
|
Assert(plan->is_saved);
|
|
ResourceOwnerForgetPlanCacheRef(t_thrd.utils_cxt.CurrentResourceOwner, plan);
|
|
}
|
|
Assert(plan->refcount > 0);
|
|
plan->refcount--;
|
|
/* we only delete the plan's context when global plancache is off or the plancache is private */
|
|
if (plan->refcount == 0) {
|
|
/* Mark it no longer valid */
|
|
plan->magic = 0;
|
|
|
|
/* One-shot plans do not own their context, so we can't free them */
|
|
if (!plan->is_oneshot) {
|
|
Assert (!plan->isShared());
|
|
/* TopTransactionResourceOwner is NULL when thread exit */
|
|
if (t_thrd.utils_cxt.TopTransactionResourceOwner)
|
|
ResourceOwnerForgetGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, plan->context);
|
|
MemoryContextDelete(plan->context);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* CachedPlanIsSimplyValid: quick check for plan still being valid
|
|
*
|
|
* This function must not be used unless CachedPlanAllowsSimpleValidityCheck
|
|
* previously said it was OK.
|
|
*
|
|
* If the plan is valid, and "owner" is not NULL, record a refcount on
|
|
* the plan in that resowner before returning. It is caller's responsibility
|
|
* to be sure that a refcount is held on any plan that's being actively used.
|
|
* long
|
|
* The code here is unconditionally safe as as the only use of this
|
|
* CachedPlanSource is in connection with the particular CachedPlan pointer
|
|
* that's passed in. If the plansource were being used for other purposes,
|
|
* it's possible that its generic plan could be invalidated and regenerated
|
|
* while the current caller wasn't looking, and then there could be a chance
|
|
* collision of address between this caller's now-stale plan pointer and the
|
|
* actual address of the new generic plan. For current uses, that scenario
|
|
* can't happen; but with a plansource shared across multiple uses, it'd be
|
|
* advisable to also save plan->generation and verify that that still matches.
|
|
*/
|
|
bool CachedPlanIsSimplyValid(CachedPlanSource *plansource, CachedPlan *plan, ResourceOwner owner)
|
|
{
|
|
/*
|
|
* Careful here: since the caller doesn't necessarily hold a refcount on
|
|
* the plan to start with, it's possible that "plan" is a dangling
|
|
* pointer. Don't dereference it until we've verified that it still
|
|
* matches the plansource's gplan (which is either valid or NULL).
|
|
*/
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/*
|
|
* Has cache invalidation fired on this plan? We can check this right
|
|
* away since there are no locks that we'd need to acquire first. Note
|
|
* that here we *do* check plansource->is_valid, so as to force plan
|
|
* rebuild if that's become false.
|
|
*/
|
|
if (!plansource->is_valid || plan != plansource->gplan || !plan->is_valid)
|
|
return false;
|
|
|
|
Assert(plan->magic == CACHEDPLAN_MAGIC);
|
|
|
|
/* Is the search_path still the same as when we made it? */
|
|
Assert(plansource->search_path != NULL);
|
|
if (!OverrideSearchPathMatchesCurrent(plansource->search_path))
|
|
return false;
|
|
|
|
/* It's still good. Bump refcount if requested. */
|
|
if (owner) {
|
|
ResourceOwnerEnlargePlanCacheRefs(owner);
|
|
plan->refcount++;
|
|
ResourceOwnerRememberPlanCacheRef(owner, plan);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* CachedPlanAllowsSimpleValidityCheck: can we use CachedPlanIsSimplyValid?
|
|
*
|
|
* This function, together with CachedPlanIsSimplyValid, provides a fast path
|
|
* for revalidating "simple" generic plans. The core requirement to be simple
|
|
* is that the plan must not require taking any locks, which translates to
|
|
* not touching any tables; this happens to match up well with an important
|
|
* use-case in PL/pgSQL. This function tests whether that's true, along
|
|
* with checking some other corner cases that we'd rather not bother with
|
|
* handling in the fast path. (Note that it's still possible for such a plan
|
|
* to be invalidated, for example due to a change in a function that was
|
|
* inlined into the plan.)
|
|
*
|
|
* If the plan is simply valid, and "owner" is not NULL, record a refcount on
|
|
* the plan in that resowner before returning. It is caller's responsibility
|
|
* to be sure that a refcount is held on any plan that's being actively used.
|
|
*
|
|
* This must only be called on known-valid generic plans (eg, ones just
|
|
* returned by GetCachedPlan). If it returns true, the caller may re-use
|
|
* the cached plan as long as CachedPlanIsSimplyValid returns true; that
|
|
* check is much cheaper than the full revalidation done by GetCachedPlan.
|
|
* Nonetheless, no required checks are omitted.
|
|
*/
|
|
bool CachedPlanAllowsSimpleValidityCheck(CachedPlanSource *plansource, CachedPlan *plan, ResourceOwner owner)
|
|
{
|
|
ListCell *lc;
|
|
|
|
/*
|
|
* Sanity-check that the caller gave us a validated generic plan. Notice
|
|
* that we *don't* assert plansource->is_valid as you might expect; that's
|
|
* because it's possible that that's already false when GetCachedPlan
|
|
* returns, e.g. because ResetPlanCache happened partway through. We
|
|
* should accept the plan as long as plan->is_valid is true, and expect to
|
|
* replan after the next CachedPlanIsSimplyValid call.
|
|
*/
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(plan->magic == CACHEDPLAN_MAGIC);
|
|
Assert(plan->is_valid);
|
|
Assert(plan == plansource->gplan);
|
|
Assert(plansource->search_path != NULL);
|
|
Assert(OverrideSearchPathMatchesCurrent(plansource->search_path));
|
|
|
|
/* We don't support oneshot plans here. */
|
|
if (plansource->is_oneshot)
|
|
return false;
|
|
Assert(!plan->is_oneshot);
|
|
|
|
/*
|
|
* If the plan is dependent on RLS considerations, or it's transient,
|
|
* reject. These things probably can't ever happen for table-free
|
|
* queries, but for safety's sake let's check.
|
|
*/
|
|
|
|
if (plan->dependsOnRole)
|
|
return false;
|
|
if (TransactionIdIsValid(plan->saved_xmin))
|
|
return false;
|
|
|
|
/*
|
|
* Reject if AcquirePlannerLocks would have anything to do. This is
|
|
* simplistic, but there's no need to inquire any more carefully; indeed,
|
|
* for current callers it shouldn't even be possible to hit any of these
|
|
* checks.
|
|
*/
|
|
foreach(lc, plansource->query_list) {
|
|
Query *query = lfirst_node(Query, lc);
|
|
|
|
if (query->commandType == CMD_UTILITY)
|
|
return false;
|
|
if (query->rtable || query->cteList || query->hasSubLinks)
|
|
return false;
|
|
}
|
|
|
|
/*
|
|
* Reject if AcquireExecutorLocks would have anything to do. This is
|
|
* probably unnecessary given the previous check, but let's be safe.
|
|
*/
|
|
foreach(lc, plan->stmt_list)
|
|
{
|
|
PlannedStmt *plannedstmt = lfirst_node(PlannedStmt, lc);
|
|
ListCell *lc2;
|
|
|
|
if (plannedstmt->commandType == CMD_UTILITY)
|
|
return false;
|
|
|
|
/*
|
|
* We have to grovel through the rtable because it's likely to contain
|
|
* an RTE_RESULT relation, rather than being totally empty.
|
|
*/
|
|
foreach(lc2, plannedstmt->rtable)
|
|
{
|
|
RangeTblEntry *rte = (RangeTblEntry *) lfirst(lc2);
|
|
|
|
if (rte->rtekind == RTE_RELATION)
|
|
return false;
|
|
}
|
|
|
|
if (!IsA(plannedstmt->planTree, BaseResult))
|
|
return false;
|
|
}
|
|
|
|
/*
|
|
* Okay, it's simple. Note that what we've primarily established here is
|
|
* that no locks need be taken before checking the plan's is_valid flag.
|
|
*/
|
|
|
|
/* Bump refcount if requested. */
|
|
if (owner) {
|
|
ResourceOwnerEnlargePlanCacheRefs(owner);
|
|
plan->refcount++;
|
|
ResourceOwnerRememberPlanCacheRef(owner, plan);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* CachedPlanSetParentContext: move a CachedPlanSource to a new memory context
|
|
*
|
|
* This can only be applied to unsaved plans; once saved, a plan always
|
|
* lives underneath u_sess->cache_mem_cxt.
|
|
*/
|
|
void CachedPlanSetParentContext(CachedPlanSource* plansource, MemoryContext newcontext)
|
|
{
|
|
/* Assert caller is doing things in a sane order */
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(plansource->is_complete);
|
|
|
|
/* These seem worth real tests, though */
|
|
if (plansource->is_saved)
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot move a saved cached plan to another context")));
|
|
if (plansource->is_oneshot)
|
|
ereport(ERROR,
|
|
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot move a one-shot cached plan to another context")));
|
|
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
/* OK, let the caller keep the plan where he wishes */
|
|
MemoryContextSetParent(plansource->context, newcontext);
|
|
}
|
|
|
|
/*
|
|
* The query_context needs no special handling, since it's a child of
|
|
* plansource->context. But if there's a generic plan, it should be
|
|
* maintained as a sibling of plansource->context.
|
|
*/
|
|
if (plansource->gplan) {
|
|
Assert(plansource->gplan->magic == CACHEDPLAN_MAGIC);
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
MemoryContextSetParent(plansource->gplan->context, newcontext);
|
|
}
|
|
}
|
|
|
|
if (plansource->cplan) {
|
|
Assert(plansource->cplan->magic == CACHEDPLAN_MAGIC);
|
|
if (plansource->gpc.status.IsPrivatePlan()) {
|
|
MemoryContextSetParent(plansource->cplan->context, newcontext);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* CopyCachedPlan: make a copy of a CachedPlanSource
|
|
*
|
|
* This is a convenience routine that does the equivalent of
|
|
* CreateCachedPlan + CompleteCachedPlan, using the data stored in the
|
|
* input CachedPlanSource. The result is therefore "unsaved" (regardless
|
|
* of the state of the source), and we don't copy any generic plan either.
|
|
* The result will be currently valid, or not, the same as the source.
|
|
*/
|
|
CachedPlanSource* CopyCachedPlan(CachedPlanSource* plansource, bool is_share)
|
|
{
|
|
CachedPlanSource* newsource = NULL;
|
|
MemoryContext source_context;
|
|
MemoryContext querytree_context;
|
|
MemoryContext oldcxt;
|
|
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(plansource->is_complete);
|
|
|
|
/*
|
|
* One-shot plans can't be copied, because we haven't taken care that
|
|
* parsing/planning didn't scribble on the raw parse tree or querytrees.
|
|
*/
|
|
if (plansource->is_oneshot)
|
|
ereport(ERROR, (errcode(ERRCODE_FEATURE_NOT_SUPPORTED), errmsg("cannot copy a one-shot cached plan")));
|
|
|
|
if (ENABLE_GPC && is_share == true) {
|
|
source_context = AllocSetContextCreate(GLOBAL_PLANCACHE_MEMCONTEXT,
|
|
"GPCCachedPlanSource",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE,
|
|
SHARED_CONTEXT);
|
|
ResourceOwnerEnlargeGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner);
|
|
ResourceOwnerRememberGMemContext(t_thrd.utils_cxt.TopTransactionResourceOwner, source_context);
|
|
} else {
|
|
source_context = AllocSetContextCreate(u_sess->cache_mem_cxt,
|
|
"CachedPlanSource",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE);
|
|
}
|
|
|
|
|
|
oldcxt = MemoryContextSwitchTo(source_context);
|
|
|
|
newsource = (CachedPlanSource*)palloc0(sizeof(CachedPlanSource));
|
|
newsource->magic = CACHEDPLANSOURCE_MAGIC;
|
|
newsource->raw_parse_tree = (Node*)copyObject(plansource->raw_parse_tree);
|
|
newsource->query_string = pstrdup(plansource->query_string);
|
|
newsource->commandTag = plansource->commandTag;
|
|
if (plansource->num_params > 0) {
|
|
newsource->param_types = (Oid*)palloc(plansource->num_params * sizeof(Oid));
|
|
errno_t rc = memcpy_s(newsource->param_types,
|
|
plansource->num_params * sizeof(Oid),
|
|
plansource->param_types,
|
|
plansource->num_params * sizeof(Oid));
|
|
securec_check(rc, "\0", "\0");
|
|
} else
|
|
newsource->param_types = NULL;
|
|
newsource->num_params = plansource->num_params;
|
|
newsource->parserSetup = plansource->parserSetup;
|
|
newsource->parserSetupArg = plansource->parserSetupArg;
|
|
newsource->cursor_options = plansource->cursor_options;
|
|
newsource->rewriteRoleId = plansource->rewriteRoleId;
|
|
newsource->dependsOnRole = plansource->dependsOnRole;
|
|
newsource->fixed_result = plansource->fixed_result;
|
|
if (plansource->resultDesc)
|
|
newsource->resultDesc = CreateTupleDescCopy(plansource->resultDesc);
|
|
else
|
|
newsource->resultDesc = NULL;
|
|
newsource->context = source_context;
|
|
|
|
if (ENABLE_GPC && is_share == true) {
|
|
querytree_context = AllocSetContextCreate(source_context,
|
|
"GPCCachedPlanQuery",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE,
|
|
SHARED_CONTEXT);
|
|
}
|
|
else {
|
|
querytree_context = AllocSetContextCreate(source_context,
|
|
"CachedPlanQuery",
|
|
ALLOCSET_SMALL_MINSIZE,
|
|
ALLOCSET_SMALL_INITSIZE,
|
|
ALLOCSET_DEFAULT_MAXSIZE);
|
|
}
|
|
MemoryContextSwitchTo(querytree_context);
|
|
newsource->query_list = (List*)copyObject(plansource->query_list);
|
|
newsource->relationOids = (List*)copyObject(plansource->relationOids);
|
|
newsource->invalItems = (List*)copyObject(plansource->invalItems);
|
|
if (plansource->search_path) {
|
|
newsource->search_path = CopyOverrideSearchPath(plansource->search_path);
|
|
}
|
|
newsource->query_context = querytree_context;
|
|
newsource->gplan = NULL;
|
|
|
|
newsource->is_oneshot = false;
|
|
newsource->is_complete = true;
|
|
newsource->is_saved = false;
|
|
newsource->is_valid = plansource->is_valid;
|
|
newsource->generation = plansource->generation;
|
|
newsource->next_saved = NULL;
|
|
|
|
/* We may as well copy any acquired cost knowledge */
|
|
newsource->generic_cost = plansource->generic_cost;
|
|
newsource->total_custom_cost = plansource->total_custom_cost;
|
|
newsource->num_custom_plans = plansource->num_custom_plans;
|
|
newsource->opFusionObj = NULL;
|
|
newsource->is_checked_opfusion = false;
|
|
newsource->spi_signature = plansource->spi_signature;
|
|
newsource->gplan_is_fqs = plansource->gplan_is_fqs;
|
|
|
|
#ifdef ENABLE_MOT
|
|
newsource->storageEngineType = SE_TYPE_UNSPECIFIED;
|
|
newsource->mot_jit_context = NULL;
|
|
#endif
|
|
|
|
#ifdef PGXC
|
|
newsource->stream_enabled = plansource->stream_enabled;
|
|
if (!is_share) {
|
|
plansource->cplan = NULL;
|
|
plansource->single_exec_node = NULL;
|
|
plansource->is_read_only = false;
|
|
plansource->lightProxyObj = NULL;
|
|
}
|
|
#endif
|
|
|
|
MemoryContextSwitchTo(oldcxt);
|
|
if (ENABLE_GPC && is_share)
|
|
GPC_LOG("copy plan when recreate", newsource, 0);
|
|
|
|
return newsource;
|
|
}
|
|
|
|
/*
|
|
* CachedPlanIsValid: test whether the rewritten querytree within a
|
|
* CachedPlanSource is currently valid (that is, not marked as being in need
|
|
* of revalidation).
|
|
*
|
|
* This result is only trustworthy (ie, free from race conditions) if
|
|
* the caller has acquired locks on all the relations used in the plan.
|
|
*/
|
|
bool CachedPlanIsValid(CachedPlanSource* plansource)
|
|
{
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
return plansource->is_valid;
|
|
}
|
|
|
|
/*
|
|
* CachedPlanGetTargetList: return tlist, if any, describing plan's output
|
|
*
|
|
* The result is guaranteed up-to-date. However, it is local storage
|
|
* within the cached plan, and may disappear next time the plan is updated.
|
|
*/
|
|
List* CachedPlanGetTargetList(CachedPlanSource* plansource)
|
|
{
|
|
Node* pstmt = NULL;
|
|
|
|
/* Assert caller is doing things in a sane order */
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
Assert(plansource->is_complete);
|
|
|
|
/*
|
|
* No work needed if statement doesn't return tuples (we assume this
|
|
* feature cannot be changed by an invalidation)
|
|
*/
|
|
if (plansource->resultDesc == NULL)
|
|
return NIL;
|
|
|
|
/* Make sure the querytree list is valid and we have parse-time locks */
|
|
(void)RevalidateCachedQuery(plansource);
|
|
|
|
/* Get the primary statement and find out what it returns */
|
|
pstmt = PortalListGetPrimaryStmt(plansource->query_list);
|
|
|
|
return FetchStatementTargetList(pstmt);
|
|
}
|
|
|
|
/*
|
|
* AcquireExecutorLocks: acquire locks needed for execution of a cached plan;
|
|
* or release them if acquire is false.
|
|
*/
|
|
void AcquireExecutorLocks(List* stmt_list, bool acquire)
|
|
{
|
|
ListCell* lc1 = NULL;
|
|
|
|
foreach (lc1, stmt_list) {
|
|
PlannedStmt* plannedstmt = (PlannedStmt*)lfirst(lc1);
|
|
int rt_index;
|
|
ListCell* lc2 = NULL;
|
|
|
|
Assert(!IsA(plannedstmt, Query));
|
|
if (!IsA(plannedstmt, PlannedStmt)) {
|
|
/*
|
|
* Ignore utility statements, except those (such as EXPLAIN) that
|
|
* contain a parsed-but-not-planned query. Note: it's okay to use
|
|
* ScanQueryForLocks, even though the query hasn't been through
|
|
* rule rewriting, because rewriting doesn't change the query
|
|
* representation.
|
|
*/
|
|
Query* query = UtilityContainsQuery((Node*)plannedstmt);
|
|
|
|
if (query != NULL)
|
|
ScanQueryForLocks(query, acquire);
|
|
continue;
|
|
}
|
|
|
|
rt_index = 0;
|
|
foreach (lc2, plannedstmt->rtable) {
|
|
RangeTblEntry* rte = (RangeTblEntry*)lfirst(lc2);
|
|
LOCKMODE lockmode;
|
|
PlanRowMark* rc = NULL;
|
|
|
|
rt_index++;
|
|
|
|
if (rte->rtekind != RTE_RELATION)
|
|
continue;
|
|
|
|
/*
|
|
* Acquire the appropriate type of lock on each relation OID. Note
|
|
* that we don't actually try to open the rel, and hence will not
|
|
* fail if it's been dropped entirely --- we'll just transiently
|
|
* acquire a non-conflicting lock.
|
|
*/
|
|
if (list_member_int(plannedstmt->resultRelations, rt_index))
|
|
lockmode = RowExclusiveLock;
|
|
else if ((rc = get_plan_rowmark(plannedstmt->rowMarks, rt_index)) != NULL &&
|
|
RowMarkRequiresRowShareLock(rc->markType))
|
|
lockmode = RowShareLock;
|
|
else
|
|
lockmode = AccessShareLock;
|
|
|
|
if (acquire)
|
|
LockRelationOid(rte->relid, lockmode);
|
|
else
|
|
UnlockRelationOid(rte->relid, lockmode);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* AcquirePlannerLocks: acquire locks needed for planning of a querytree list;
|
|
* or release them if acquire is false.
|
|
*
|
|
* Note that we don't actually try to open the relations, and hence will not
|
|
* fail if one has been dropped entirely --- we'll just transiently acquire
|
|
* a non-conflicting lock.
|
|
*/
|
|
void AcquirePlannerLocks(List* stmt_list, bool acquire)
|
|
{
|
|
ListCell* lc = NULL;
|
|
|
|
foreach (lc, stmt_list) {
|
|
Query* query = (Query*)lfirst(lc);
|
|
|
|
Assert(IsA(query, Query));
|
|
|
|
if (query->commandType == CMD_UTILITY) {
|
|
/* Ignore utility statements, unless they contain a Query */
|
|
query = UtilityContainsQuery(query->utilityStmt);
|
|
if (query != NULL)
|
|
ScanQueryForLocks(query, acquire);
|
|
continue;
|
|
}
|
|
|
|
ScanQueryForLocks(query, acquire);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* ScanQueryForLocks: recursively scan one Query for AcquirePlannerLocks.
|
|
*/
|
|
static void ScanQueryForLocks(Query* parsetree, bool acquire)
|
|
{
|
|
ListCell* lc = NULL;
|
|
int rt_index;
|
|
|
|
/* Shouldn't get called on utility commands */
|
|
Assert(parsetree->commandType != CMD_UTILITY);
|
|
|
|
/*
|
|
* First, process RTEs of the current query level.
|
|
*/
|
|
rt_index = 0;
|
|
foreach (lc, parsetree->rtable) {
|
|
RangeTblEntry* rte = (RangeTblEntry*)lfirst(lc);
|
|
LOCKMODE lockmode;
|
|
|
|
rt_index++;
|
|
switch (rte->rtekind) {
|
|
case RTE_RELATION:
|
|
/* Acquire or release the appropriate type of lock */
|
|
if (rt_index == parsetree->resultRelation)
|
|
lockmode = RowExclusiveLock;
|
|
else if (get_parse_rowmark(parsetree, rt_index) != NULL)
|
|
lockmode = RowShareLock;
|
|
else
|
|
lockmode = AccessShareLock;
|
|
if (acquire)
|
|
LockRelationOid(rte->relid, lockmode);
|
|
else
|
|
UnlockRelationOid(rte->relid, lockmode);
|
|
break;
|
|
|
|
case RTE_SUBQUERY:
|
|
/* Recurse into subquery-in-FROM */
|
|
ScanQueryForLocks(rte->subquery, acquire);
|
|
break;
|
|
|
|
default:
|
|
/* ignore other types of RTEs */
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* Recurse into subquery-in-WITH */
|
|
foreach (lc, parsetree->cteList) {
|
|
CommonTableExpr* cte = (CommonTableExpr*)lfirst(lc);
|
|
|
|
ScanQueryForLocks((Query*)cte->ctequery, acquire);
|
|
}
|
|
|
|
/*
|
|
* Recurse into sublink subqueries, too. But we already did the ones in
|
|
* the rtable and cteList.
|
|
*/
|
|
if (parsetree->hasSubLinks) {
|
|
query_tree_walker(parsetree, (bool (*)())ScanQueryWalker, (void*)&acquire, QTW_IGNORE_RC_SUBQUERIES);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Walker to find sublink subqueries for ScanQueryForLocks
|
|
*/
|
|
static bool ScanQueryWalker(Node* node, bool* acquire)
|
|
{
|
|
if (node == NULL)
|
|
return false;
|
|
if (IsA(node, SubLink)) {
|
|
SubLink* sub = (SubLink*)node;
|
|
|
|
/* Do what we came for */
|
|
ScanQueryForLocks((Query*)sub->subselect, *acquire);
|
|
/* Fall through to process lefthand args of SubLink */
|
|
}
|
|
|
|
/*
|
|
* Do NOT recurse into Query nodes, because ScanQueryForLocks already
|
|
* processed subselects of subselects for us.
|
|
*/
|
|
return expression_tree_walker(node, (bool (*)())ScanQueryWalker, (void*)acquire);
|
|
}
|
|
|
|
/*
|
|
* PlanCacheComputeResultDesc: given a list of analyzed-and-rewritten Queries,
|
|
* determine the result tupledesc it will produce. Returns NULL if the
|
|
* execution will not return tuples.
|
|
*
|
|
* Note: the result is created or copied into current memory context.
|
|
*/
|
|
static TupleDesc PlanCacheComputeResultDesc(List* stmt_list)
|
|
{
|
|
Query* query = NULL;
|
|
|
|
switch (ChoosePortalStrategy(stmt_list)) {
|
|
case PORTAL_ONE_SELECT:
|
|
case PORTAL_ONE_MOD_WITH:
|
|
query = (Query*)linitial(stmt_list);
|
|
Assert(IsA(query, Query));
|
|
return ExecCleanTypeFromTL(query->targetList, false);
|
|
|
|
case PORTAL_ONE_RETURNING:
|
|
query = (Query*)PortalListGetPrimaryStmt(stmt_list);
|
|
Assert(IsA(query, Query));
|
|
Assert(query->returningList);
|
|
return ExecCleanTypeFromTL(query->returningList, false);
|
|
|
|
case PORTAL_UTIL_SELECT:
|
|
query = (Query*)linitial(stmt_list);
|
|
Assert(IsA(query, Query));
|
|
Assert(query->utilityStmt);
|
|
return UtilityTupleDescriptor(query->utilityStmt);
|
|
|
|
case PORTAL_MULTI_QUERY:
|
|
/* will not return tuples */
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* CheckRelDependency: go through the plansource's dependency list to see if it
|
|
* depends on the given relation by ID.
|
|
*/
|
|
void
|
|
CheckRelDependency(CachedPlanSource *plansource, Oid relid)
|
|
{
|
|
/*
|
|
* Check the dependency list for the rewritten querytree.
|
|
*/
|
|
if ((relid == InvalidOid) ? (plansource->relationOids != NIL) :
|
|
list_member_oid(plansource->relationOids, relid))
|
|
{
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
CN_GPC_LOG("invalid shared in CheckRelDependency", plansource, 0);
|
|
plansource->gpc.status.SetStatus(GPC_INVALID);
|
|
return;
|
|
} else {
|
|
/* Invalidate the querytree and generic plan */
|
|
CN_GPC_LOG("invalid in CheckRelDependency", plansource, 0);
|
|
plansource->is_valid = false;
|
|
if (plansource->gplan) {
|
|
plansource->gplan->is_valid = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* The generic plan, if any, could have more dependencies than the
|
|
* querytree does, so we have to check it too.
|
|
*/
|
|
if (plansource->gplan && plansource->gplan->is_valid)
|
|
{
|
|
ListCell *lc = NULL;
|
|
|
|
foreach(lc, plansource->gplan->stmt_list)
|
|
{
|
|
PlannedStmt *plannedstmt = (PlannedStmt *) lfirst(lc);
|
|
|
|
Assert(!IsA(plannedstmt, Query));
|
|
if (!IsA(plannedstmt, PlannedStmt))
|
|
continue; /* Ignore utility statements */
|
|
if ((relid == InvalidOid) ? (plannedstmt->relationOids != NIL) :
|
|
list_member_oid(plannedstmt->relationOids, relid))
|
|
{
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
plansource->gpc.status.SetStatus(GPC_INVALID);
|
|
return;
|
|
} else {
|
|
/* Invalidate the generic plan only */
|
|
plansource->gplan->is_valid = false;
|
|
}
|
|
break; /* out of stmt_list scan */
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* CheckInvalItemDependency: go through the plansource's dependency list to see if it
|
|
* depends on the given object by ID.
|
|
*/
|
|
void
|
|
CheckInvalItemDependency(CachedPlanSource *plansource, int cacheid, uint32 hashvalue)
|
|
{
|
|
ListCell *lc = NULL;
|
|
|
|
/*
|
|
* Check the dependency list for the rewritten querytree.
|
|
*/
|
|
foreach(lc, plansource->invalItems)
|
|
{
|
|
PlanInvalItem *item = (PlanInvalItem *) lfirst(lc);
|
|
|
|
if (item->cacheId != cacheid)
|
|
continue;
|
|
if (hashvalue == 0 ||
|
|
item->hashValue == hashvalue)
|
|
{
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
CN_GPC_LOG("invalid shared in CheckInvalItemDependency", plansource, 0);
|
|
plansource->gpc.status.SetStatus(GPC_INVALID);
|
|
return;
|
|
} else {
|
|
/* Invalidate the querytree and generic plan */
|
|
CN_GPC_LOG("invalid in CheckInvalItemDependency", plansource, 0);
|
|
plansource->is_valid = false;
|
|
if (plansource->gplan) {
|
|
plansource->gplan->is_valid = false;
|
|
}
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* The generic plan, if any, could have more dependencies than the
|
|
* querytree does, so we have to check it too.
|
|
*/
|
|
if (plansource->gplan && plansource->gplan->is_valid)
|
|
{
|
|
foreach(lc, plansource->gplan->stmt_list)
|
|
{
|
|
PlannedStmt *plannedstmt = (PlannedStmt *) lfirst(lc);
|
|
ListCell *lc3 = NULL;
|
|
|
|
Assert(!IsA(plannedstmt, Query));
|
|
if (!IsA(plannedstmt, PlannedStmt))
|
|
continue; /* Ignore utility statements */
|
|
foreach(lc3, plannedstmt->invalItems)
|
|
{
|
|
PlanInvalItem *item = (PlanInvalItem *) lfirst(lc3);
|
|
|
|
if (item->cacheId != cacheid)
|
|
continue;
|
|
if (hashvalue == 0 ||
|
|
item->hashValue == hashvalue)
|
|
{
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
plansource->gpc.status.SetStatus(GPC_INVALID);
|
|
return;
|
|
} else {
|
|
/* Invalidate the generic plan only */
|
|
plansource->gplan->is_valid = false;
|
|
}
|
|
break; /* out of invalItems scan */
|
|
}
|
|
}
|
|
if (!plansource->gplan->is_valid)
|
|
break; /* out of stmt_list scan */
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* PlanCacheRelCallback
|
|
* Relcache inval callback function
|
|
*
|
|
* Invalidate all plans mentioning the given rel, or all plans mentioning
|
|
* any rel at all if relid == InvalidOid.
|
|
*/
|
|
void PlanCacheRelCallback(Datum arg, Oid relid)
|
|
{
|
|
CachedPlanSource* plansource = NULL;
|
|
|
|
for (plansource = u_sess->pcache_cxt.first_saved_plan; plansource; plansource = plansource->next_saved) {
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/* No work if it's already invalidated */
|
|
if (!plansource->is_valid)
|
|
continue;
|
|
|
|
/* Never invalidate transaction control commands */
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
continue;
|
|
}
|
|
|
|
CheckRelDependency(plansource, relid);
|
|
}
|
|
|
|
if (ENABLE_CN_GPC) {
|
|
for (plansource = u_sess->pcache_cxt.ungpc_saved_plan; plansource; plansource = plansource->next_saved) {
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/* No work if it's already invalidated */
|
|
if (!plansource->is_valid)
|
|
continue;
|
|
|
|
/* Never invalidate transaction control commands */
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
continue;
|
|
}
|
|
|
|
CheckRelDependency(plansource, relid);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* PlanCacheFuncCallback
|
|
* Syscache inval callback function for PROCOID cache
|
|
*
|
|
* Invalidate all plans mentioning the object with the specified hash value,
|
|
* or all plans mentioning any member of this cache if hashvalue == 0.
|
|
*
|
|
* Note that the coding would support use for multiple caches, but right
|
|
* now only user-defined functions are tracked this way.
|
|
*/
|
|
void PlanCacheFuncCallback(Datum arg, int cacheid, uint32 hashvalue)
|
|
{
|
|
CachedPlanSource* plansource = NULL;
|
|
|
|
for (plansource = u_sess->pcache_cxt.first_saved_plan; plansource; plansource = plansource->next_saved) {
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/* No work if it's already invalidated */
|
|
if (!plansource->is_valid)
|
|
continue;
|
|
|
|
/* Never invalidate transaction control commands */
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
continue;
|
|
}
|
|
CheckInvalItemDependency(plansource, cacheid, hashvalue);
|
|
}
|
|
if (ENABLE_CN_GPC) {
|
|
for (plansource = u_sess->pcache_cxt.ungpc_saved_plan; plansource; plansource = plansource->next_saved) {
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/* No work if it's already invalidated */
|
|
if (!plansource->is_valid)
|
|
continue;
|
|
|
|
/* Never invalidate transaction control commands */
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
continue;
|
|
}
|
|
CheckInvalItemDependency(plansource, cacheid, hashvalue);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* PlanCacheSysCallback
|
|
* Syscache inval callback function for other caches
|
|
*
|
|
* Just invalidate everything...
|
|
*/
|
|
void PlanCacheSysCallback(Datum arg, int cacheid, uint32 hashvalue)
|
|
{
|
|
ResetPlanCache();
|
|
}
|
|
|
|
/*
|
|
* ResetPlanCache: invalidate a cached plans.
|
|
*/
|
|
void
|
|
ResetPlanCache(CachedPlanSource *plansource)
|
|
{
|
|
ListCell *lc = NULL;
|
|
|
|
/*
|
|
* In general there is no point in invalidating utility statements
|
|
* since they have no plans anyway. So invalidate it only if it
|
|
* contains at least one non-utility statement, or contains a utility
|
|
* statement that contains a pre-analyzed query (which could have
|
|
* dependencies.)
|
|
*/
|
|
foreach(lc, plansource->query_list)
|
|
{
|
|
Query *query = (Query *) lfirst(lc);
|
|
|
|
Assert(IsA(query, Query));
|
|
if (query->commandType != CMD_UTILITY ||
|
|
UtilityContainsQuery(query->utilityStmt))
|
|
{
|
|
/* non-utility statement, so invalidate */
|
|
if (plansource->gpc.status.InShareTable()) {
|
|
CN_GPC_LOG("invalid shared in ResetPlanCache", plansource, 0);
|
|
plansource->gpc.status.SetStatus(GPC_INVALID);
|
|
return;
|
|
} else {
|
|
plansource->is_valid = false;
|
|
CN_GPC_LOG("invalid in ResetPlanCache", plansource, 0);
|
|
if (plansource->gplan) {
|
|
plansource->gplan->is_valid = false;
|
|
}
|
|
}
|
|
/* no need to look further */
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* ResetPlanCache: invalidate all cached plans.
|
|
*/
|
|
void ResetPlanCache(void)
|
|
{
|
|
CachedPlanSource* plansource = NULL;
|
|
|
|
for (plansource = u_sess->pcache_cxt.first_saved_plan; plansource; plansource = plansource->next_saved) {
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/* No work if it's already invalidated */
|
|
if (!plansource->is_valid)
|
|
continue;
|
|
|
|
/*
|
|
* We *must not* mark transaction control statements as invalid,
|
|
* particularly not ROLLBACK, because they may need to be executed in
|
|
* aborted transactions when we can't revalidate them (cf bug #5269).
|
|
*/
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
continue;
|
|
}
|
|
|
|
ResetPlanCache(plansource);
|
|
}
|
|
if (ENABLE_CN_GPC) {
|
|
for (plansource = u_sess->pcache_cxt.ungpc_saved_plan; plansource; plansource = plansource->next_saved) {
|
|
Assert(plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
|
|
/* No work if it's already invalidated */
|
|
if (!plansource->is_valid)
|
|
continue;
|
|
|
|
/*
|
|
* We *must not* mark transaction control statements as invalid,
|
|
* particularly not ROLLBACK, because they may need to be executed in
|
|
* aborted transactions when we can't revalidate them (cf bug #5269).
|
|
*/
|
|
if (IsTransactionStmtPlan(plansource)) {
|
|
continue;
|
|
}
|
|
|
|
ResetPlanCache(plansource);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Drop lightProxyObj/gplan/cplan inside CachedPlanSource,
|
|
* and drop prepared statements on DN.
|
|
*/
|
|
void DropCachedPlanInternal(CachedPlanSource* plansource)
|
|
{
|
|
#ifdef ENABLE_MOT
|
|
/* MOT: clean any JIT context */
|
|
if (plansource->mot_jit_context) {
|
|
JitExec::DestroyJitContext(plansource->mot_jit_context);
|
|
plansource->mot_jit_context = NULL;
|
|
}
|
|
#endif
|
|
|
|
if (plansource->lightProxyObj != NULL) {
|
|
/* always use light proxy */
|
|
Assert(plansource->gplan == NULL && plansource->cplan == NULL);
|
|
|
|
lightProxy* lp = (lightProxy*)plansource->lightProxyObj;
|
|
if (lp->m_cplan->stmt_name) {
|
|
lp->m_entry = NULL;
|
|
DropDatanodeStatement(lp->m_cplan->stmt_name);
|
|
}
|
|
if (lp->m_portalName == NULL || lightProxy::locateLightProxy(lp->m_portalName) == NULL) {
|
|
lightProxy::tearDown(lp);
|
|
}
|
|
plansource->lightProxyObj = NULL;
|
|
} else {
|
|
if (!plansource->gpc.status.InShareTable() && plansource->opFusionObj != NULL) {
|
|
OpFusion *opfusion = (OpFusion *)plansource->opFusionObj;
|
|
if (opfusion->m_local.m_portalName == NULL ||
|
|
OpFusion::locateFusion(opfusion->m_local.m_portalName) == NULL) {
|
|
OpFusion::tearDown(opfusion);
|
|
} else {
|
|
opfusion->m_global->m_psrc = NULL;
|
|
}
|
|
plansource->opFusionObj = NULL;
|
|
}
|
|
/* Decrement generic CachePlan's refcount and drop if no longer needed */
|
|
ReleaseGenericPlan(plansource);
|
|
}
|
|
|
|
#ifdef MEMORY_CONTEXT_CHECKING
|
|
CachedPlanSource* cur_plansource = u_sess->pcache_cxt.first_saved_plan;
|
|
while (cur_plansource != NULL) {
|
|
Assert(cur_plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
cur_plansource = cur_plansource->next_saved;
|
|
}
|
|
if (ENABLE_CN_GPC) {
|
|
cur_plansource = u_sess->pcache_cxt.ungpc_saved_plan;
|
|
while (cur_plansource != NULL) {
|
|
Assert(cur_plansource->magic == CACHEDPLANSOURCE_MAGIC);
|
|
cur_plansource = cur_plansource->next_saved;
|
|
}
|
|
}
|
|
#endif
|
|
}
|