Files
openGauss-server/src/gausskernel/optimizer/util/plancat.cpp
2020-12-18 17:27:54 +08:00

1494 lines
58 KiB
C++

/* -------------------------------------------------------------------------
*
* plancat.cpp
* routines for accessing the system catalogs
*
*
* Portions Copyright (c) 2020 Huawei Technologies Co.,Ltd.
* Portions Copyright (c) 1996-2012, PostgreSQL Global Development Group
* Portions Copyright (c) 1994, Regents of the University of California
*
*
* IDENTIFICATION
* src/gausskernel/optimizer/util/plancat.cpp
*
* -------------------------------------------------------------------------
*/
#include "postgres.h"
#include "knl/knl_variable.h"
#include <math.h>
#include "access/genam.h"
#include "access/heapam.h"
#include "access/sysattr.h"
#include "access/transam.h"
#include "catalog/catalog.h"
#include "catalog/pg_partition_fn.h"
#include "catalog/pg_statistic.h"
#include "catalog/heap.h"
#include "catalog/storage_gtt.h"
#include "commands/dbcommands.h"
#include "executor/nodeModifyTable.h"
#include "foreign/fdwapi.h"
#include "miscadmin.h"
#include "nodes/makefuncs.h"
#include "nodes/nodeFuncs.h"
#include "optimizer/autoanalyzer.h"
#include "optimizer/clauses.h"
#include "optimizer/cost.h"
#include "optimizer/plancat.h"
#include "optimizer/predtest.h"
#include "optimizer/prep.h"
#include "optimizer/streamplan.h"
#include "parser/parse_relation.h"
#include "parser/parsetree.h"
#include "rewrite/rewriteManip.h"
#include "storage/bufmgr.h"
#include "utils/acl.h"
#include "utils/lsyscache.h"
#include "utils/partcache.h"
#include "utils/rel.h"
#include "utils/rel_gs.h"
#include "utils/snapmgr.h"
#include "utils/syscache.h"
#ifdef PGXC
#include "pgxc/pgxc.h"
#endif
#define ESTIMATE_PARTITION_NUMBER 10
#define DEFAULT_PAGES_NUM (u_sess->attr.attr_sql.enable_global_stats ? 10 * u_sess->pgxc_cxt.NumDataNodes : 10)
#define DEFAULT_TUPLES_NUM DEFAULT_PAGES_NUM
/* Hook for plugins to get control in get_relation_info() */
THR_LOCAL get_relation_info_hook_type get_relation_info_hook = NULL;
extern void AcceptInvalidationMessages(void);
extern bool check_relation_analyzed(Oid relid);
static int32 get_rel_data_width(Relation rel, int32* attr_widths, bool vectorized = false);
static List* get_relation_constraints(PlannerInfo* root, Oid relationObjectId, RelOptInfo* rel, bool include_notnull);
List* build_index_tlist(PlannerInfo* root, IndexOptInfo* index, Relation heapRelation);
static void setRelStoreInfo(RelOptInfo* relOptInfo, Relation relation);
static void acquireSamplesForPartitionedRelation(
Relation relation, LOCKMODE lmode, RelPageType* samplePages, List** sampledPartitionOids)
{
if (RelationIsPartitioned(relation)) {
if (relation->rd_rel->relkind == RELKIND_RELATION) {
RangePartitionMap* partMap = (RangePartitionMap*)(relation->partMap);
int totalPartitionNumber = getNumberOfRangePartitions(relation);
int partitionNumber = 0;
int nonzeroPartitionNumber = 0;
BlockNumber partPages = 0;
BlockNumber currentPartPages = 0;
Partition part = NULL;
for (partitionNumber = 0; partitionNumber < totalPartitionNumber; partitionNumber++) {
Oid partitionOid = partMap->rangeElements[partitionNumber].partitionOid;
if (!OidIsValid(partitionOid))
continue;
part = partitionOpen(relation, partitionOid, lmode);
currentPartPages = PartitionGetNumberOfBlocks(relation, part);
partitionClose(relation, part, lmode);
// for empty heap, PartitionGetNumberOfBlocks() return 0
if (currentPartPages > 0) {
if (sampledPartitionOids != NULL)
*sampledPartitionOids = lappend_oid(*sampledPartitionOids, partitionOid);
partPages += currentPartPages;
if (++nonzeroPartitionNumber == ESTIMATE_PARTITION_NUMBER) {
break;
}
}
}
// compute the total pages
if (nonzeroPartitionNumber >= 0 && nonzeroPartitionNumber < ESTIMATE_PARTITION_NUMBER) {
*samplePages = partPages;
} else if (nonzeroPartitionNumber == ESTIMATE_PARTITION_NUMBER) {
*samplePages = partPages * (totalPartitionNumber / nonzeroPartitionNumber);
}
}
}
}
/*
* get_relation_info -
* Retrieves catalog information for a given relation.
*
* Given the Oid of the relation, return the following info into fields
* of the RelOptInfo struct:
*
* min_attr lowest valid AttrNumber
* max_attr highest valid AttrNumber
* indexlist list of IndexOptInfos for relation's indexes
* fdwroutine if it's a foreign table, the FDW function pointers
* pages number of pages
* tuples number of tuples
*
* Also, initialize the attr_needed[] and attr_widths[] arrays. In most
* cases these are left as zeroes, but sometimes we need to compute attr
* widths here, and we may as well cache the results for costsize.c.
*
* If inhparent is true, all we need to do is set up the attr arrays:
* the RelOptInfo actually represents the appendrel formed by an inheritance
* tree, and so the parent rel's physical size and index information isn't
* important for it.
*/
void get_relation_info(PlannerInfo* root, Oid relationObjectId, bool inhparent, RelOptInfo* rel)
{
Index varno = rel->relid;
Relation relation;
bool hasindex = false;
List* indexinfos = NIL;
List* sampledPartitionIds = NIL;
/*
* We need not lock the relation since it was already locked, either by
* the rewriter or when expand_inherited_rtentry() added it to the query's
* rangetable.
*/
relation = heap_open(relationObjectId, NoLock);
/* Temporary and unlogged relations are inaccessible during recovery. */
if (!RelationNeedsWAL(relation) && RecoveryInProgress())
ereport(ERROR,
(errmodule(MOD_OPT),
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
errmsg("cannot access temporary or unlogged relations during recovery"))));
rel->min_attr = FirstLowInvalidHeapAttributeNumber + 1;
rel->max_attr = RelationGetNumberOfAttributes(relation);
rel->reltablespace = RelationGetForm(relation)->reltablespace;
AssertEreport(rel->max_attr >= rel->min_attr, MOD_OPT, "Max attribute no is less than the min attribute number.");
rel->attr_needed = (Relids*)palloc0((rel->max_attr - rel->min_attr + 1) * sizeof(Relids));
rel->attr_widths = (int32*)palloc0((rel->max_attr - rel->min_attr + 1) * sizeof(int32));
/*
* Estimate relation size --- unless it's an inheritance parent, in which
* case the size will be computed later in set_append_rel_pathlist, and we
* must leave it zero for now to avoid bollixing the total_table_pages
* calculation.
*/
if (!inhparent)
estimate_rel_size(relation,
rel->attr_widths - rel->min_attr,
&rel->pages,
&rel->tuples,
&rel->allvisfrac,
&sampledPartitionIds);
/* Retrieve the parallel_workers reloption, or -1 if not set. */
rel->rel_parallel_workers = RelationGetParallelWorkers(relation, -1);
/*
* Make list of indexes. Ignore indexes on system catalogs if told to.
* Don't bother with indexes for an inheritance parent, either.
*/
if (inhparent || (u_sess->attr.attr_common.IgnoreSystemIndexes && IsSystemClass(relation->rd_rel)))
hasindex = false;
else
hasindex = relation->rd_rel->relhasindex;
if (hasindex) {
List* indexoidlist = NIL;
ListCell* l = NULL;
LOCKMODE lmode;
indexoidlist = RelationGetIndexList(relation);
/*
* For each index, we get the same type of lock that the executor will
* need, and do not release it. This saves a couple of trips to the
* shared lock manager while not creating any real loss of
* concurrency, because no schema changes could be happening on the
* index while we hold lock on the parent rel, and neither lock type
* blocks any other kind of index operation.
*/
if (rel->relid == (unsigned int)root->parse->resultRelation)
lmode = RowExclusiveLock;
else
lmode = AccessShareLock;
foreach (l, indexoidlist) {
Oid indexoid = lfirst_oid(l);
Relation indexRelation;
Form_pg_index index;
IndexOptInfo* info = NULL;
int ncolumns;
int nkeycolumns;
int i;
/*
* Extract info from the relation descriptor for the index.
*/
indexRelation = index_open(indexoid, lmode);
index = indexRelation->rd_index;
/*
* Ignore invalid indexes, since they can't safely be used for
* queries. Note that this is OK because the data structure we
* are constructing is only used by the planner --- the executor
* still needs to insert into "invalid" indexes, if they're marked
* IndexIsReady.
*/
if (!IndexIsValid(index)) {
index_close(indexRelation, NoLock);
continue;
}
/* Ignore empty index for global temp table */
if (RELATION_IS_GLOBAL_TEMP(indexRelation) &&
!gtt_storage_attached(RelationGetRelid(indexRelation))) {
index_close(indexRelation, NoLock);
continue;
}
/*
* If the index is valid, but cannot yet be used, ignore it; but
* mark the plan we are generating as transient. See
* src/backend/access/heap/README.HOT for discussion.
*/
if (index->indcheckxmin) {
TransactionId xmin = HeapTupleGetRawXmin(indexRelation->rd_indextuple);
if (!TransactionIdPrecedes(xmin, u_sess->utils_cxt.TransactionXmin)) {
/*
* Since the TransactionXmin won't advance immediately(see CalculateLocalLatestSnapshot),
* we need to check CSN for the visibility.
*/
CommitSeqNo csn = TransactionIdGetCommitSeqNo(xmin, true, true, false);
if (csn >= u_sess->utils_cxt.CurrentSnapshot->snapshotcsn) {
root->glob->transientPlan = true;
index_close(indexRelation, NoLock);
continue;
}
}
}
info = makeNode(IndexOptInfo);
info->indexoid = index->indexrelid;
info->reltablespace = RelationGetForm(indexRelation)->reltablespace;
info->rel = rel;
info->ncolumns = ncolumns = index->indnatts;
info->nkeycolumns = nkeycolumns = IndexRelationGetNumberOfKeyAttributes(indexRelation);
info->indexkeys = (int*)palloc(sizeof(int) * ncolumns);
info->indexcollations = (Oid*)palloc(sizeof(Oid) * nkeycolumns);
info->opfamily = (Oid*)palloc(sizeof(Oid) * nkeycolumns);
info->opcintype = (Oid*)palloc(sizeof(Oid) * nkeycolumns);
info->isGlobal = RelationIsGlobalIndex(indexRelation);
for (i = 0; i < ncolumns; i++) {
info->indexkeys[i] = index->indkey.values[i];
}
for (i = 0; i < nkeycolumns; i++) {
info->opfamily[i] = indexRelation->rd_opfamily[i];
info->opcintype[i] = indexRelation->rd_opcintype[i];
info->indexcollations[i] = indexRelation->rd_indcollation[i];
}
info->relam = indexRelation->rd_rel->relam;
info->amcostestimate = indexRelation->rd_am->amcostestimate;
info->canreturn = index_can_return(indexRelation);
info->amcanorderbyop = indexRelation->rd_am->amcanorderbyop;
info->amoptionalkey = indexRelation->rd_am->amoptionalkey;
info->amsearcharray = indexRelation->rd_am->amsearcharray;
info->amsearchnulls = indexRelation->rd_am->amsearchnulls;
info->amcanparallel = indexRelation->rd_rel->relam == BTREE_AM_OID;
info->amhasgettuple = OidIsValid(indexRelation->rd_am->amgettuple);
info->amhasgetbitmap = OidIsValid(indexRelation->rd_am->amgetbitmap);
/*
* Fetch the ordering information for the index, if any.
*/
if (info->relam == BTREE_AM_OID) {
/*
* If it's a btree index, we can use its opfamily OIDs
* directly as the sort ordering opfamily OIDs.
*/
AssertEreport(indexRelation->rd_am->amcanorder, MOD_OPT, "amcanorder is NULL.");
info->sortopfamily = info->opfamily;
info->reverse_sort = (bool*)palloc(sizeof(bool) * nkeycolumns);
info->nulls_first = (bool*)palloc(sizeof(bool) * nkeycolumns);
for (i = 0; i < nkeycolumns; i++) {
int16 opt = indexRelation->rd_indoption[i];
info->reverse_sort[i] = (opt & INDOPTION_DESC) != 0;
info->nulls_first[i] = (opt & INDOPTION_NULLS_FIRST) != 0;
}
} else if (indexRelation->rd_am->amcanorder) {
/*
* Otherwise, identify the corresponding btree opfamilies by
* trying to map this index's "<" operators into btree. Since
* "<" uniquely defines the behavior of a sort order, this is
* a sufficient test.
*
* XXX This method is rather slow and also requires the
* undesirable assumption that the other index AM numbers its
* strategies the same as btree. It'd be better to have a way
* to explicitly declare the corresponding btree opfamily for
* each opfamily of the other index type. But given the lack
* of current or foreseeable amcanorder index types, it's not
* worth expending more effort on now.
*/
info->sortopfamily = (Oid*)palloc(sizeof(Oid) * nkeycolumns);
info->reverse_sort = (bool*)palloc(sizeof(bool) * nkeycolumns);
info->nulls_first = (bool*)palloc(sizeof(bool) * nkeycolumns);
for (i = 0; i < nkeycolumns; i++) {
int16 opt = indexRelation->rd_indoption[i];
Oid ltopr;
Oid btopfamily;
Oid btopcintype;
int16 btstrategy;
info->reverse_sort[i] = (opt & INDOPTION_DESC) != 0;
info->nulls_first[i] = (opt & INDOPTION_NULLS_FIRST) != 0;
ltopr = get_opfamily_member(
info->opfamily[i], info->opcintype[i], info->opcintype[i], BTLessStrategyNumber);
if (OidIsValid(ltopr) &&
get_ordering_op_properties(ltopr, &btopfamily, &btopcintype, &btstrategy) &&
btopcintype == info->opcintype[i] && btstrategy == BTLessStrategyNumber) {
/* Successful mapping */
info->sortopfamily[i] = btopfamily;
} else {
/* Fail ... quietly treat index as unordered */
info->sortopfamily = NULL;
info->reverse_sort = NULL;
info->nulls_first = NULL;
break;
}
}
} else {
info->sortopfamily = NULL;
info->reverse_sort = NULL;
info->nulls_first = NULL;
}
/*
* Fetch the index expressions and predicate, if any. We must
* modify the copies we obtain from the relcache to have the
* correct varno for the parent relation, so that they match up
* correctly against qual clauses.
*/
info->indexprs = RelationGetIndexExpressions(indexRelation);
info->indpred = RelationGetIndexPredicate(indexRelation);
if (info->indexprs && varno != 1)
ChangeVarNodes((Node*)info->indexprs, 1, varno, 0);
if (info->indpred && varno != 1)
ChangeVarNodes((Node*)info->indpred, 1, varno, 0);
/* Build targetlist using the completed indexprs data */
info->indextlist = build_index_tlist(root, info, relation);
info->predOK = false; /* set later in indxpath.c */
info->unique = index->indisunique;
info->immediate = index->indimmediate;
info->hypothetical = false;
/*
* Estimate the index size. If it's not a partial index, we lock
* the number-of-tuples estimate to equal the parent table; if it
* is partial then we have to use the same methods as we would for
* a table, except we can be sure that the index is not larger
* than the table.
*/
if (info->indpred == NIL) {
#ifdef PGXC
/*
* If parent relation is distributed the local storage manager
* does not have actual information about index size.
* We have to get relation statistics instead.
*/
if (IS_PGXC_COORDINATOR && relation->rd_locator_info != NULL) {
info->pages = indexRelation->rd_rel->relpages;
} else {
#endif
// non-partitioned index or global partition index
if (!RelationIsPartitioned(indexRelation) || RelationIsGlobalIndex(indexRelation)) {
info->pages = RelationGetNumberOfBlocks(indexRelation);
} else { // partitioned index
ListCell* cell = NULL;
BlockNumber partIndexPages = 0;
int partitionNum = getNumberOfPartitions(relation);
foreach (cell, sampledPartitionIds) {
Oid partOid = lfirst_oid(cell);
Oid partIndexOid = getPartitionIndexOid(indexRelation->rd_id, partOid);
Partition partIndex = partitionOpen(indexRelation, partIndexOid, AccessShareLock);
partIndexPages += PartitionGetNumberOfBlocks(indexRelation, partIndex);
partitionClose(indexRelation, partIndex, AccessShareLock);
}
// if sampled ESTIMATE_PARTITION_NUMBER, infer the pages of index,
// else partIndexPages is the actrual pages of index.
if (sampledPartitionIds != NIL) {
if (sampledPartitionIds->length == ESTIMATE_PARTITION_NUMBER)
partIndexPages *= partitionNum / ESTIMATE_PARTITION_NUMBER;
}
info->pages = partIndexPages;
}
#ifdef PGXC
}
#endif
info->tuples = rel->tuples;
} else {
double allvisfrac; /* dummy */
estimate_rel_size(indexRelation, NULL, &info->pages, &info->tuples, &allvisfrac, NULL);
if (info->tuples > rel->tuples)
info->tuples = rel->tuples;
}
info->ispartitionedindex = RelationIsPartitioned(relation);
info->partitionindex = InvalidOid;
index_close(indexRelation, NoLock);
indexinfos = lcons(info, indexinfos);
}
list_free_ext(indexoidlist);
}
rel->indexlist = indexinfos;
setRelStoreInfo(rel, relation);
/* Grab the fdwroutine info using the relcache, while we have it */
if (relation->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
rel->fdwroutine = GetFdwRoutineForRelation(relation, true);
else
rel->fdwroutine = NULL;
heap_close(relation, NoLock);
/*
* Allow a plugin to editorialize on the info we obtained from the
* catalogs. Actions might include altering the assumed relation size,
* removing an index, or adding a hypothetical index to the indexlist.
*/
if (get_relation_info_hook)
(*get_relation_info_hook) (root, relationObjectId, inhparent, rel);
}
/*
* estimate_rel_size - estimate # pages and # tuples in a table or index
*
* We also estimate the fraction of the pages that are marked all-visible in
* the visibility map, for use in estimation of index-only scans.
*
* If attr_widths isn't NULL, it points to the zero-index entry of the
* relation's attr_widths[] cache; we fill this in if we have need to compute
* the attribute widths for estimation purposes.
*/
void estimate_rel_size(Relation rel, int32* attr_widths, RelPageType* pages, double* tuples, double* allvisfrac,
List** sampledPartitionIds)
{
RelPageType curpages = 0;
RelPageType relpages;
double reltuples;
BlockNumber relallvisible;
double density;
switch (rel->rd_rel->relkind) {
case RELKIND_RELATION:
#ifdef PGXC
/*
* This is a remote table... we have no idea how many pages/rows
* we may get from a scan of this table. However, we should set the
* costs in such a manner that cheapest paths should pick up the
* ones involving these remote rels
*
* These allow for maximum query shipping to the remote
* side later during the planning phase
*
* This has to be set on a remote Coordinator only
* as it hugely penalizes performance on backend Nodes.
*
* Override the estimates only for remote tables (currently
* identified by non-NULL rd_locator_info)
*/
if (IS_PGXC_COORDINATOR && !IsConnFromCoord() && rel->rd_locator_info) {
*pages = (RelPageType)DEFAULT_PAGES_NUM;
*tuples = (double)DEFAULT_TUPLES_NUM;
*tuples = clamp_row_est(*tuples);
#ifdef STREAMPLAN
if (IS_PGXC_COORDINATOR) {
if ((rel->rd_rel->relpages > 0) || (rel->rd_rel->reltuples > 0)) {
*pages = rel->rd_rel->relpages;
*tuples = (double)rel->rd_rel->reltuples;
*tuples = clamp_row_est(*tuples);
} else if (rel->rd_id >= FirstNormalObjectId && !check_relation_analyzed(rel->rd_id)) {
if (u_sess->analyze_cxt.need_autoanalyze) {
bool is_analyzed = AutoAnaProcess::runAutoAnalyze(rel);
if (is_analyzed) {
/* refresh the statistic info */
AcceptInvalidationMessages();
/* keep the default value for empty table */
if ((rel->rd_rel->relpages > 0) || (rel->rd_rel->reltuples > 0)) {
*pages = rel->rd_rel->relpages;
*tuples = (double)rel->rd_rel->reltuples;
*tuples = clamp_row_est(*tuples);
}
} else {
elog(LOG,
"[AUTO-ANALYZE] fail to do autoanalyze on table \"%s.%s\"",
get_namespace_name(RelationGetNamespace(rel), true),
RelationGetRelationName(rel));
set_noanalyze_rellist(rel->rd_id, 0);
}
} else
set_noanalyze_rellist(rel->rd_id, 0);
}
}
#endif
break;
}
#endif
/* fall through */
case RELKIND_INDEX:
case RELKIND_GLOBAL_INDEX:
case RELKIND_MATVIEW:
/* fall through */
case RELKIND_TOASTVALUE:
/*
* 1. if it has storage, ok to call the smgr;(for non-partitioned relation)
* 2. if it doesnot have storage(for partitioned relation),
* get the curpages of average first
* ESTIMATE_PARTITION_NUMBER non-zero-pages partitions
* multiply total number of partitions
*/
if (RelationIsPartitioned(rel) && !RelationIsColStore(rel) && !RelationIsGlobalIndex(rel)) {
acquireSamplesForPartitionedRelation(rel, AccessShareLock, &curpages, sampledPartitionIds);
} else if (RelationIsValuePartitioned(rel)) {
/*
* For value partitioned rels, the effort of getting curpages might be
* very consuming(high uniqueness of partKey), as we have to recursively
* get HDFS file sizes under each partition directories, so for this
* case we'd prefer to look at relpages in pg_class to get a rough estimation
* as workaround.
*
* Please note, as comparing to none-partitioned HDFS table, partitioned
* HDFS's execution plan is more heavily depending on STATS so we are
* strongly recommanding to run ANALYZE command before submitting a
* query with partitioned HDFS tables.
*/
curpages = rel->rd_rel->relpages;
} else {
curpages = RelationGetNumberOfBlocks(rel);
}
/*
* HACK: if the relation has never yet been vacuumed, use a
* minimum size estimate of 10 pages. The idea here is to avoid
* assuming a newly-created table is really small, even if it
* currently is, because that may not be true once some data gets
* loaded into it. Once a vacuum or analyze cycle has been done
* on it, it's more reasonable to believe the size is somewhat
* stable.
*
* (Note that this is only an issue if the plan gets cached and
* used again after the table has been filled. What we're trying
* to avoid is using a nestloop-type plan on a table that has
* grown substantially since the plan was made. Normally,
* autovacuum/autoanalyze will occur once enough inserts have
* happened and cause cached-plan invalidation; but that doesn't
* happen instantaneously, and it won't happen at all for cases
* such as temporary tables.)
*
* We approximate "never vacuumed" by "has relpages = 0", which
* means this will also fire on genuinely empty relations. Not
* great, but fortunately that's a seldom-seen case in the real
* world, and it shouldn't degrade the quality of the plan too
* much anyway to err in this direction.
*
* There are two exceptions wherein we don't apply this heuristic.
* One is if the table has inheritance children. Totally empty
* parent tables are quite common, so we should be willing to
* believe that they are empty. Also, we don't apply the 10-page
* minimum to indexes.
*/
if (curpages < 10 && rel->rd_rel->relpages == 0 && !rel->rd_rel->relhassubclass &&
rel->rd_rel->relkind != RELKIND_INDEX)
curpages = 10;
/* report estimated # pages */
*pages = curpages;
/* quick exit if rel is clearly empty */
if (curpages == 0) {
*tuples = 0;
*allvisfrac = 0;
break;
}
/* coerce values in pg_class to more desirable types */
relpages = rel->rd_rel->relpages;
reltuples = (double)rel->rd_rel->reltuples;
relallvisible = (BlockNumber)rel->rd_rel->relallvisible;
/*
* If it's an index, discount the metapage while estimating the
* number of tuples. This is a kluge because it assumes more than
* it ought to about index structure. Currently it's OK for
* btree, hash, and GIN indexes but suspect for GiST indexes.
*/
if (rel->rd_rel->relkind == RELKIND_INDEX && relpages > 0) {
curpages--;
relpages--;
}
/* estimate number of tuples from previous tuple density */
if (relpages > 0)
density = reltuples / (double)relpages;
else {
/*
* When we have no data because the relation was truncated,
* estimate tuple width from attribute datatypes. We assume
* here that the pages are completely full, which is OK for
* tables (since they've presumably not been VACUUMed yet) but
* is probably an overestimate for indexes. Fortunately
* get_relation_info() can clamp the overestimate to the
* parent table's size.
*
* Note: this code intentionally disregards alignment
* considerations, because (a) that would be gilding the lily
* considering how crude the estimate is, and (b) it creates
* platform dependencies in the default plans which are kind
* of a headache for regression testing.
*/
int32 tuple_width;
tuple_width = get_rel_data_width(rel, attr_widths);
tuple_width += sizeof(HeapTupleHeaderData);
tuple_width += sizeof(ItemPointerData);
/* note: integer division is intentional here */
density = (BLCKSZ - SizeOfPageHeaderData) / (double)tuple_width;
}
*tuples = rint(density * curpages);
*tuples = clamp_row_est(*tuples);
/*
* We use relallvisible as-is, rather than scaling it up like we
* do for the pages and tuples counts, on the theory that any
* pages added since the last VACUUM are most likely not marked
* all-visible. But costsize.c wants it converted to a fraction.
*/
if (relallvisible == 0 || curpages <= 0)
*allvisfrac = 0;
else if ((double)relallvisible >= curpages)
*allvisfrac = 1;
else
*allvisfrac = (double)relallvisible / curpages;
break;
case RELKIND_SEQUENCE:
/* Sequences always have a known size */
*pages = 1;
*tuples = 1;
*allvisfrac = 0;
break;
case RELKIND_FOREIGN_TABLE:
/* Just use whatever's in pg_class */
*pages = rel->rd_rel->relpages;
*tuples = rel->rd_rel->reltuples;
*tuples = clamp_row_est(*tuples);
*allvisfrac = 0;
/*
* Append no analyze relation to g_NoAnalyzeRelNameList
* in order to print warning and output to log for hdfs foreign table.
*/
if ((*pages == 0) && (*tuples == 0) && (rel->rd_id >= FirstNormalObjectId))
set_noanalyze_rellist(rel->rd_id, 0);
break;
default:
/* else it has no disk storage; probably shouldn't get here? */
*pages = 0;
*tuples = 0;
*allvisfrac = 0;
break;
}
}
/*
* get_rel_data_width
*
* Estimate the average width of (the data part of) the relation's tuples.
*
* If attr_widths isn't NULL, it points to the zero-index entry of the
* relation's attr_widths[] cache; use and update that cache as appropriate.
*
* Currently we ignore dropped columns. Ideally those should be included
* in the result, but we haven't got any way to get info about them; and
* since they might be mostly NULLs, treating them as zero-width is not
* necessarily the wrong thing anyway.
*/
static int32 get_rel_data_width(Relation rel, int32* attr_widths, bool vectorized)
{
int32 tuple_width = 0;
int i;
bool isPartition = RelationIsPartition(rel);
bool hasencoded = false;
for (i = 1; i <= RelationGetNumberOfAttributes(rel); i++) {
Form_pg_attribute att = rel->rd_att->attrs[i - 1];
int32 item_width;
int4 att_typmod = att->atttypmod;
Oid att_typid = att->atttypid;
if (att->attisdropped)
continue;
/* use previously cached data, if any */
if (attr_widths != NULL && attr_widths[i] > 0) {
tuple_width += attr_widths[i];
continue;
}
/* This should match set_rel_width() in costsize.c */
item_width = get_attavgwidth(RelationGetRelid(rel), i, isPartition);
if (item_width <= 0) {
item_width = get_typavgwidth(att_typid, att_typmod);
if (unlikely(item_width <= 0)) {
ereport(ERROR, (errmodule(MOD_OPT), errcode(ERRCODE_DATA_EXCEPTION),
errmsg("Expected positive width estimation.")));
}
}
if (attr_widths != NULL)
attr_widths[i] = item_width;
if (vectorized) {
if (COL_IS_ENCODE((int)att_typid)) {
hasencoded = true;
tuple_width += alloc_trunk_size(item_width);
}
} else
tuple_width += item_width;
}
if (vectorized)
tuple_width += sizeof(Datum) * (i + (hasencoded ? 1 : 0));
return tuple_width;
}
/*
* get_relation_data_width
*
* External API for get_rel_data_width: same behavior except we have to
* open the relcache entry.
*/
int32 get_relation_data_width(Oid relid, Oid partid, int32* attr_widths, bool vectorized)
{
int32 result;
Relation relation;
Partition partition = NULL;
Relation targetrel = NULL;
/* As above, assume relation is already locked */
relation = heap_open(relid, NoLock);
if (OidIsValid(partid)) {
partition = partitionOpen(relation, partid, NoLock);
targetrel = partitionGetRelation(relation, partition);
} else {
targetrel = relation;
}
result = get_rel_data_width(targetrel, attr_widths, vectorized);
if (OidIsValid(partid)) {
releaseDummyRelation(&targetrel);
partitionClose(relation, partition, NoLock);
}
heap_close(relation, NoLock);
return result;
}
int32 getPartitionDataWidth(Relation partRel, int32* attr_widths)
{
return get_rel_data_width(partRel, attr_widths);
}
int32 getIdxDataWidth(Relation rel, IndexInfo* info, bool vectorized)
{
int32 width = 0;
bool isPartition = RelationIsPartition(rel);
bool hasencoded = false;
int i = 0;
int expr_i = 0;
for (i = 0; i < info->ii_NumIndexAttrs; i++) {
AttrNumber attnum = info->ii_KeyAttrNumbers[i];
Oid typid = InvalidOid;
int32 item_width = 0;
if (attnum > 0) {
Form_pg_attribute att = rel->rd_att->attrs[attnum - 1];
/* This should match set_rel_width() in costsize.c */
item_width = get_attavgwidth(RelationGetRelid(rel), attnum, isPartition);
if (item_width <= 0) {
item_width = get_typavgwidth(att->atttypid, att->atttypmod);
AssertEreport(item_width > 0, MOD_OPT, "");
}
typid = att->atttypid;
} else if (expr_i < list_length(info->ii_Expressions)) {
Node* expr = (Node*)list_nth(info->ii_Expressions, expr_i);
typid = exprType(expr);
item_width = get_typavgwidth(typid, exprTypmod(expr));
expr_i++;
}
if (vectorized) {
if (COL_IS_ENCODE(typid)) {
hasencoded = true;
width += alloc_trunk_size(item_width);
}
} else
width += item_width;
}
if (vectorized) {
/* we include index columns, index ctid, and sort column for encoded column */
int numCol = info->ii_NumIndexAttrs + 1 + (hasencoded ? 1 : 0);
width += alloc_trunk_size(sizeof(Datum) * numCol) + alloc_trunk_size(sizeof(uint8) * numCol);
}
return width;
}
/*
* get_relation_constraints
*
* Retrieve the validated CHECK constraint expressions of the given relation.
*
* Returns a List (possibly empty) of constraint expressions. Each one
* has been canonicalized, and its Vars are changed to have the varno
* indicated by rel->relid. This allows the expressions to be easily
* compared to expressions taken from WHERE.
*
* If include_notnull is true, "col IS NOT NULL" expressions are generated
* and added to the result for each column that's marked attnotnull.
*
* Note: at present this is invoked at most once per relation per planner
* run, and in many cases it won't be invoked at all, so there seems no
* point in caching the data in RelOptInfo.
*/
static List* get_relation_constraints(PlannerInfo* root, Oid relationObjectId, RelOptInfo* rel, bool include_notnull)
{
List* result = NIL;
Index varno = rel->relid;
Relation relation;
TupleConstr* constr = NULL;
/*
* We assume the relation has already been safely locked.
*/
relation = heap_open(relationObjectId, NoLock);
constr = relation->rd_att->constr;
if (constr != NULL) {
int num_check = constr->num_check;
int i;
for (i = 0; i < num_check; i++) {
Node* cexpr = NULL;
/*
* If this constraint hasn't been fully validated yet, we must
* ignore it here.
*/
if (!constr->check[i].ccvalid)
continue;
cexpr = (Node*)stringToNode(constr->check[i].ccbin);
/*
* Run each expression through const-simplification and
* canonicalization. This is not just an optimization, but is
* necessary, because we will be comparing it to
* similarly-processed qual clauses, and may fail to detect valid
* matches without this. This must match the processing done to
* qual clauses in preprocess_expression()! (We can skip the
* stuff involving subqueries, however, since we don't allow any
* in check constraints.)
*/
cexpr = eval_const_expressions(root, cexpr);
cexpr = (Node*)canonicalize_qual((Expr*)cexpr);
/* Fix Vars to have the desired varno */
if (varno != 1)
ChangeVarNodes(cexpr, 1, varno, 0);
/*
* Finally, convert to implicit-AND format (that is, a List) and
* append the resulting item(s) to our output list.
*/
result = list_concat(result, make_ands_implicit((Expr*)cexpr));
}
/* Add NOT NULL constraints in expression form, if requested */
if (include_notnull && constr->has_not_null) {
int natts = relation->rd_att->natts;
for (i = 1; i <= natts; i++) {
Form_pg_attribute att = relation->rd_att->attrs[i - 1];
if (att->attnotnull && !att->attisdropped) {
NullTest* ntest = makeNode(NullTest);
ntest->arg = (Expr*)makeVar(varno, i, att->atttypid, att->atttypmod, att->attcollation, 0);
ntest->nulltesttype = IS_NOT_NULL;
ntest->argisrow = type_is_rowtype(att->atttypid);
result = lappend(result, ntest);
}
}
}
}
heap_close(relation, NoLock);
return result;
}
/*
* relation_excluded_by_constraints
*
* Detect whether the relation need not be scanned because it has either
* self-inconsistent restrictions, or restrictions inconsistent with the
* relation's validated CHECK constraints.
*
* Note: this examines only rel->relid, rel->reloptkind, and
* rel->baserestrictinfo; therefore it can be called before filling in
* other fields of the RelOptInfo.
*/
bool relation_excluded_by_constraints(PlannerInfo* root, RelOptInfo* rel, RangeTblEntry* rte)
{
List* safe_restrictions = NIL;
List* constraint_pred = NIL;
List* safe_constraints = NIL;
ListCell* lc = NULL;
/* Skip the test if constraint exclusion is disabled for the rel */
if (u_sess->attr.attr_sql.constraint_exclusion == CONSTRAINT_EXCLUSION_OFF ||
(u_sess->attr.attr_sql.constraint_exclusion == CONSTRAINT_EXCLUSION_PARTITION &&
!(rel->reloptkind == RELOPT_OTHER_MEMBER_REL ||
(root->hasInheritedTarget && rel->reloptkind == RELOPT_BASEREL &&
rel->relid == (unsigned int)root->parse->resultRelation))))
return false;
/*
* Check for self-contradictory restriction clauses. We dare not make
* deductions with non-immutable functions, but any immutable clauses that
* are self-contradictory allow us to conclude the scan is unnecessary.
*
* Note: strip off RestrictInfo because predicate_refuted_by() isn't
* expecting to see any in its predicate argument.
*/
foreach (lc, rel->baserestrictinfo) {
RestrictInfo* rinfo = (RestrictInfo*)lfirst(lc);
if (!contain_mutable_functions((Node*)rinfo->clause))
safe_restrictions = lappend(safe_restrictions, rinfo->clause);
}
if (predicate_refuted_by(safe_restrictions, safe_restrictions, false))
return true;
/* Only plain relations have constraints */
if (rte->rtekind != RTE_RELATION || rte->inh)
return false;
/*
* OK to fetch the constraint expressions. Include "col IS NOT NULL"
* expressions for attnotnull columns, in case we can refute those.
*/
constraint_pred = get_relation_constraints(root, rte->relid, rel, true);
/*
* We do not currently enforce that CHECK constraints contain only
* immutable functions, so it's necessary to check here. We daren't draw
* conclusions from plan-time evaluation of non-immutable functions. Since
* they're ANDed, we can just ignore any mutable constraints in the list,
* and reason about the rest.
*/
safe_constraints = NIL;
foreach (lc, constraint_pred) {
Node* pred = (Node*)lfirst(lc);
if (!contain_mutable_functions(pred))
safe_constraints = lappend(safe_constraints, pred);
}
/*
* The constraints are effectively ANDed together, so we can just try to
* refute the entire collection at once. This may allow us to make proofs
* that would fail if we took them individually.
*
* Note: we use rel->baserestrictinfo, not safe_restrictions as might seem
* an obvious optimization. Some of the clauses might be OR clauses that
* have volatile and nonvolatile subclauses, and it's OK to make
* deductions with the nonvolatile parts.
*/
if (predicate_refuted_by(safe_constraints, rel->baserestrictinfo, false))
return true;
return false;
}
/*
* build_physical_tlist
*
* Build a targetlist consisting of exactly the relation's user attributes,
* in order. The executor can special-case such tlists to avoid a projection
* step at runtime, so we use such tlists preferentially for scan nodes.
*
* Exception: if there are any dropped columns, we punt and return NIL.
* Ideally we would like to handle the dropped-column case too. However this
* creates problems for ExecTypeFromTL, which may be asked to build a tupdesc
* for a tlist that includes vars of no-longer-existent types. In theory we
* could dig out the required info from the pg_attribute entries of the
* relation, but that data is not readily available to ExecTypeFromTL.
* For now, we don't apply the physical-tlist optimization when there are
* dropped cols.
*
* We also support building a "physical" tlist for subqueries, functions,
* values lists, and CTEs, since the same optimization can occur in
* SubqueryScan, FunctionScan, ValuesScan, CteScan, and WorkTableScan nodes.
*/
List* build_physical_tlist(PlannerInfo* root, RelOptInfo* rel)
{
List* tlist = NIL;
Index varno = rel->relid;
RangeTblEntry* rte = planner_rt_fetch(varno, root);
Relation relation;
Query* subquery = NULL;
Var* var = NULL;
ListCell* l = NULL;
int attrno, numattrs;
List* colvars = NIL;
switch (rte->rtekind) {
case RTE_RELATION:
/* Assume we already have adequate lock */
relation = heap_open(rte->relid, NoLock);
numattrs = RelationGetNumberOfAttributes(relation);
for (attrno = 1; attrno <= numattrs; attrno++) {
Form_pg_attribute att_tup = relation->rd_att->attrs[attrno - 1];
if (att_tup->attisdropped) {
/* found a dropped col, so punt */
tlist = NIL;
break;
}
var = makeVar(varno, attrno, att_tup->atttypid, att_tup->atttypmod, att_tup->attcollation, 0);
tlist = lappend(tlist, makeTargetEntry((Expr*)var, attrno, NULL, false));
}
heap_close(relation, NoLock);
break;
case RTE_SUBQUERY:
subquery = rte->subquery;
foreach (l, subquery->targetList) {
TargetEntry* tle = (TargetEntry*)lfirst(l);
/*
* A resjunk column of the subquery can be reflected as
* resjunk in the physical tlist; we need not punt.
*/
var = makeVarFromTargetEntry(varno, tle);
tlist = lappend(tlist, makeTargetEntry((Expr*)var, tle->resno, NULL, tle->resjunk));
}
break;
case RTE_FUNCTION:
case RTE_VALUES:
case RTE_CTE:
/* Not all of these can have dropped cols, but share code anyway */
expandRTE(rte, varno, 0, -1, true /* include dropped */, NULL, &colvars);
foreach (l, colvars) {
var = (Var*)lfirst(l);
/*
* A non-Var in expandRTE's output means a dropped column;
* must punt.
*/
if (!IsA(var, Var)) {
tlist = NIL;
break;
}
tlist = lappend(tlist, makeTargetEntry((Expr*)var, var->varattno, NULL, false));
}
break;
default:
/* caller error */
ereport(ERROR,
(errmodule(MOD_OPT),
errcode(ERRCODE_UNRECOGNIZED_NODE_TYPE),
errmsg("unsupported RTE kind %d in build_physical_tlist", (int)rte->rtekind)));
break;
}
return tlist;
}
/*
* mark_index_col
* mark real index column in g_index_vars.
* vars with type conversion in g_index_vars are from parser.
*/
static void mark_index_col(Oid relid, AttrNumber attno, Oid indexoid)
{
if (g_index_vars == NULL || attno < 0)
return;
ListCell* lc = NULL;
foreach (lc, g_index_vars) {
IndexVar* var = (IndexVar*)lfirst(lc);
if (var->relid == relid && var->attno == attno) {
var->indexcol = true;
var->indexoids = lappend_oid(var->indexoids, indexoid);
break;
}
}
}
/*
* build_index_tlist
*
* Build a targetlist representing the columns of the specified index.
* Each column is represented by a Var for the corresponding base-relation
* column, or an expression in base-relation Vars, as appropriate.
*
* There are never any dropped columns in indexes, so unlike
* build_physical_tlist, we need no failure case.
*/
List* build_index_tlist(PlannerInfo* root, IndexOptInfo* index, Relation heapRelation)
{
List* tlist = NIL;
Index varno = index->rel->relid;
ListCell* indexpr_item = NULL;
int i;
indexpr_item = list_head(index->indexprs);
for (i = 0; i < index->ncolumns; i++) {
int indexkey = index->indexkeys[i];
Expr* indexvar = NULL;
if (indexkey != 0) {
/* simple column */
Form_pg_attribute att_tup;
if (indexkey < 0) {
att_tup = SystemAttributeDefinition(indexkey, heapRelation->rd_rel->relhasoids,
RELATION_HAS_BUCKET(heapRelation));
} else {
att_tup = heapRelation->rd_att->attrs[indexkey - 1];
}
indexvar = (Expr*)makeVar(varno, indexkey, att_tup->atttypid, att_tup->atttypmod, att_tup->attcollation, 0);
if (enable_check_implicit_cast()) {
mark_index_col(heapRelation->rd_id, (AttrNumber)indexkey, index->indexoid);
}
} else {
/* expression column */
if (indexpr_item == NULL) {
ereport(ERROR,
(errmodule(MOD_OPT),
errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
errmsg("wrong number of index expressions")));
}
indexvar = (Expr*)lfirst(indexpr_item);
indexpr_item = lnext(indexpr_item);
}
tlist = lappend(tlist, makeTargetEntry(indexvar, i + 1, NULL, false));
}
if (indexpr_item != NULL) {
ereport(ERROR,
(errmodule(MOD_OPT),
errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
errmsg("wrong number of index expressions")));
}
return tlist;
}
/*
* restriction_selectivity
*
* Returns the selectivity of a specified restriction operator clause.
* This code executes registered procedures stored in the
* operator relation, by calling the function manager.
*
* See clause_selectivity() for the meaning of the additional parameters.
*/
Selectivity restriction_selectivity(PlannerInfo* root, Oid operatorid, List* args, Oid inputcollid, int varRelid)
{
RegProcedure oprrest = get_oprrest(operatorid);
float8 result;
/*
* if the oprrest procedure is missing for whatever reason, use a
* selectivity of 0.5
*/
if (!oprrest)
return (Selectivity)0.5;
result = DatumGetFloat8(OidFunctionCall4Coll(oprrest,
inputcollid,
PointerGetDatum(root),
ObjectIdGetDatum(operatorid),
PointerGetDatum(args),
Int32GetDatum(varRelid)));
if (result < 0.0 || result > 1.0)
ereport(ERROR,
(errmodule(MOD_OPT),
errcode(ERRCODE_OPTIMIZER_INCONSISTENT_STATE),
errmsg("invalid restriction selectivity: %f", result)));
return (Selectivity)result;
}
/*
* join_selectivity
*
* Returns the selectivity of a specified join operator clause.
* This code executes registered procedures stored in the
* operator relation, by calling the function manager.
*/
Selectivity join_selectivity(
PlannerInfo* root, Oid operatorid, List* args, Oid inputcollid, JoinType jointype, SpecialJoinInfo* sjinfo)
{
RegProcedure oprjoin = get_oprjoin(operatorid);
float8 result;
/*
* if the oprjoin procedure is missing for whatever reason, use a
* selectivity of 0.5
*/
if (!oprjoin)
return (Selectivity)0.5;
result = DatumGetFloat8(OidFunctionCall5Coll(oprjoin,
inputcollid,
PointerGetDatum(root),
ObjectIdGetDatum(operatorid),
PointerGetDatum(args),
Int16GetDatum(jointype),
PointerGetDatum(sjinfo)));
if (result < 0.0 || result > 1.0)
ereport(ERROR,
(errmodule(MOD_OPT),
errcode(ERRCODE_OPTIMIZER_INCONSISTENT_STATE),
errmsg("invalid join selectivity: %f", result)));
return (Selectivity)result;
}
/*
* has_unique_index
*
* Detect whether there is a unique index on the specified attribute
* of the specified relation, thus allowing us to conclude that all
* the (non-null) values of the attribute are distinct.
*
* This function does not check the index's indimmediate property, which
* means that uniqueness may transiently fail to hold intra-transaction.
* That's appropriate when we are making statistical estimates, but beware
* of using this for any correctness proofs.
*/
bool has_unique_index(RelOptInfo* rel, AttrNumber attno)
{
ListCell* ilist = NULL;
foreach (ilist, rel->indexlist) {
IndexOptInfo* index = (IndexOptInfo*)lfirst(ilist);
/*
* Note: ignore partial indexes, since they don't allow us to conclude
* that all attr values are distinct, *unless* they are marked predOK
* which means we know the index's predicate is satisfied by the
* query. We don't take any interest in expressional indexes either.
* Also, a multicolumn unique index doesn't allow us to conclude that
* just the specified attr is unique.
*/
if (index->unique && index->nkeycolumns == 1 && index->indexkeys[0] == attno &&
(index->indpred == NIL || index->predOK))
return true;
}
return false;
}
/*
* @@GaussDB@@
* Target : data partition
* Brief : estimate # pages and # tuples in a partition, and get the
* : tablespace that the partition locates
* Description :
* Notes : simulate the caculation of the ordinary table
*/
void estimatePartitionSize(
Relation relation, Oid partitionid, int32* attr_widths, RelPageType* pages, double* tuples, double* allvisfrac)
{
BlockNumber curpages = 0;
BlockNumber partitionpages = 0;
double partitiontuples = 0;
BlockNumber partitionallvisible = 0;
double density = 0.0;
// partition already locked by caller
Partition partition = partitionOpen(relation, partitionid, NoLock);
/* calculate the number of blocks in the partition */
curpages = PartitionGetNumberOfBlocks(relation, partition);
/*
* use a minimum size estimate of 10 pages. The idea here is to avoid assuming
* a newly-created table is really small, even if it currently is, because that may
* not be true once some data gets loaded into it. Once a vacuum or analyze
* cycle has been done on it, it's more reasonable to believe the size is somewhat
* stable.
*/
if (curpages < 10 && !partition->pd_part->relpages && PartitionIsTablePartition(partition)) {
curpages = 10;
}
/* report estimated # pages */
*pages = curpages;
/* quick exit if partition is clearly empty */
if (curpages == 0) {
*tuples = 0;
*allvisfrac = 0;
partitionClose(relation, partition, NoLock);
return;
}
/* coerce values in pg_partition to more desirable types */
partitionpages = (BlockNumber)partition->pd_part->relpages;
partitiontuples = (double)partition->pd_part->reltuples;
partitionallvisible = (BlockNumber)partition->pd_part->relallvisible;
/*
* If it's an index, discount the metapage while estimating the number of tuples.
* This is a kluge because it assumes more than it ought to about index structure.
* Currently it's OK for btree, hash, and GIN indexes but suspect for GiST indexes.
*/
if (partitionpages > 0 && PartitionIsIndexPartition(partition)) {
curpages--;
partitionpages--;
}
/* estimate number of tuples from previous tuple density */
if (partitionpages > 0) {
density = partitiontuples / (double)partitionpages;
} else {
/*
* When we have no data because the partition was truncated,
* estimate tuple width from attribute datatypes. We assume
* here that the pages are completely full, which is OK for
* tables (since they've presumably not been VACUUMed yet) but
* is probably an overestimate for indexes. Fortunately
* get_relation_info() can clamp the overestimate to the
* parent table's size.
*
* Note: this code intentionally disregards alignment
* considerations, because (a) that would be gilding the lily
* considering how crude the estimate is, and (b) it creates
* platform dependencies in the default plans which are kind
* of a headache for regression testing.
*/
int32 tuple_width;
Relation fakerel = partitionGetRelation(relation, partition);
tuple_width = get_rel_data_width(fakerel, attr_widths);
tuple_width += sizeof(HeapTupleHeaderData);
tuple_width += sizeof(ItemPointerData);
density = (BLCKSZ - SizeOfPageHeaderData) / (double)tuple_width;
releaseDummyRelation(&fakerel);
}
*tuples = rint(density * (double)curpages);
/*
* We use relallvisible as-is, rather than scaling it up like we
* do for the pages and tuples counts, on the theory that any
* pages added since the last VACUUM are most likely not marked
* all-visible. But costsize.c wants it converted to a fraction.
*/
if (partitionallvisible == 0 || curpages <= 0) {
*allvisfrac = 0;
} else if ((double)partitionallvisible >= curpages) {
*allvisfrac = 1;
} else {
*allvisfrac = (double)partitionallvisible / curpages;
}
partitionClose(relation, partition, NoLock);
}
/*
* Brief : Set the relation store information.
* Input : relOptInfo, the RelOptInfo stuct.
* relaiton, the relation to be seted.
* Output : None.
* Return Value : None.
* Notes : None.
*/
static void setRelStoreInfo(RelOptInfo* relOptInfo, Relation relation)
{
AssertEreport(relation != NULL, MOD_OPT, "Relation is null.");
if (RelationIsColStore(relation)) {
/*
* This is a column store table.
*/
/*
* Set store location type.
*/
if (RelationIsDfsStore(relation)) {
relOptInfo->relStoreLocation = HDFS_STORE;
} else {
relOptInfo->relStoreLocation = LOCAL_STORE;
}
/*
* Set store format type.
*/
if (RelationIsPAXFormat(relation)) {
relOptInfo->orientation = REL_PAX_ORIENTED;
} else {
AssertEreport(RelationIsCUFormat(relation), MOD_OPT, "Unexpected relation store format.");
relOptInfo->orientation = REL_COL_ORIENTED;
}
} else if(RelationIsTsStore(relation)) {
relOptInfo->orientation = REL_TIMESERIES_ORIENTED;
relOptInfo->relStoreLocation = LOCAL_STORE;
} else {
/*
* This is a row store table.
*/
relOptInfo->orientation = REL_ROW_ORIENTED;
relOptInfo->relStoreLocation = LOCAL_STORE;
}
}