diff options
Diffstat (limited to 'src/backend')
36 files changed, 43 insertions, 50 deletions
diff --git a/src/backend/access/gin/README b/src/backend/access/gin/README index 838fdc0d630..76c12ae2f6b 100644 --- a/src/backend/access/gin/README +++ b/src/backend/access/gin/README @@ -270,7 +270,7 @@ is stored in the higher bits. That requires 43 bits in total, which conveniently fits in at most 6 bytes. A compressed posting list is passed around and stored on disk in a -PackedPostingList struct. The first item in the list is stored uncompressed +GinPostingList struct. The first item in the list is stored uncompressed as a regular ItemPointerData, followed by the length of the list in bytes, followed by the packed items. diff --git a/src/backend/access/gin/ginbtree.c b/src/backend/access/gin/ginbtree.c index 11a8ed7bbc2..4c29261256a 100644 --- a/src/backend/access/gin/ginbtree.c +++ b/src/backend/access/gin/ginbtree.c @@ -650,7 +650,7 @@ ginPlaceToPage(GinBtree btree, GinBtreeStack *stack, } else { - elog(ERROR, "invalid return code from GIN placeToPage method: %d", rc); + elog(ERROR, "invalid return code from GIN beginPlaceToPage method: %d", rc); result = false; /* keep compiler quiet */ } diff --git a/src/backend/access/gist/gistget.c b/src/backend/access/gist/gistget.c index 4e0c500a22e..95ac35e3bb3 100644 --- a/src/backend/access/gist/gistget.c +++ b/src/backend/access/gist/gistget.c @@ -74,7 +74,7 @@ gistkillitems(IndexScanDesc scan) /* * Mark all killedItems as dead. We need no additional recheck, because, - * if page was modified, pageLSN must have changed. + * if page was modified, curPageLSN must have changed. */ for (i = 0; i < so->numKilled; i++) { @@ -379,11 +379,11 @@ gistScanPage(IndexScanDesc scan, GISTSearchItem *pageItem, double *myDistances, /* * Check if the page was deleted after we saw the downlink. There's - * nothing of interest on a deleted page. Note that we must do this - * after checking the NSN for concurrent splits! It's possible that - * the page originally contained some tuples that are visible to us, - * but was split so that all the visible tuples were moved to another - * page, and then this page was deleted. + * nothing of interest on a deleted page. Note that we must do this after + * checking the NSN for concurrent splits! It's possible that the page + * originally contained some tuples that are visible to us, but was split + * so that all the visible tuples were moved to another page, and then + * this page was deleted. */ if (GistPageIsDeleted(page)) { diff --git a/src/backend/access/hash/hash_xlog.c b/src/backend/access/hash/hash_xlog.c index d7b70981101..86e7210acb8 100644 --- a/src/backend/access/hash/hash_xlog.c +++ b/src/backend/access/hash/hash_xlog.c @@ -706,7 +706,7 @@ hash_xlog_squeeze_page(XLogReaderState *record) /* * if the page on which are adding tuples is a page previous to freed - * overflow page, then update its nextblno. + * overflow page, then update its nextblkno. */ if (xldata->is_prev_bucket_same_wrt) { diff --git a/src/backend/access/hash/hashinsert.c b/src/backend/access/hash/hashinsert.c index 5321762d5ea..89876d2ccd0 100644 --- a/src/backend/access/hash/hashinsert.c +++ b/src/backend/access/hash/hashinsert.c @@ -257,8 +257,8 @@ restart_insert: * _hash_pgaddtup() -- add a tuple to a particular page in the index. * * This routine adds the tuple to the page as requested; it does not write out - * the page. It is an error to call pgaddtup() without pin and write lock on - * the target buffer. + * the page. It is an error to call this function without pin and write lock + * on the target buffer. * * Returns the offset number at which the tuple was inserted. This function * is responsible for preserving the condition that tuples in a hash index diff --git a/src/backend/access/heap/heapam.c b/src/backend/access/heap/heapam.c index 94309949fac..e33f019939d 100644 --- a/src/backend/access/heap/heapam.c +++ b/src/backend/access/heap/heapam.c @@ -74,7 +74,7 @@ static HeapTuple heap_prepare_insert(Relation relation, HeapTuple tup, TransactionId xid, CommandId cid, int options); static XLogRecPtr log_heap_update(Relation reln, Buffer oldbuf, Buffer newbuf, HeapTuple oldtup, - HeapTuple newtup, HeapTuple old_key_tup, + HeapTuple newtup, HeapTuple old_key_tuple, bool all_visible_cleared, bool new_all_visible_cleared); static Bitmapset *HeapDetermineModifiedColumns(Relation relation, Bitmapset *interesting_cols, diff --git a/src/backend/access/heap/visibilitymap.c b/src/backend/access/heap/visibilitymap.c index 64dfe06b261..a08922b0798 100644 --- a/src/backend/access/heap/visibilitymap.c +++ b/src/backend/access/heap/visibilitymap.c @@ -124,7 +124,7 @@ /* prototypes for internal routines */ static Buffer vm_readbuf(Relation rel, BlockNumber blkno, bool extend); -static void vm_extend(Relation rel, BlockNumber nvmblocks); +static void vm_extend(Relation rel, BlockNumber vm_nblocks); /* diff --git a/src/backend/access/nbtree/nbtsort.c b/src/backend/access/nbtree/nbtsort.c index d0b9013caf4..b30cf9e9898 100644 --- a/src/backend/access/nbtree/nbtsort.c +++ b/src/backend/access/nbtree/nbtsort.c @@ -657,7 +657,7 @@ _bt_blwritepage(BTWriteState *wstate, Page page, BlockNumber blkno) /* XLOG stuff */ if (wstate->btws_use_wal) { - /* We use the heap NEWPAGE record type for this */ + /* We use the XLOG_FPI record type for this */ log_newpage(&wstate->index->rd_node, MAIN_FORKNUM, blkno, page, true); } diff --git a/src/backend/access/transam/README b/src/backend/access/transam/README index ad4083eb6b5..b5a2cb2de83 100644 --- a/src/backend/access/transam/README +++ b/src/backend/access/transam/README @@ -530,7 +530,7 @@ void XLogResetInsertion(void) construction workspace. This is only needed if you have already called XLogBeginInsert(), but decide to not insert the record after all. -void XLogEnsureRecordSpace(int max_block_id, int nrdatas) +void XLogEnsureRecordSpace(int max_block_id, int ndatas) Normally, the WAL record construction buffers have the following limits: diff --git a/src/backend/access/transam/clog.c b/src/backend/access/transam/clog.c index 34c74d96f8e..595c860aaaf 100644 --- a/src/backend/access/transam/clog.c +++ b/src/backend/access/transam/clog.c @@ -92,7 +92,7 @@ static int ZeroCLOGPage(int pageno, bool writeXlog); static bool CLOGPagePrecedes(int page1, int page2); static void WriteZeroPageXlogRec(int pageno); static void WriteTruncateXlogRec(int pageno, TransactionId oldestXact, - Oid oldestXidDb); + Oid oldestXactDb); static void TransactionIdSetPageStatus(TransactionId xid, int nsubxids, TransactionId *subxids, XidStatus status, XLogRecPtr lsn, int pageno, diff --git a/src/backend/access/transam/multixact.c b/src/backend/access/transam/multixact.c index 2e169dd3f92..7b2448e05ba 100644 --- a/src/backend/access/transam/multixact.c +++ b/src/backend/access/transam/multixact.c @@ -282,7 +282,7 @@ typedef struct MultiXactStateData } MultiXactStateData; /* - * Last element of OldestMemberMXactID and OldestVisibleMXactId arrays. + * Last element of OldestMemberMXactId and OldestVisibleMXactId arrays. * Valid elements are (1..MaxOldestSlot); element 0 is never used. */ #define MaxOldestSlot (MaxBackends + max_prepared_xacts) diff --git a/src/backend/catalog/aclchk.c b/src/backend/catalog/aclchk.c index 291412e305b..88ce37c08de 100644 --- a/src/backend/catalog/aclchk.c +++ b/src/backend/catalog/aclchk.c @@ -4114,7 +4114,7 @@ pg_largeobject_aclmask_snapshot(Oid lobj_oid, Oid roleid, return mask; /* - * Get the largeobject's ACL from pg_language_metadata + * Get the largeobject's ACL from pg_largeobject_metadata */ pg_lo_meta = table_open(LargeObjectMetadataRelationId, AccessShareLock); diff --git a/src/backend/catalog/information_schema.sql b/src/backend/catalog/information_schema.sql index 9c21ac7c629..c42e76ea894 100644 --- a/src/backend/catalog/information_schema.sql +++ b/src/backend/catalog/information_schema.sql @@ -48,10 +48,6 @@ CREATE FUNCTION _pg_expandarray(IN anyarray, OUT x anyelement, OUT n int) pg_catalog.array_upper($1,1), 1) as g(s)'; -CREATE FUNCTION _pg_keysequal(smallint[], smallint[]) RETURNS boolean - LANGUAGE sql IMMUTABLE PARALLEL SAFE -- intentionally not STRICT, to allow inlining - AS 'select $1 operator(pg_catalog.<@) $2 and $2 operator(pg_catalog.<@) $1'; - /* Given an index's OID and an underlying-table column number, return the * column's position in the index (NULL if not there) */ CREATE FUNCTION _pg_index_position(oid, smallint) RETURNS int diff --git a/src/backend/catalog/objectaddress.c b/src/backend/catalog/objectaddress.c index caf48cefa98..ce8a4e927da 100644 --- a/src/backend/catalog/objectaddress.c +++ b/src/backend/catalog/objectaddress.c @@ -530,7 +530,7 @@ static const ObjectPropertyType ObjectProperty[] = /* * This struct maps the string object types as returned by - * getObjectTypeDescription into ObjType enum values. Note that some enum + * getObjectTypeDescription into ObjectType enum values. Note that some enum * values can be obtained by different names, and that some string object types * do not have corresponding values in the output enum. The user of this map * must be careful to test for invalid values being returned. diff --git a/src/backend/commands/copy.c b/src/backend/commands/copy.c index 4f04d122c30..3aeef30b281 100644 --- a/src/backend/commands/copy.c +++ b/src/backend/commands/copy.c @@ -176,7 +176,6 @@ typedef struct CopyStateData * Working state for COPY FROM */ AttrNumber num_defaults; - FmgrInfo oid_in_function; FmgrInfo *in_functions; /* array of input functions for each attrs */ Oid *typioparams; /* array of element types for in_functions */ int *defmap; /* array of default att numbers */ diff --git a/src/backend/commands/define.c b/src/backend/commands/define.c index b6c618aa00f..9c93e415f14 100644 --- a/src/backend/commands/define.c +++ b/src/backend/commands/define.c @@ -139,7 +139,7 @@ defGetBoolean(DefElem *def) /* * The set of strings accepted here should match up with the - * grammar's opt_boolean production. + * grammar's opt_boolean_or_string production. */ if (pg_strcasecmp(sval, "true") == 0) return true; diff --git a/src/backend/commands/prepare.c b/src/backend/commands/prepare.c index c12b6137633..7e0a041fabf 100644 --- a/src/backend/commands/prepare.c +++ b/src/backend/commands/prepare.c @@ -143,7 +143,7 @@ PrepareQuery(PrepareStmt *stmt, const char *queryString, } /* - * grammar only allows OptimizableStmt, so this check should be redundant + * grammar only allows PreparableStmt, so this check should be redundant */ switch (query->commandType) { diff --git a/src/backend/commands/vacuum.c b/src/backend/commands/vacuum.c index e7b379dfda1..046bb047321 100644 --- a/src/backend/commands/vacuum.c +++ b/src/backend/commands/vacuum.c @@ -847,7 +847,7 @@ get_all_vacuum_rels(int options) } /* - * vacuum_set_xid_limits() -- compute oldest-Xmin and freeze cutoff points + * vacuum_set_xid_limits() -- compute oldestXmin and freeze cutoff points * * The output parameters are: * - oldestXmin is the cutoff value used to distinguish whether tuples are diff --git a/src/backend/executor/nodeWindowAgg.c b/src/backend/executor/nodeWindowAgg.c index def00cd7c5f..cb813573044 100644 --- a/src/backend/executor/nodeWindowAgg.c +++ b/src/backend/executor/nodeWindowAgg.c @@ -93,7 +93,7 @@ typedef struct WindowStatePerFuncData bool resulttypeByVal; bool plain_agg; /* is it just a plain aggregate function? */ - int aggno; /* if so, index of its PerAggData */ + int aggno; /* if so, index of its WindowStatePerAggData */ WindowObject winobj; /* object used in window function API */ } WindowStatePerFuncData; @@ -142,7 +142,7 @@ typedef struct WindowStatePerAggData resulttypeByVal, transtypeByVal; - int wfuncno; /* index of associated PerFuncData */ + int wfuncno; /* index of associated WindowStatePerFuncData */ /* Context holding transition value and possibly other subsidiary data */ MemoryContext aggcontext; /* may be private, or winstate->aggcontext */ diff --git a/src/backend/libpq/pqcomm.c b/src/backend/libpq/pqcomm.c index 384887e70d9..d8d863dda20 100644 --- a/src/backend/libpq/pqcomm.c +++ b/src/backend/libpq/pqcomm.c @@ -44,8 +44,8 @@ * StreamClose - Close a client/backend connection * TouchSocketFiles - Protect socket files against /tmp cleaners * pq_init - initialize libpq at backend startup - * pq_comm_reset - reset libpq during error recovery - * pq_close - shutdown libpq at backend exit + * socket_comm_reset - reset libpq during error recovery + * socket_close - shutdown libpq at backend exit * * low-level I/O: * pq_getbytes - get a known number of bytes from connection diff --git a/src/backend/nodes/params.c b/src/backend/nodes/params.c index f5d56138ee5..cf4387e40f0 100644 --- a/src/backend/nodes/params.c +++ b/src/backend/nodes/params.c @@ -144,7 +144,7 @@ EstimateParamListSpace(ParamListInfo paramLI) } /* - * Serialize a paramListInfo structure into caller-provided storage. + * Serialize a ParamListInfo structure into caller-provided storage. * * We write the number of parameters first, as a 4-byte integer, and then * write details for each parameter in turn. The details for each parameter diff --git a/src/backend/nodes/tidbitmap.c b/src/backend/nodes/tidbitmap.c index bf534599969..23905a9c352 100644 --- a/src/backend/nodes/tidbitmap.c +++ b/src/backend/nodes/tidbitmap.c @@ -1021,7 +1021,7 @@ tbm_iterate(TBMIterator *iterator) PagetableEntry *page; int ntuples; - /* In ONE_PAGE state, we don't allocate an spages[] array */ + /* In TBM_ONE_PAGE state, we don't allocate an spages[] array */ if (tbm->status == TBM_ONE_PAGE) page = &tbm->entry1; else diff --git a/src/backend/optimizer/geqo/geqo_selection.c b/src/backend/optimizer/geqo/geqo_selection.c index f8b75e60918..0a6bacc2f2a 100644 --- a/src/backend/optimizer/geqo/geqo_selection.c +++ b/src/backend/optimizer/geqo/geqo_selection.c @@ -91,7 +91,7 @@ geqo_selection(PlannerInfo *root, Chromosome *momma, Chromosome *daddy, static int linear_rand(PlannerInfo *root, int pool_size, double bias) { - double index; /* index between 0 and pop_size */ + double index; /* index between 0 and pool_size */ double max = (double) pool_size; /* diff --git a/src/backend/optimizer/plan/planner.c b/src/backend/optimizer/plan/planner.c index 8f51f59f8ac..0f918dd358d 100644 --- a/src/backend/optimizer/plan/planner.c +++ b/src/backend/optimizer/plan/planner.c @@ -4242,7 +4242,7 @@ consider_groupingsets_paths(PlannerInfo *root, * 2) If there are no empty sets and only unsortable sets, then the * rollups list will be empty (and thus l_start == NULL), and * group_pathkeys will be NIL; we must ensure that the vacuously-true - * pathkeys_contain_in test doesn't cause us to crash. + * pathkeys_contained_in test doesn't cause us to crash. */ if (l_start != NULL && pathkeys_contained_in(root->group_pathkeys, path->pathkeys)) @@ -5177,7 +5177,7 @@ make_group_input_target(PlannerInfo *root, PathTarget *final_target) * a regular aggregation node would, plus any aggregates used in HAVING; * except that the Aggref nodes should be marked as partial aggregates. * - * In addition, we'd better emit any Vars and PlaceholderVars that are + * In addition, we'd better emit any Vars and PlaceHolderVars that are * used outside of Aggrefs in the aggregation tlist and HAVING. (Presumably, * these would be Vars that are grouped by or used in grouping expressions.) * diff --git a/src/backend/parser/parse_oper.c b/src/backend/parser/parse_oper.c index 0e1015962d7..20abbb3b55a 100644 --- a/src/backend/parser/parse_oper.c +++ b/src/backend/parser/parse_oper.c @@ -134,7 +134,7 @@ LookupOperName(ParseState *pstate, List *opername, Oid oprleft, Oid oprright, /* * LookupOperWithArgs * Like LookupOperName, but the argument types are specified by - * a ObjectWithArg node. + * a ObjectWithArgs node. */ Oid LookupOperWithArgs(ObjectWithArgs *oper, bool noError) diff --git a/src/backend/postmaster/pgstat.c b/src/backend/postmaster/pgstat.c index b4f2b28b517..2bb14cdd026 100644 --- a/src/backend/postmaster/pgstat.c +++ b/src/backend/postmaster/pgstat.c @@ -6122,7 +6122,7 @@ pgstat_recv_resetcounter(PgStat_MsgResetcounter *msg, int len) } /* ---------- - * pgstat_recv_resetshared() - + * pgstat_recv_resetsharedcounter() - * * Reset some shared statistics of the cluster. * ---------- diff --git a/src/backend/regex/regcomp.c b/src/backend/regex/regcomp.c index 8765af71d4d..91078dcd806 100644 --- a/src/backend/regex/regcomp.c +++ b/src/backend/regex/regcomp.c @@ -288,7 +288,6 @@ struct vars #define NWBDRY 'W' /* non-word-boundary constraint */ #define SBEGIN 'A' /* beginning of string (even if not BOL) */ #define SEND 'Z' /* end of string (even if not EOL) */ -#define PREFER 'P' /* length preference */ /* is an arc colored, and hence on a color chain? */ #define COLORED(a) \ diff --git a/src/backend/replication/logical/reorderbuffer.c b/src/backend/replication/logical/reorderbuffer.c index e8ffa0492f1..5fa3d7323e9 100644 --- a/src/backend/replication/logical/reorderbuffer.c +++ b/src/backend/replication/logical/reorderbuffer.c @@ -1381,7 +1381,7 @@ ReorderBufferCopySnap(ReorderBuffer *rb, Snapshot orig_snap, snap->subxip[i++] = txn->xid; /* - * nsubxcnt isn't decreased when subtransactions abort, so count manually. + * subxcnt isn't decreased when subtransactions abort, so count manually. * Since it's an upper boundary it is safe to use it for the allocation * above. */ diff --git a/src/backend/storage/ipc/procarray.c b/src/backend/storage/ipc/procarray.c index fadab62950d..469aac38c40 100644 --- a/src/backend/storage/ipc/procarray.c +++ b/src/backend/storage/ipc/procarray.c @@ -265,7 +265,6 @@ CreateSharedProcArray(void) &found); } - /* Register and initialize fields of ProcLWLockTranche */ LWLockRegisterTranche(LWTRANCHE_PROC, "proc"); } diff --git a/src/backend/storage/ipc/shm_mq.c b/src/backend/storage/ipc/shm_mq.c index 4c245d1f85e..551515960d1 100644 --- a/src/backend/storage/ipc/shm_mq.c +++ b/src/backend/storage/ipc/shm_mq.c @@ -1251,7 +1251,7 @@ shm_mq_inc_bytes_written(shm_mq *mq, Size n) pg_atomic_read_u64(&mq->mq_bytes_written) + n); } -/* Shim for on_dsm_callback. */ +/* Shim for on_dsm_detach callback. */ static void shm_mq_detach_callback(dsm_segment *seg, Datum arg) { diff --git a/src/backend/storage/lmgr/lock.c b/src/backend/storage/lmgr/lock.c index 1b7053cb1cf..9089733ecc7 100644 --- a/src/backend/storage/lmgr/lock.c +++ b/src/backend/storage/lmgr/lock.c @@ -996,8 +996,8 @@ LockAcquireExtended(const LOCKTAG *locktag, /* * We can't acquire the lock immediately. If caller specified no - * blocking, remove useless table entries and return NOT_AVAIL without - * waiting. + * blocking, remove useless table entries and return + * LOCKACQUIRE_NOT_AVAIL without waiting. */ if (dontWait) { diff --git a/src/backend/storage/lmgr/predicate.c b/src/backend/storage/lmgr/predicate.c index 2d709420c3d..85a629f4fce 100644 --- a/src/backend/storage/lmgr/predicate.c +++ b/src/backend/storage/lmgr/predicate.c @@ -364,7 +364,7 @@ static SERIALIZABLEXACT *OldCommittedSxact; * These configuration variables are used to set the predicate lock table size * and to control promotion of predicate locks to coarser granularity in an * attempt to degrade performance (mostly as false positive serialization - * failure) gracefully in the face of memory pressurel + * failure) gracefully in the face of memory pressure. */ int max_predicate_locks_per_xact; /* set by guc.c */ int max_predicate_locks_per_relation; /* set by guc.c */ @@ -4809,7 +4809,7 @@ OnConflict_CheckForSerializationFailure(const SERIALIZABLEXACT *reader, } /* - * PreCommit_CheckForSerializableConflicts + * PreCommit_CheckForSerializationFailure * Check for dangerous structures in a serializable transaction * at commit. * diff --git a/src/backend/utils/adt/genfile.c b/src/backend/utils/adt/genfile.c index f5261065305..5d4f26a4901 100644 --- a/src/backend/utils/adt/genfile.c +++ b/src/backend/utils/adt/genfile.c @@ -325,7 +325,7 @@ pg_read_binary_file(PG_FUNCTION_ARGS) /* * Wrapper functions for the 1 and 3 argument variants of pg_read_file_v2() - * and pg_binary_read_file(). + * and pg_read_binary_file(). * * These are necessary to pass the sanity check in opr_sanity, which checks * that all built-in functions that share the implementing C function take diff --git a/src/backend/utils/adt/ruleutils.c b/src/backend/utils/adt/ruleutils.c index 0c58f1f1096..54816dc0287 100644 --- a/src/backend/utils/adt/ruleutils.c +++ b/src/backend/utils/adt/ruleutils.c @@ -115,7 +115,7 @@ typedef struct List *windowTList; /* targetlist for resolving WINDOW clause */ int prettyFlags; /* enabling of pretty-print functions */ int wrapColumn; /* max line length, or -1 for no limit */ - int indentLevel; /* current indent level for prettyprint */ + int indentLevel; /* current indent level for pretty-print */ bool varprefix; /* true to print prefixes on Vars */ ParseExprKind special_exprkind; /* set only for exprkinds needing special * handling */ diff --git a/src/backend/utils/mmgr/slab.c b/src/backend/utils/mmgr/slab.c index e23669fb4fb..bd8d2009829 100644 --- a/src/backend/utils/mmgr/slab.c +++ b/src/backend/utils/mmgr/slab.c @@ -40,7 +40,7 @@ * * For each block, we maintain pointer to the first free chunk - this is quite * cheap and allows us to skip all the preceding used chunks, eliminating - * a significant number of lookups in many common usage patters. In the worst + * a significant number of lookups in many common usage patterns. In the worst * case this performs as if the pointer was not maintained. * * We cache the freelist index for the blocks with the fewest free chunks diff --git a/src/backend/utils/time/snapmgr.c b/src/backend/utils/time/snapmgr.c index 40fe6ed3d30..47b0517596e 100644 --- a/src/backend/utils/time/snapmgr.c +++ b/src/backend/utils/time/snapmgr.c @@ -1171,7 +1171,7 @@ AtEOXact_Snapshot(bool isCommit, bool resetXmin) /* * During normal commit processing, we call ProcArrayEndTransaction() to - * reset the PgXact->xmin. That call happens prior to the call to + * reset the MyPgXact->xmin. That call happens prior to the call to * AtEOXact_Snapshot(), so we need not touch xmin here at all. */ if (resetXmin) |