diff options
author | Michael Paquier <michael@paquier.xyz> | 2023-05-02 12:23:08 +0900 |
---|---|---|
committer | Michael Paquier <michael@paquier.xyz> | 2023-05-02 12:23:08 +0900 |
commit | 8961cb9a0315fa23458587b3df547ca8d8e00f85 (patch) | |
tree | f9ea45d43d53950d56062dc9eb0cebe547503718 /src/backend | |
parent | 6fd8ae6888847da842ee4810fcd83b013cbed27f (diff) | |
download | postgresql-8961cb9a0315fa23458587b3df547ca8d8e00f85.tar.gz |
Fix typos in comments
The changes done in this commit impact comments with no direct
user-visible changes, with fixes for incorrect function, variable or
structure names.
Author: Alexander Lakhin
Discussion: https://postgr.es/m/e8c38840-596a-83d6-bd8d-cebc51111572@gmail.com
Diffstat (limited to 'src/backend')
29 files changed, 41 insertions, 42 deletions
diff --git a/src/backend/access/gist/gistxlog.c b/src/backend/access/gist/gistxlog.c index 9a86fb3fef..a2ddfd5e69 100644 --- a/src/backend/access/gist/gistxlog.c +++ b/src/backend/access/gist/gistxlog.c @@ -580,8 +580,8 @@ gistXLogAssignLSN(void) int dummy = 0; /* - * Records other than SWITCH_WAL must have content. We use an integer 0 to - * follow the restriction. + * Records other than XLOG_SWITCH must have content. We use an integer 0 + * to follow the restriction. */ XLogBeginInsert(); XLogSetRecordFlags(XLOG_MARK_UNIMPORTANT); diff --git a/src/backend/access/heap/heapam.c b/src/backend/access/heap/heapam.c index b300a4675e..0124f37911 100644 --- a/src/backend/access/heap/heapam.c +++ b/src/backend/access/heap/heapam.c @@ -1223,7 +1223,7 @@ heap_set_tidrange(TableScanDesc sscan, ItemPointer mintid, * Calculate the first block and the number of blocks we must scan. We * could be more aggressive here and perform some more validation to try * and further narrow the scope of blocks to scan by checking if the - * lowerItem has an offset above MaxOffsetNumber. In this case, we could + * lowestItem has an offset above MaxOffsetNumber. In this case, we could * advance startBlk by one. Likewise, if highestItem has an offset of 0 * we could scan one fewer blocks. However, such an optimization does not * seem worth troubling over, currently. diff --git a/src/backend/access/heap/heapam_handler.c b/src/backend/access/heap/heapam_handler.c index e2e35b71ea..cbb35aa73d 100644 --- a/src/backend/access/heap/heapam_handler.c +++ b/src/backend/access/heap/heapam_handler.c @@ -816,7 +816,7 @@ heapam_relation_copy_for_cluster(Relation OldHeap, Relation NewHeap, * If the last pages of the scan were empty, we would go to * the next phase while heap_blks_scanned != heap_blks_total. * Instead, to ensure that heap_blks_scanned is equivalent to - * total_heap_blks after the table scan phase, this parameter + * heap_blks_total after the table scan phase, this parameter * is manually updated to the correct value when the table * scan finishes. */ diff --git a/src/backend/access/heap/pruneheap.c b/src/backend/access/heap/pruneheap.c index 3f0342351f..20df39c149 100644 --- a/src/backend/access/heap/pruneheap.c +++ b/src/backend/access/heap/pruneheap.c @@ -198,8 +198,7 @@ heap_page_prune_opt(Relation relation, Buffer buffer) /* * Now that we have buffer lock, get accurate information about the * page's free space, and recheck the heuristic about whether to - * prune. (We needn't recheck PageIsPrunable, since no one else could - * have pruned while we hold pin.) + * prune. */ if (PageIsFull(page) || PageGetHeapFreeSpace(page) < minfree) { @@ -490,7 +489,7 @@ heap_page_prune(Relation relation, Buffer buffer, * * Due to its cost we also only want to call * TransactionIdLimitedForOldSnapshots() if necessary, i.e. we might not have - * done so in heap_hot_prune_opt() if pd_prune_xid was old enough. But we + * done so in heap_page_prune_opt() if pd_prune_xid was old enough. But we * still want to be able to remove rows that are too new to be removed * according to prstate->vistest, but that can be removed based on * old_snapshot_threshold. So we call TransactionIdLimitedForOldSnapshots() on diff --git a/src/backend/access/heap/vacuumlazy.c b/src/backend/access/heap/vacuumlazy.c index 442db3551b..cda8889f5e 100644 --- a/src/backend/access/heap/vacuumlazy.c +++ b/src/backend/access/heap/vacuumlazy.c @@ -2575,7 +2575,7 @@ lazy_vacuum_heap_page(LVRelState *vacrel, BlockNumber blkno, Buffer buffer, END_CRIT_SECTION(); /* - * Now that we have removed the LD_DEAD items from the page, once again + * Now that we have removed the LP_DEAD items from the page, once again * check if the page has become all-visible. The page is already marked * dirty, exclusively locked, and, if needed, a full page image has been * emitted. diff --git a/src/backend/access/nbtree/nbtsplitloc.c b/src/backend/access/nbtree/nbtsplitloc.c index ecb49bb471..43b67893d9 100644 --- a/src/backend/access/nbtree/nbtsplitloc.c +++ b/src/backend/access/nbtree/nbtsplitloc.c @@ -119,7 +119,7 @@ static inline IndexTuple _bt_split_firstright(FindSplitData *state, * righthand page (which is called firstrightoff), plus a boolean * indicating whether the new tuple goes on the left or right page. You * can think of the returned state as a point _between_ two adjacent data - * items (laftleft and firstright data items) on an imaginary version of + * items (lastleft and firstright data items) on an imaginary version of * origpage that already includes newitem. The bool is necessary to * disambiguate the case where firstrightoff == newitemoff (i.e. it is * sometimes needed to determine if the firstright tuple for the split is diff --git a/src/backend/access/transam/slru.c b/src/backend/access/transam/slru.c index 2a42f31ec2..71ac70fb40 100644 --- a/src/backend/access/transam/slru.c +++ b/src/backend/access/transam/slru.c @@ -809,7 +809,7 @@ SlruPhysicalWritePage(SlruCtl ctl, int pageno, int slotno, SlruWriteAll fdata) } /* - * During a WriteAll, we may already have the desired file open. + * During a SimpleLruWriteAll, we may already have the desired file open. */ if (fdata) { @@ -864,7 +864,7 @@ SlruPhysicalWritePage(SlruCtl ctl, int pageno, int slotno, SlruWriteAll fdata) else { /* - * In the unlikely event that we exceed MAX_FLUSH_BUFFERS, + * In the unlikely event that we exceed MAX_WRITEALL_BUFFERS, * fall back to treating it as a standalone write. */ fdata = NULL; @@ -1478,7 +1478,7 @@ SlruPagePrecedesTestOffset(SlruCtl ctl, int per_page, uint32 offset) * * This assumes every uint32 >= FirstNormalTransactionId is a valid key. It * assumes each value occupies a contiguous, fixed-size region of SLRU bytes. - * (MultiXactMemberCtl separates flags from XIDs. AsyncCtl has + * (MultiXactMemberCtl separates flags from XIDs. NotifyCtl has * variable-length entries, no keys, and no random access. These unit tests * do not apply to them.) */ diff --git a/src/backend/access/transam/xlogreader.c b/src/backend/access/transam/xlogreader.c index cadea21b37..631f260f79 100644 --- a/src/backend/access/transam/xlogreader.c +++ b/src/backend/access/transam/xlogreader.c @@ -1476,7 +1476,7 @@ err: } /* - * Helper function to ease writing of XLogRoutine->page_read callbacks. + * Helper function to ease writing of XLogReaderRoutine->page_read callbacks. * If this function is used, caller must supply a segment_open callback in * 'state', as that is used here. * @@ -1513,7 +1513,7 @@ WALRead(XLogReaderState *state, /* * If the data we want is not in a segment we have open, close what we * have (if anything) and open the next one, using the caller's - * provided openSegment callback. + * provided segment_open callback. */ if (state->seg.ws_file < 0 || !XLByteInSeg(recptr, state->seg.ws_segno, state->segcxt.ws_segsize) || diff --git a/src/backend/commands/opclasscmds.c b/src/backend/commands/opclasscmds.c index 864e35e45b..5f7ee23886 100644 --- a/src/backend/commands/opclasscmds.c +++ b/src/backend/commands/opclasscmds.c @@ -819,7 +819,7 @@ AlterOpFamily(AlterOpFamilyStmt *stmt) Oid amoid, /* our AM's oid */ opfamilyoid; /* oid of opfamily */ int maxOpNumber, /* amstrategies value */ - optsProcNumber, /* amopclassopts value */ + optsProcNumber, /* amoptsprocnum value */ maxProcNumber; /* amsupport value */ HeapTuple tup; Form_pg_am amform; @@ -1321,7 +1321,7 @@ assignProcTypes(OpFamilyMember *member, Oid amoid, Oid typeoid, /* * pg_amproc functions are indexed by (lefttype, righttype), but * an equalimage function can only be called at CREATE INDEX time. - * The same opclass opcintype OID is always used for leftype and + * The same opclass opcintype OID is always used for lefttype and * righttype. Providing a cross-type routine isn't sensible. * Reject cross-type ALTER OPERATOR FAMILY ... ADD FUNCTION 4 * statements here. diff --git a/src/backend/commands/subscriptioncmds.c b/src/backend/commands/subscriptioncmds.c index 56eafbff10..8b3de032ee 100644 --- a/src/backend/commands/subscriptioncmds.c +++ b/src/backend/commands/subscriptioncmds.c @@ -974,8 +974,8 @@ AlterSubscription_refresh(Subscription *sub, bool copy_data, * * Even if new worker for this particular rel is restarted it * won't be able to make any progress as we hold exclusive - * lock on subscription_rel till the transaction end. It will - * simply exit as there is no corresponding rel entry. + * lock on pg_subscription_rel till the transaction end. It + * will simply exit as there is no corresponding rel entry. * * This locking also ensures that the state of rels won't * change till we are done with this refresh operation. diff --git a/src/backend/commands/tablecmds.c b/src/backend/commands/tablecmds.c index 343fe61115..c7a8a689b7 100644 --- a/src/backend/commands/tablecmds.c +++ b/src/backend/commands/tablecmds.c @@ -2984,7 +2984,7 @@ MergeAttributes(List *schema, List *supers, char relpersistence, * generated column, we'll take its generation expression in * preference to the parent's. We must check that the child * column doesn't specify a default value or identity, which - * matches the rules for a single column in parse_util.c. + * matches the rules for a single column in parse_utilcmd.c. * * Conversely, if the parent column is not generated, the * child column can't be either. (We used to allow that, but @@ -8017,7 +8017,7 @@ ATPrepDropExpression(Relation rel, AlterTableCmd *cmd, bool recurse, bool recurs * is a bit complicated. GENERATED clauses must be attached to the column * definition and cannot be added later like DEFAULT, so if a child table * has a generation expression that the parent does not have, the child - * column will necessarily be an attlocal column. So to implement ONLY + * column will necessarily be an attislocal column. So to implement ONLY * here, we'd need extra code to update attislocal of the direct child * tables, somewhat similar to how DROP COLUMN does it, so that the * resulting state can be properly dumped and restored. @@ -10930,7 +10930,7 @@ ATExecAlterConstraint(Relation rel, AlterTableCmd *cmd, bool recurse, } /* - * ATExecConstrRecurse already invalidated relcache for the relations + * ATExecAlterConstrRecurse already invalidated relcache for the relations * having the constraint itself; here we also invalidate for relations * that have any triggers that are part of the constraint. */ diff --git a/src/backend/executor/nodeAgg.c b/src/backend/executor/nodeAgg.c index cfadef0942..a8a57b7063 100644 --- a/src/backend/executor/nodeAgg.c +++ b/src/backend/executor/nodeAgg.c @@ -3004,7 +3004,7 @@ hashagg_batch_new(LogicalTape *input_tape, int setno, } /* - * read_spilled_tuple + * hashagg_batch_read * read the next tuple from a batch's tape. Return NULL if no more. */ static MinimalTuple diff --git a/src/backend/jit/llvm/llvmjit_expr.c b/src/backend/jit/llvm/llvmjit_expr.c index a1b33d6e6c..2c3d64ea6e 100644 --- a/src/backend/jit/llvm/llvmjit_expr.c +++ b/src/backend/jit/llvm/llvmjit_expr.c @@ -2121,8 +2121,8 @@ llvm_compile_expr(ExprState *state) /* * pergroup = &aggstate->all_pergroups - * [op->d.agg_strict_trans_check.setoff] - * [op->d.agg_init_trans_check.transno]; + * [op->d.agg_trans.setoff] + * [op->d.agg_trans.transno]; */ v_allpergroupsp = l_load_struct_gep(b, v_aggstatep, diff --git a/src/backend/optimizer/plan/planagg.c b/src/backend/optimizer/plan/planagg.c index d168620665..a6090167f5 100644 --- a/src/backend/optimizer/plan/planagg.c +++ b/src/backend/optimizer/plan/planagg.c @@ -386,7 +386,7 @@ build_minmax_path(PlannerInfo *root, MinMaxAggInfo *mminfo, ntest = makeNode(NullTest); ntest->nulltesttype = IS_NOT_NULL; ntest->arg = copyObject(mminfo->target); - /* we checked it wasn't a rowtype in find_minmax_aggs_walker */ + /* we checked it wasn't a rowtype in can_minmax_aggs */ ntest->argisrow = false; ntest->location = -1; diff --git a/src/backend/parser/gram.y b/src/backend/parser/gram.y index acf6cf4866..a723d9db78 100644 --- a/src/backend/parser/gram.y +++ b/src/backend/parser/gram.y @@ -957,7 +957,7 @@ stmtmulti: stmtmulti ';' toplevel_stmt /* * toplevel_stmt includes BEGIN and END. stmt does not include them, because - * those words have different meanings in function bodys. + * those words have different meanings in function bodies. */ toplevel_stmt: stmt diff --git a/src/backend/partitioning/partbounds.c b/src/backend/partitioning/partbounds.c index cf1156b842..c685621416 100644 --- a/src/backend/partitioning/partbounds.c +++ b/src/backend/partitioning/partbounds.c @@ -2340,7 +2340,7 @@ merge_default_partitions(PartitionMap *outer_map, /* * The default partitions have to be joined with each other, so merge * them. Note that each of the default partitions isn't merged yet - * (see, process_outer_partition()/process_innerer_partition()), so + * (see, process_outer_partition()/process_inner_partition()), so * they should be merged successfully. The merged partition will act * as the default partition of the join relation. */ diff --git a/src/backend/partitioning/partdesc.c b/src/backend/partitioning/partdesc.c index de06caccd2..7a2b5e57ff 100644 --- a/src/backend/partitioning/partdesc.c +++ b/src/backend/partitioning/partdesc.c @@ -55,7 +55,7 @@ static PartitionDesc RelationBuildPartitionDesc(Relation rel, * RelationGetPartitionDesc -- get partition descriptor, if relation is partitioned * * We keep two partdescs in relcache: rd_partdesc includes all partitions - * (even those being concurrently marked detached), while rd_partdesc_nodetach + * (even those being concurrently marked detached), while rd_partdesc_nodetached * omits (some of) those. We store the pg_inherits.xmin value for the latter, * to determine whether it can be validly reused in each case, since that * depends on the active snapshot. diff --git a/src/backend/replication/logical/reorderbuffer.c b/src/backend/replication/logical/reorderbuffer.c index 9f44974473..b85b890010 100644 --- a/src/backend/replication/logical/reorderbuffer.c +++ b/src/backend/replication/logical/reorderbuffer.c @@ -1885,7 +1885,7 @@ ReorderBufferStreamCommit(ReorderBuffer *rb, ReorderBufferTXN *txn) /* * This is a PREPARED transaction, part of a two-phase commit. The * full cleanup will happen as part of the COMMIT PREPAREDs, so now - * just truncate txn by removing changes and tuple_cids. + * just truncate txn by removing changes and tuplecids. */ ReorderBufferTruncateTXN(rb, txn, true); /* Reset the CheckXidAlive */ @@ -4010,7 +4010,7 @@ ReorderBufferStreamTXN(ReorderBuffer *rb, ReorderBufferTXN *txn) * After that we need to reuse the snapshot from the previous run. * * Unlike DecodeCommit which adds xids of all the subtransactions in - * snapshot's xip array via SnapBuildCommittedTxn, we can't do that here + * snapshot's xip array via SnapBuildCommitTxn, we can't do that here * but we do add them to subxip array instead via ReorderBufferCopySnap. * This allows the catalog changes made in subtransactions decoded till * now to be visible. diff --git a/src/backend/replication/syncrep.c b/src/backend/replication/syncrep.c index 889e20b5dd..c263a59690 100644 --- a/src/backend/replication/syncrep.c +++ b/src/backend/replication/syncrep.c @@ -804,7 +804,7 @@ standby_priority_comparator(const void *a, const void *b) /* * We might have equal priority values; arbitrarily break ties by position - * in the WALSnd array. (This is utterly bogus, since that is arrival + * in the WalSnd array. (This is utterly bogus, since that is arrival * order dependent, but there are regression tests that rely on it.) */ return sa->walsnd_index - sb->walsnd_index; diff --git a/src/backend/storage/file/fd.c b/src/backend/storage/file/fd.c index 277a28fc13..4f86d44d6e 100644 --- a/src/backend/storage/file/fd.c +++ b/src/backend/storage/file/fd.c @@ -3069,7 +3069,7 @@ AtEOXact_Files(bool isCommit) /* * BeforeShmemExit_Files * - * before_shmem_access hook to clean up temp files during backend shutdown. + * before_shmem_exit hook to clean up temp files during backend shutdown. * Here, we want to clean up *all* temp files including interXact ones. */ static void diff --git a/src/backend/storage/ipc/latch.c b/src/backend/storage/ipc/latch.c index f4123e7de7..cdb95c1931 100644 --- a/src/backend/storage/ipc/latch.c +++ b/src/backend/storage/ipc/latch.c @@ -151,7 +151,7 @@ struct WaitEventSet #endif }; -/* A common WaitEventSet used to implement WatchLatch() */ +/* A common WaitEventSet used to implement WaitLatch() */ static WaitEventSet *LatchWaitSet; /* The position of the latch in LatchWaitSet. */ diff --git a/src/backend/storage/ipc/procarray.c b/src/backend/storage/ipc/procarray.c index 80153fb250..8c8d728ba8 100644 --- a/src/backend/storage/ipc/procarray.c +++ b/src/backend/storage/ipc/procarray.c @@ -1754,7 +1754,7 @@ ComputeXidHorizons(ComputeXidHorizonsResult *h) * current top-level xid any. * * Without an assigned xid we could use a horizon as aggressive as - * ReadNewTransactionid(), but we can get away with the much cheaper + * GetNewTransactionId(), but we can get away with the much cheaper * latestCompletedXid + 1: If this backend has no xid there, by * definition, can't be any newer changes in the temp table than * latestCompletedXid. diff --git a/src/backend/storage/lmgr/spin.c b/src/backend/storage/lmgr/spin.c index e406bea201..6052779d9c 100644 --- a/src/backend/storage/lmgr/spin.c +++ b/src/backend/storage/lmgr/spin.c @@ -37,7 +37,7 @@ #define NUM_EMULATION_SEMAPHORES (NUM_SPINLOCK_SEMAPHORES + NUM_ATOMICS_SEMAPHORES) #else #define NUM_EMULATION_SEMAPHORES (NUM_SPINLOCK_SEMAPHORES) -#endif /* DISABLE_ATOMICS */ +#endif /* HAVE_ATOMICS */ PGSemaphore *SpinlockSemaArray; diff --git a/src/backend/utils/activity/backend_status.c b/src/backend/utils/activity/backend_status.c index 391d5de043..a8f927f540 100644 --- a/src/backend/utils/activity/backend_status.c +++ b/src/backend/utils/activity/backend_status.c @@ -283,7 +283,7 @@ pgstat_beinit(void) * * Apart from auxiliary processes, MyBackendId, MyDatabaseId, * session userid, and application_name must be set for a - * backend (hence, this cannot be combined with pgbestat_beinit). + * backend (hence, this cannot be combined with pgstat_beinit). * Note also that we must be inside a transaction if this isn't an aux * process, as we may need to do encoding conversion on some strings. * ---------- diff --git a/src/backend/utils/activity/pgstat_wal.c b/src/backend/utils/activity/pgstat_wal.c index 94f1a3b06e..bcaed14d02 100644 --- a/src/backend/utils/activity/pgstat_wal.c +++ b/src/backend/utils/activity/pgstat_wal.c @@ -24,7 +24,7 @@ PgStat_PendingWalStats PendingWalStats = {0}; /* - * WAL usage counters saved from pgWALUsage at the previous call to + * WAL usage counters saved from pgWalUsage at the previous call to * pgstat_report_wal(). This is used to calculate how much WAL usage * happens between pgstat_report_wal() calls, by subtracting * the previous counters from the current ones. diff --git a/src/backend/utils/adt/acl.c b/src/backend/utils/adt/acl.c index 729ebd86d1..c660fd3e70 100644 --- a/src/backend/utils/adt/acl.c +++ b/src/backend/utils/adt/acl.c @@ -4717,8 +4717,8 @@ pg_has_role_id_id(PG_FUNCTION_ARGS) * Convert text string to AclMode value. * * We use USAGE to denote whether the privileges of the role are accessible - * (has_privs), MEMBER to denote is_member, and MEMBER WITH GRANT OPTION - * (or ADMIN OPTION) to denote is_admin. There is no ACL bit corresponding + * (has_privs_of_role), MEMBER to denote is_member, and MEMBER WITH GRANT + * (or ADMIN) OPTION to denote is_admin. There is no ACL bit corresponding * to MEMBER so we cheat and use ACL_CREATE for that. This convention * is shared only with pg_role_aclcheck, below. */ diff --git a/src/backend/utils/adt/pg_locale.c b/src/backend/utils/adt/pg_locale.c index 51df570ce9..f0b6567da1 100644 --- a/src/backend/utils/adt/pg_locale.c +++ b/src/backend/utils/adt/pg_locale.c @@ -1012,7 +1012,7 @@ search_locale_enum(LPWSTR pStr, DWORD dwFlags, LPARAM lparam) { /* * If the enumerated locale does not have a hyphen ("en") OR the - * lc_message input does not have an underscore ("English"), we only + * locale_name input does not have an underscore ("English"), we only * need to compare the <Language> tags. */ if (wcsrchr(pStr, '-') == NULL || wcsrchr(argv[0], '_') == NULL) diff --git a/src/backend/utils/adt/selfuncs.c b/src/backend/utils/adt/selfuncs.c index fe37e65af0..c4fcd0076e 100644 --- a/src/backend/utils/adt/selfuncs.c +++ b/src/backend/utils/adt/selfuncs.c @@ -4210,7 +4210,7 @@ estimate_multivariate_ndistinct(PlannerInfo *root, RelOptInfo *rel, * Process complex expressions, not just simple Vars. * * First, we search for an exact match of an expression. If we - * find one, we can just discard the whole GroupExprInfo, with all + * find one, we can just discard the whole GroupVarInfo, with all * the variables we extracted from it. * * Otherwise we inspect the individual vars, and try matching it diff --git a/src/backend/utils/mb/mbutils.c b/src/backend/utils/mb/mbutils.c index 033647011b..67a1ab2ab2 100644 --- a/src/backend/utils/mb/mbutils.c +++ b/src/backend/utils/mb/mbutils.c @@ -453,7 +453,7 @@ pg_do_encoding_conversion(unsigned char *src, int len, * * The output is null-terminated. * - * If destlen < srclen * MAX_CONVERSION_LENGTH + 1, the converted output + * If destlen < srclen * MAX_CONVERSION_INPUT_LENGTH + 1, the converted output * wouldn't necessarily fit in the output buffer, and the function will not * convert the whole input. * |