Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
Diffstat (limited to 'src/backend')
-rw-r--r--src/backend/catalog/index.c20
-rw-r--r--src/backend/optimizer/path/costsize.c6
2 files changed, 13 insertions, 13 deletions
diff --git a/src/backend/catalog/index.c b/src/backend/catalog/index.c
index 062691c9c63..401c59f9a1a 100644
--- a/src/backend/catalog/index.c
+++ b/src/backend/catalog/index.c
@@ -2381,8 +2381,8 @@ IndexBuildHeapRangeScan(Relation heapRelation,
case HEAPTUPLE_INSERT_IN_PROGRESS:
/*
- * In "anyvisible" mode, this tuple is visible and we don't
- * need any further checks.
+ * In "anyvisible" mode, this tuple is visible and we
+ * don't need any further checks.
*/
if (anyvisible)
{
@@ -2437,8 +2437,8 @@ IndexBuildHeapRangeScan(Relation heapRelation,
/*
* As with INSERT_IN_PROGRESS case, this is unexpected
- * unless it's our own deletion or a system catalog;
- * but in anyvisible mode, this tuple is visible.
+ * unless it's our own deletion or a system catalog; but
+ * in anyvisible mode, this tuple is visible.
*/
if (anyvisible)
{
@@ -2892,9 +2892,9 @@ validate_index(Oid heapId, Oid indexId, Snapshot snapshot)
static inline int64
itemptr_encode(ItemPointer itemptr)
{
- BlockNumber block = ItemPointerGetBlockNumber(itemptr);
- OffsetNumber offset = ItemPointerGetOffsetNumber(itemptr);
- int64 encoded;
+ BlockNumber block = ItemPointerGetBlockNumber(itemptr);
+ OffsetNumber offset = ItemPointerGetOffsetNumber(itemptr);
+ int64 encoded;
/*
* Use the 16 least significant bits for the offset. 32 adjacent bits are
@@ -2913,8 +2913,8 @@ itemptr_encode(ItemPointer itemptr)
static inline void
itemptr_decode(ItemPointer itemptr, int64 encoded)
{
- BlockNumber block = (BlockNumber) (encoded >> 16);
- OffsetNumber offset = (OffsetNumber) (encoded & 0xFFFF);
+ BlockNumber block = (BlockNumber) (encoded >> 16);
+ OffsetNumber offset = (OffsetNumber) (encoded & 0xFFFF);
ItemPointerSet(itemptr, block, offset);
}
@@ -2960,7 +2960,7 @@ validate_index_heapscan(Relation heapRelation,
/* state variables for the merge */
ItemPointer indexcursor = NULL;
- ItemPointerData decoded;
+ ItemPointerData decoded;
bool tuplesort_empty = false;
/*
diff --git a/src/backend/optimizer/path/costsize.c b/src/backend/optimizer/path/costsize.c
index 459368e0fd2..4e17fac6c13 100644
--- a/src/backend/optimizer/path/costsize.c
+++ b/src/backend/optimizer/path/costsize.c
@@ -228,9 +228,9 @@ cost_seqscan(Path *path, PlannerInfo *root,
/*
* Primitive parallel cost model. Assume the leader will do half as much
* work as a regular worker, because it will also need to read the tuples
- * returned by the workers when they percolate up to the gather ndoe.
- * This is almost certainly not exactly the right way to model this, so
- * this will probably need to be changed at some point...
+ * returned by the workers when they percolate up to the gather node. This
+ * is almost certainly not exactly the right way to model this, so this
+ * will probably need to be changed at some point...
*/
if (nworkers > 0)
run_cost = run_cost / (nworkers + 0.5);