aboutsummaryrefslogtreecommitdiff
path: root/src/backend
diff options
context:
space:
mode:
Diffstat (limited to 'src/backend')
-rw-r--r--src/backend/access/gist/gist.c2
-rw-r--r--src/backend/access/heap/heapam_handler.c2
-rw-r--r--src/backend/jit/llvm/llvmjit_inline.cpp2
-rw-r--r--src/backend/rewrite/rewriteManip.c2
-rw-r--r--src/backend/storage/buffer/buf_init.c2
-rw-r--r--src/backend/storage/ipc/standby.c2
-rw-r--r--src/backend/utils/adt/geo_ops.c2
-rw-r--r--src/backend/utils/adt/inet_net_pton.c4
-rw-r--r--src/backend/utils/adt/jsonpath.c2
9 files changed, 10 insertions, 10 deletions
diff --git a/src/backend/access/gist/gist.c b/src/backend/access/gist/gist.c
index 45c00aaa87a..470b121e7da 100644
--- a/src/backend/access/gist/gist.c
+++ b/src/backend/access/gist/gist.c
@@ -1533,7 +1533,7 @@ initGISTstate(Relation index)
* The truncated tupdesc for non-leaf index tuples, which doesn't contain
* the INCLUDE attributes.
*
- * It is used to form tuples during tuple adjustement and page split.
+ * It is used to form tuples during tuple adjustment and page split.
* B-tree creates shortened tuple descriptor for every truncated tuple,
* because it is doing this less often: it does not have to form truncated
* tuples during page split. Also, B-tree is not adjusting tuples on
diff --git a/src/backend/access/heap/heapam_handler.c b/src/backend/access/heap/heapam_handler.c
index a4a28e88ec9..92ea1d163e6 100644
--- a/src/backend/access/heap/heapam_handler.c
+++ b/src/backend/access/heap/heapam_handler.c
@@ -2362,7 +2362,7 @@ heapam_scan_sample_next_block(TableScanDesc scan, SampleScanState *scanstate)
if (blockno >= hscan->rs_nblocks)
{
- /* wrap to begining of rel, might not have started at 0 */
+ /* wrap to beginning of rel, might not have started at 0 */
blockno = 0;
}
diff --git a/src/backend/jit/llvm/llvmjit_inline.cpp b/src/backend/jit/llvm/llvmjit_inline.cpp
index 07b5fc7b38e..8005d43a849 100644
--- a/src/backend/jit/llvm/llvmjit_inline.cpp
+++ b/src/backend/jit/llvm/llvmjit_inline.cpp
@@ -308,7 +308,7 @@ llvm_build_inline_plan(llvm::Module *mod)
* Check whether function and all its dependencies are too
* big. Dependencies already counted for other functions that
* will get inlined are not counted again. While this make
- * things somewhat order dependant, I can't quite see a point
+ * things somewhat order dependent, I can't quite see a point
* in a different behaviour.
*/
if (running_instcount > inlineState.costLimit)
diff --git a/src/backend/rewrite/rewriteManip.c b/src/backend/rewrite/rewriteManip.c
index e20158b9626..6189a068535 100644
--- a/src/backend/rewrite/rewriteManip.c
+++ b/src/backend/rewrite/rewriteManip.c
@@ -1015,7 +1015,7 @@ AddQual(Query *parsetree, Node *qual)
errmsg("conditional UNION/INTERSECT/EXCEPT statements are not implemented")));
}
- /* INTERSECT want's the original, but we need to copy - Jan */
+ /* INTERSECT wants the original, but we need to copy - Jan */
copy = copyObject(qual);
parsetree->jointree->quals = make_and_qual(parsetree->jointree->quals,
diff --git a/src/backend/storage/buffer/buf_init.c b/src/backend/storage/buffer/buf_init.c
index b0ee3a26d6e..ccd2c31c0b3 100644
--- a/src/backend/storage/buffer/buf_init.c
+++ b/src/backend/storage/buffer/buf_init.c
@@ -179,7 +179,7 @@ BufferShmemSize(void)
* and benchmarking has shown that keeping every BufferDesc aligned on a
* cache line boundary is important for performance. So, instead, the
* array of I/O locks is allocated in a separate tranche. Because those
- * locks are not highly contentended, we lay out the array with minimal
+ * locks are not highly contended, we lay out the array with minimal
* padding.
*/
size = add_size(size, mul_size(NBuffers, sizeof(LWLockMinimallyPadded)));
diff --git a/src/backend/storage/ipc/standby.c b/src/backend/storage/ipc/standby.c
index 842fcabd979..25b7e314afc 100644
--- a/src/backend/storage/ipc/standby.c
+++ b/src/backend/storage/ipc/standby.c
@@ -202,7 +202,7 @@ WaitExceedsMaxStandbyDelay(void)
/*
* Progressively increase the sleep times, but not to more than 1s, since
- * pg_usleep isn't interruptable on some platforms.
+ * pg_usleep isn't interruptible on some platforms.
*/
standbyWait_us *= 2;
if (standbyWait_us > 1000000)
diff --git a/src/backend/utils/adt/geo_ops.c b/src/backend/utils/adt/geo_ops.c
index 2ecc73b7724..373784fcc1e 100644
--- a/src/backend/utils/adt/geo_ops.c
+++ b/src/backend/utils/adt/geo_ops.c
@@ -5315,7 +5315,7 @@ lseg_crossing(float8 x, float8 y, float8 prev_x, float8 prev_y)
/* both non-positive so do not cross positive X-axis */
return 0;
- /* x and y cross axises, see URL above point_inside() */
+ /* x and y cross axes, see URL above point_inside() */
z = float8_mi(float8_mul(float8_mi(x, prev_x), y),
float8_mul(float8_mi(y, prev_y), x));
if (FPzero(z))
diff --git a/src/backend/utils/adt/inet_net_pton.c b/src/backend/utils/adt/inet_net_pton.c
index 6f3ece1209c..c28809fad68 100644
--- a/src/backend/utils/adt/inet_net_pton.c
+++ b/src/backend/utils/adt/inet_net_pton.c
@@ -189,7 +189,7 @@ inet_cidr_pton_ipv4(const char *src, u_char *dst, size_t size)
goto emsgsize;
}
- /* Firey death and destruction unless we prefetched EOS. */
+ /* Fiery death and destruction unless we prefetched EOS. */
if (ch != '\0')
goto enoent;
@@ -309,7 +309,7 @@ inet_net_pton_ipv4(const char *src, u_char *dst)
goto emsgsize;
}
- /* Firey death and destruction unless we prefetched EOS. */
+ /* Fiery death and destruction unless we prefetched EOS. */
if (ch != '\0')
goto enoent;
diff --git a/src/backend/utils/adt/jsonpath.c b/src/backend/utils/adt/jsonpath.c
index 10cd38a5315..d5da1558670 100644
--- a/src/backend/utils/adt/jsonpath.c
+++ b/src/backend/utils/adt/jsonpath.c
@@ -229,7 +229,7 @@ static int
flattenJsonPathParseItem(StringInfo buf, JsonPathParseItem *item,
int nestingLevel, bool insideArraySubscript)
{
- /* position from begining of jsonpath data */
+ /* position from beginning of jsonpath data */
int32 pos = buf->len - JSONPATH_HDRSZ;
int32 chld;
int32 next;