It's now split more or less to your suggestion:
https://github.com/jsoref/postgres/commits/spelling
diff --git a/configure b/configure
--- a/configure
+++ b/configure
@@ -7088,7 +7088,7 @@ test -z "$INSTALL_SCRIPT" && INSTALL_SCR
test -z "$INSTALL_DATA" && INSTALL_DATA='${INSTALL} -m 644'
# When Autoconf chooses install-sh as install program it tries to generate
-# a relative path to it in each makefile where it subsitutes it. This clashes
+# a relative path to it in each makefile where it substitutes it. This clashes
# with our Makefile.global concept. This workaround helps.
case $INSTALL in
*install-sh*) install_bin='';;
@@ -7232,7 +7232,7 @@ fi
$as_echo "$MKDIR_P" >&6; }
# When Autoconf chooses install-sh as mkdir -p program it tries to generate
-# a relative path to it in each makefile where it subsitutes it. This clashes
+# a relative path to it in each makefile where it substitutes it. This clashes
# with our Makefile.global concept. This workaround helps.
case $MKDIR_P in
*install-sh*) MKDIR_P='\${SHELL} \${top_srcdir}/config/install-sh -c -d';;
diff --git a/configure.in b/configure.in
--- a/configure.in
+++ b/configure.in
@@ -887,7 +887,7 @@ fi
AC_PROG_INSTALL
# When Autoconf chooses install-sh as install program it tries to generate
-# a relative path to it in each makefile where it subsitutes it. This clashes
+# a relative path to it in each makefile where it substitutes it. This clashes
# with our Makefile.global concept. This workaround helps.
case $INSTALL in
*install-sh*) install_bin='';;
@@ -900,7 +900,7 @@ AC_PROG_LN_S
AC_PROG_AWK
AC_PROG_MKDIR_P
# When Autoconf chooses install-sh as mkdir -p program it tries to generate
-# a relative path to it in each makefile where it subsitutes it. This clashes
+# a relative path to it in each makefile where it substitutes it. This clashes
# with our Makefile.global concept. This workaround helps.
case $MKDIR_P in
*install-sh*) MKDIR_P='\${SHELL} \${top_srcdir}/config/install-sh -c -d';;
diff --git a/contrib/bloom/blvacuum.c b/contrib/bloom/blvacuum.c
--- a/contrib/bloom/blvacuum.c
+++ b/contrib/bloom/blvacuum.c
@@ -51,7 +51,7 @@ blbulkdelete(IndexVacuumInfo *info, Inde
initBloomState(&state, index);
/*
- * Interate over the pages. We don't care about concurrently added
pages,
+ * Iterate over the pages. We don't care about concurrently added pages,
* they can't contain tuples to delete.
*/
npages = RelationGetNumberOfBlocks(index);
diff --git a/contrib/cube/sql/cube.sql b/contrib/cube/sql/cube.sql
--- a/contrib/cube/sql/cube.sql
+++ b/contrib/cube/sql/cube.sql
@@ -256,7 +256,7 @@ SELECT cube_dim('(0,0,0)'::cube);
SELECT cube_dim('(42,42,42),(42,42,42)'::cube);
SELECT cube_dim('(4,8,15,16,23),(4,8,15,16,23)'::cube);
--- Test of cube_ll_coord function (retrieves LL coodinate values)
+-- Test of cube_ll_coord function (retrieves LL coordinate values)
--
SELECT cube_ll_coord('(-1,1),(2,-2)'::cube, 1);
SELECT cube_ll_coord('(-1,1),(2,-2)'::cube, 2);
@@ -268,7 +268,7 @@ SELECT cube_ll_coord('(42,137)'::cube, 1
SELECT cube_ll_coord('(42,137)'::cube, 2);
SELECT cube_ll_coord('(42,137)'::cube, 3);
--- Test of cube_ur_coord function (retrieves UR coodinate values)
+-- Test of cube_ur_coord function (retrieves UR coordinate values)
--
SELECT cube_ur_coord('(-1,1),(2,-2)'::cube, 1);
SELECT cube_ur_coord('(-1,1),(2,-2)'::cube, 2);
diff --git a/contrib/earthdistance/earthdistance--1.1.sql
b/contrib/earthdistance/earthdistance--1.1.sql
--- a/contrib/earthdistance/earthdistance--1.1.sql
+++ b/contrib/earthdistance/earthdistance--1.1.sql
@@ -11,7 +11,7 @@ CREATE FUNCTION earth() RETURNS float8
LANGUAGE SQL IMMUTABLE PARALLEL SAFE
AS 'SELECT ''6378168''::float8';
--- Astromers may want to change the earth function so that distances will be
+-- Astronomers may want to change the earth function so that distances will be
-- returned in degrees. To do this comment out the above definition and
-- uncomment the one below. Note that doing this will break the regression
-- tests.
diff --git a/contrib/isn/ISSN.h b/contrib/isn/ISSN.h
--- a/contrib/isn/ISSN.h
+++ b/contrib/isn/ISSN.h
@@ -23,7 +23,7 @@
* Product 9 + 21 + 7 + 3 + 1 + 12 + 4 + 24 + 7 + 15 + 0 + 0 = 103
* 103 / 10 = 10 remainder 3
* Check digit 10 - 3 = 7
- * => 977-1144875-00-7 ?? <- suplemental number (number of the week, month,
etc.)
+ * => 977-1144875-00-7 ?? <- supplemental number (number of the week, month,
etc.)
* ^^ 00 for non-daily publications (01=Monday,
02=Tuesday, ...)
*
* The hyphenation is always in after the four digits of the ISSN code.
diff --git a/contrib/isn/isn.c b/contrib/isn/isn.c
--- a/contrib/isn/isn.c
+++ b/contrib/isn/isn.c
@@ -160,7 +160,7 @@ dehyphenate(char *bufO, char *bufI)
* into bufO using the given hyphenation range
TABLE.
* Assumes the input string to be used is of
only digits.
*
- * Returns the number of characters acctually hyphenated.
+ * Returns the number of characters actually hyphenated.
*/
static unsigned
hyphenate(char *bufO, char *bufI, const char *(*TABLE)[2], const unsigned
TABLE_index[10][2])
@@ -748,7 +748,7 @@ string2ean(const char *str, bool errorOK
}
else if (*aux2 == '!' && *(aux2 + 1) == '\0')
{
- /* the invalid check digit sufix was found, set it */
+ /* the invalid check digit suffix was found, set it */
if (!magic)
valid = false;
magic = true;
diff --git a/contrib/ltree/ltxtquery_io.c b/contrib/ltree/ltxtquery_io.c
--- a/contrib/ltree/ltxtquery_io.c
+++ b/contrib/ltree/ltxtquery_io.c
@@ -197,7 +197,7 @@ pushval_asis(QPRS_STATE *state, int type
#define STACKDEPTH 32
/*
- * make polish notaion of query
+ * make polish notation of query
*/
static int32
makepol(QPRS_STATE *state)
diff --git a/contrib/ltree/sql/ltree.sql b/contrib/ltree/sql/ltree.sql
--- a/contrib/ltree/sql/ltree.sql
+++ b/contrib/ltree/sql/ltree.sql
@@ -209,7 +209,7 @@ SELECT 'a.b.c.d.e'::ltree ? '{A.b.c.d.e,
SELECT '{a.b.c.d.e,B.df}'::ltree[] ? '{A.b.c.d.e}';
SELECT '{a.b.c.d.e,B.df}'::ltree[] ? '{A.b.c.d.e,*.df}';
---exractors
+--extractors
SELECT ('{3456,1.2.3.34}'::ltree[] ?@> '1.2.3.4') is null;
SELECT '{3456,1.2.3}'::ltree[] ?@> '1.2.3.4';
SELECT '{3456,1.2.3.4}'::ltree[] ?<@ '1.2.3';
diff --git a/contrib/pg_standby/pg_standby.c b/contrib/pg_standby/pg_standby.c
--- a/contrib/pg_standby/pg_standby.c
+++ b/contrib/pg_standby/pg_standby.c
@@ -779,7 +779,7 @@ main(int argc, char **argv)
{
/*
* Once we have restored this file successfully we can
remove some
- * prior WAL files. If this restore fails we musn't
remove any
+ * prior WAL files. If this restore fails we mustn't
remove any
* file because some of them will be requested again
immediately
* after the failed restore, or when we restart
recovery.
*/
diff --git a/contrib/pg_stat_statements/pg_stat_statements.c
b/contrib/pg_stat_statements/pg_stat_statements.c
--- a/contrib/pg_stat_statements/pg_stat_statements.c
+++ b/contrib/pg_stat_statements/pg_stat_statements.c
@@ -139,7 +139,7 @@ typedef struct Counters
{
int64 calls; /* # of times executed */
double total_time; /* total execution time, in
msec */
- double min_time; /* minimim execution time in
msec */
+ double min_time; /* minimum execution time in
msec */
double max_time; /* maximum execution time in
msec */
double mean_time; /* mean execution time in msec
*/
double sum_var_time; /* sum of variances in execution time
in msec */
diff --git a/contrib/pg_trgm/trgm_op.c b/contrib/pg_trgm/trgm_op.c
--- a/contrib/pg_trgm/trgm_op.c
+++ b/contrib/pg_trgm/trgm_op.c
@@ -413,7 +413,7 @@ comp_ptrgm(const void *v1, const void *v
* ulen1: count of unique trigrams of array "trg1".
* len2: length of array "trg2" and array "trg2indexes".
* len: length of the array "found".
- * check_only: if true then only check existaince of similar search pattern in
+ * check_only: if true then only check existence of similar search pattern in
* text.
*
* Returns word similarity.
@@ -456,7 +456,7 @@ iterate_word_similarity(int *trg2indexes
lastpos[trgindex] = i;
}
- /* Adjust lower bound if this trigram is present in required
substing */
+ /* Adjust lower bound if this trigram is present in required
substring */
if (found[trgindex])
{
int prev_lower,
@@ -547,7 +547,7 @@ iterate_word_similarity(int *trg2indexes
*
* str1: search pattern string, of length slen1 bytes.
* str2: text in which we are looking for a word, of length slen2 bytes.
- * check_only: if true then only check existaince of similar search pattern in
+ * check_only: if true then only check existence of similar search pattern in
* text.
*
* Returns word similarity.
diff --git a/contrib/pg_trgm/trgm_regexp.c b/contrib/pg_trgm/trgm_regexp.c
--- a/contrib/pg_trgm/trgm_regexp.c
+++ b/contrib/pg_trgm/trgm_regexp.c
@@ -844,7 +844,7 @@ convertPgWchar(pg_wchar c, trgm_mb_char
if (c == 0)
return false;
- /* Do the conversion, making sure the result is NUL-terminated */
+ /* Do the conversion, making sure the result is NULL-terminated */
memset(s, 0, sizeof(s));
pg_wchar2mb_with_len(&c, s, 1);
diff --git a/contrib/pgcrypto/crypt-des.c b/contrib/pgcrypto/crypt-des.c
--- a/contrib/pgcrypto/crypt-des.c
+++ b/contrib/pgcrypto/crypt-des.c
@@ -416,7 +416,7 @@ des_setkey(const char *key)
&& rawkey1 == old_rawkey1)
{
/*
- * Already setup for this key. This optimisation fails on a
zero key
+ * Already setup for this key. This optimization fails on a
zero key
* (which is weak and has bad parity anyway) in order to
simplify the
* starting conditions.
*/
diff --git a/contrib/pgcrypto/imath.h b/contrib/pgcrypto/imath.h
--- a/contrib/pgcrypto/imath.h
+++ b/contrib/pgcrypto/imath.h
@@ -163,7 +163,7 @@ mp_result mp_int_sqrt(mp_int a, mp_int c
/* Convert to an int, if representable (returns MP_RANGE if not). */
mp_result mp_int_to_int(mp_int z, int *out);
-/* Convert to nul-terminated string with the specified radix, writing at
+/* Convert to null-terminated string with the specified radix, writing at
most limit characters including the nul terminator */
mp_result mp_int_to_string(mp_int z, mp_size radix,
char *str, int limit);
diff --git a/contrib/pgcrypto/mbuf.c b/contrib/pgcrypto/mbuf.c
--- a/contrib/pgcrypto/mbuf.c
+++ b/contrib/pgcrypto/mbuf.c
@@ -311,7 +311,7 @@ pullf_read_max(PullFilter *pf, int len,
}
/*
- * caller wants exatly len bytes and dont bother with references
+ * caller wants exactly len bytes and dont bother with references
*/
int
pullf_read_fixed(PullFilter *src, int len, uint8 *dst)
diff --git a/contrib/pgcrypto/pgp-mpi-internal.c
b/contrib/pgcrypto/pgp-mpi-internal.c
--- a/contrib/pgcrypto/pgp-mpi-internal.c
+++ b/contrib/pgcrypto/pgp-mpi-internal.c
@@ -141,7 +141,7 @@ bn_to_mpi(mpz_t *bn)
}
/*
- * Decide the number of bits in the random componont k
+ * Decide the number of bits in the random component k
*
* It should be in the same range as p for signing (which
* is deprecated), but can be much smaller for encrypting.
@@ -149,8 +149,8 @@ bn_to_mpi(mpz_t *bn)
* Until I research it further, I just mimic gpg behaviour.
* It has a special mapping table, for values <= 5120,
* above that it uses 'arbitrary high number'. Following
- * algorihm hovers 10-70 bits above gpg values. And for
- * larger p, it uses gpg's algorihm.
+ * algorithm hovers 10-70 bits above gpg values. And for
+ * larger p, it uses gpg's algorithm.
*
* The point is - if k gets large, encryption will be
* really slow. It does not matter for decryption.
diff --git a/contrib/pgcrypto/pgp-mpi-openssl.c
b/contrib/pgcrypto/pgp-mpi-openssl.c
--- a/contrib/pgcrypto/pgp-mpi-openssl.c
+++ b/contrib/pgcrypto/pgp-mpi-openssl.c
@@ -74,7 +74,7 @@ bn_to_mpi(BIGNUM *bn)
}
/*
- * Decide the number of bits in the random componont k
+ * Decide the number of bits in the random component k
*
* It should be in the same range as p for signing (which
* is deprecated), but can be much smaller for encrypting.
@@ -82,8 +82,8 @@ bn_to_mpi(BIGNUM *bn)
* Until I research it further, I just mimic gpg behaviour.
* It has a special mapping table, for values <= 5120,
* above that it uses 'arbitrary high number'. Following
- * algorihm hovers 10-70 bits above gpg values. And for
- * larger p, it uses gpg's algorihm.
+ * algorithm hovers 10-70 bits above gpg values. And for
+ * larger p, it uses gpg's algorithm.
*
* The point is - if k gets large, encryption will be
* really slow. It does not matter for decryption.
diff --git a/contrib/postgres_fdw/sql/postgres_fdw.sql
b/contrib/postgres_fdw/sql/postgres_fdw.sql
--- a/contrib/postgres_fdw/sql/postgres_fdw.sql
+++ b/contrib/postgres_fdw/sql/postgres_fdw.sql
@@ -493,7 +493,7 @@ EXPLAIN (VERBOSE, COSTS OFF)
SELECT t1."C 1" FROM "S 1"."T 1" t1, LATERAL (SELECT DISTINCT t2.c1, t3.c1
FROM ft1 t2, ft2 t3 WHERE t2.c1 = t3.c1 AND t2.c2 = t1.c2) q ORDER BY t1."C 1"
OFFSET 10 LIMIT 10;
SELECT t1."C 1" FROM "S 1"."T 1" t1, LATERAL (SELECT DISTINCT t2.c1, t3.c1
FROM ft1 t2, ft2 t3 WHERE t2.c1 = t3.c1 AND t2.c2 = t1.c2) q ORDER BY t1."C 1"
OFFSET 10 LIMIT 10;
--- non-Var items in targelist of the nullable rel of a join preventing
+-- non-Var items in targetlist of the nullable rel of a join preventing
-- push-down in some cases
-- unable to push {ft1, ft2}
EXPLAIN (VERBOSE, COSTS OFF)
diff --git a/contrib/seg/seg.c b/contrib/seg/seg.c
--- a/contrib/seg/seg.c
+++ b/contrib/seg/seg.c
@@ -888,7 +888,7 @@ restore(char *result, float val, int n)
if (Abs(exp) <= 4)
{
/*
- * remove the decimal point from the mantyssa and write
the digits
+ * remove the decimal point from the mantissa and write
the digits
* to the buf array
*/
for (p = result + sign, i = 10, dp = 0; *p != 'e'; p++,
i++)
diff --git a/contrib/sepgsql/selinux.c b/contrib/sepgsql/selinux.c
--- a/contrib/sepgsql/selinux.c
+++ b/contrib/sepgsql/selinux.c
@@ -23,7 +23,7 @@
* When we ask SELinux whether the required privileges are allowed or not,
* we use security_compute_av(3). It needs us to represent object classes
* and access vectors using 'external' codes defined in the security policy.
- * It is determinded in the runtime, not build time. So, it needs an internal
+ * It is determined in the runtime, not build time. So, it needs an internal
* service to translate object class/access vectors which we want to check
* into the code which kernel want to be given.
*/
diff --git a/contrib/sepgsql/sql/label.sql b/contrib/sepgsql/sql/label.sql
--- a/contrib/sepgsql/sql/label.sql
+++ b/contrib/sepgsql/sql/label.sql
@@ -206,7 +206,7 @@ SELECT * FROM auth_tbl; -- failed
SELECT sepgsql_setcon(NULL); -- end of session
SELECT sepgsql_getcon();
--- the pooler cannot touch these tables directry
+-- the pooler cannot touch these tables directly
SELECT * FROM foo_tbl; -- failed
SELECT * FROM var_tbl; -- failed
diff --git a/contrib/spi/refint.c b/contrib/spi/refint.c
--- a/contrib/spi/refint.c
+++ b/contrib/spi/refint.c
@@ -89,7 +89,7 @@ check_primary_key(PG_FUNCTION_ARGS)
/* internal error */
elog(ERROR, "check_primary_key: cannot process DELETE events");
- /* If UPDATion the must check new Tuple, not old one */
+ /* If UPDATE, then must check new Tuple, not old one */
else
tuple = trigdata->tg_newtuple;
diff --git a/contrib/spi/timetravel.c b/contrib/spi/timetravel.c
--- a/contrib/spi/timetravel.c
+++ b/contrib/spi/timetravel.c
@@ -51,7 +51,7 @@ static EPlan *find_plan(char *ident, EPl
* and stop_date eq INFINITY [ and update_user eq current
user ]
* and all other column values as in new tuple, and insert
tuple
* with old data and stop_date eq current date
- * ELSE - skip updation of tuple.
+ * ELSE - skip UPDATE of tuple.
* 2. IF a delete affects tuple with stop_date eq INFINITY
* then insert the same tuple with stop_date eq current
date
* [ and delete_user eq current user ]
diff --git a/contrib/start-scripts/osx/PostgreSQL
b/contrib/start-scripts/osx/PostgreSQL
--- a/contrib/start-scripts/osx/PostgreSQL
+++ b/contrib/start-scripts/osx/PostgreSQL
@@ -29,7 +29,7 @@
# modified by Ray Aspeitia 12-03-2003 :
# added log rotation script to db startup
# modified StartupParameters.plist "Provides" parameter to make it easier to
-# start and stop with the SystemStarter utitlity
+# start and stop with the SystemStarter utility
# use the below command in order to correctly start/stop/restart PG with log
rotation script:
# SystemStarter [start|stop|restart] PostgreSQL
diff --git a/contrib/tsearch2/tsearch2--1.0.sql
b/contrib/tsearch2/tsearch2--1.0.sql
--- a/contrib/tsearch2/tsearch2--1.0.sql
+++ b/contrib/tsearch2/tsearch2--1.0.sql
@@ -414,7 +414,7 @@ CREATE FUNCTION stat(text,text)
LANGUAGE INTERNAL
RETURNS NULL ON NULL INPUT;
---reset - just for debuging
+--reset - just for debugging
CREATE FUNCTION reset_tsearch()
RETURNS void
as 'MODULE_PATHNAME', 'tsa_reset_tsearch'
diff --git a/contrib/xml2/xpath.c b/contrib/xml2/xpath.c
--- a/contrib/xml2/xpath.c
+++ b/contrib/xml2/xpath.c
@@ -610,7 +610,7 @@ xpath_table(PG_FUNCTION_ARGS)
/*
* At the moment we assume that the returned attributes make sense for
the
- * XPath specififed (i.e. we trust the caller). It's not fatal if they
get
+ * XPath specified (i.e. we trust the caller). It's not fatal if they
get
* it wrong - the input function for the column type will raise an error
* if the path result can't be converted into the correct binary
* representation.
diff --git a/src/Makefile.shlib b/src/Makefile.shlib
--- a/src/Makefile.shlib
+++ b/src/Makefile.shlib
@@ -377,7 +377,7 @@ DLL_DEFFILE = lib$(NAME)dll.def
$(CC) $(CFLAGS) -shared -static-libgcc -o $@ $(OBJS) $(DLL_DEFFILE)
$(LDFLAGS) $(LDFLAGS_SL) $(SHLIB_LINK) $(LIBS) -Wl,--out-implib=$(stlib)
endif
-endif # PORTNAME == cgywin
+endif # PORTNAME == cygwin
endif # PORTNAME == cygwin || PORTNAME == win32
diff --git a/src/backend/access/gist/README b/src/backend/access/gist/README
--- a/src/backend/access/gist/README
+++ b/src/backend/access/gist/README
@@ -28,7 +28,7 @@ The current implementation of GiST suppo
The support for concurrency implemented in PostgreSQL was developed based on
the paper "Access Methods for Next-Generation Database Systems" by
-Marcel Kornaker:
+Marcel Kornacker:
http://www.sai.msu.su/~megera/postgres/gist/papers/concurrency/access-methods-for-next-generation.pdf.gz
diff --git a/src/backend/access/hash/hashpage.c
b/src/backend/access/hash/hashpage.c
--- a/src/backend/access/hash/hashpage.c
+++ b/src/backend/access/hash/hashpage.c
@@ -1077,7 +1077,7 @@ static void
* already moved before the split operation was previously interrupted.
*
* The caller must hold a pin, but no lock, on the metapage and old bucket's
- * primay page buffer. The buffers are returned in the same state. (The
+ * primary page buffer. The buffers are returned in the same state. (The
* metapage is only touched if it becomes necessary to add or remove overflow
* pages.)
*/
diff --git a/src/backend/access/heap/rewriteheap.c
b/src/backend/access/heap/rewriteheap.c
--- a/src/backend/access/heap/rewriteheap.c
+++ b/src/backend/access/heap/rewriteheap.c
@@ -209,7 +209,7 @@ typedef struct RewriteMappingFile
} RewriteMappingFile;
/*
- * A single In-Memeory logical rewrite mapping, hanging of
+ * A single In-Memory logical rewrite mapping, hanging of
* RewriteMappingFile->mappings.
*/
typedef struct RewriteMappingDataEntry
diff --git a/src/backend/access/transam/commit_ts.c
b/src/backend/access/transam/commit_ts.c
--- a/src/backend/access/transam/commit_ts.c
+++ b/src/backend/access/transam/commit_ts.c
@@ -615,7 +615,7 @@ CommitTsParameterChange(bool newvalue, b
/*
* Activate this module whenever necessary.
- * This must happen during postmaster or standalong-backend
startup,
+ * This must happen during postmaster or standalone-backend
startup,
* or during WAL replay anytime the track_commit_timestamp setting
is
* changed in the master.
*
diff --git a/src/backend/access/transam/twophase.c
b/src/backend/access/transam/twophase.c
--- a/src/backend/access/transam/twophase.c
+++ b/src/backend/access/transam/twophase.c
@@ -1613,7 +1613,7 @@ CheckPointTwoPhase(XLogRecPtr redo_horiz
*
* It's also possible to move I/O out of the lock, but on every error we
* should check whether somebody committed our transaction in different
- * backend. Let's leave this optimisation for future, if somebody will
+ * backend. Let's leave this optimization for future, if somebody will
* spot that this place cause bottleneck.
*
* Note that it isn't possible for there to be a GXACT with a
diff --git a/src/backend/access/transam/xact.c
b/src/backend/access/transam/xact.c
--- a/src/backend/access/transam/xact.c
+++ b/src/backend/access/transam/xact.c
@@ -2752,7 +2752,7 @@ CommitTransactionCommand(void)
* These shouldn't happen. TBLOCK_DEFAULT means the
previous
* StartTransactionCommand didn't set the STARTED state
* appropriately, while TBLOCK_PARALLEL_INPROGRESS
should be ended
- * by EndParallelWorkerTranaction(), not this function.
+ * by EndParallelWorkerTransaction(), not this function.
*/
case TBLOCK_DEFAULT:
case TBLOCK_PARALLEL_INPROGRESS:
diff --git a/src/backend/catalog/objectaddress.c
b/src/backend/catalog/objectaddress.c
--- a/src/backend/catalog/objectaddress.c
+++ b/src/backend/catalog/objectaddress.c
@@ -770,7 +770,7 @@ static void getRelationIdentity(StringIn
*
* Note: If the object is not found, we don't give any indication of the
* reason. (It might have been a missing schema if the name was qualified, or
- * an inexistant type name in case of a cast, function or operator; etc).
+ * a nonexistent type name in case of a cast, function or operator; etc).
* Currently there is only one caller that might be interested in such info, so
* we don't spend much effort here. If more callers start to care, it might be
* better to add some support for that in this function.
diff --git a/src/backend/commands/amcmds.c b/src/backend/commands/amcmds.c
--- a/src/backend/commands/amcmds.c
+++ b/src/backend/commands/amcmds.c
@@ -34,7 +34,7 @@ static const char *get_am_type_string(ch
/*
- * CreateAcessMethod
+ * CreateAccessMethod
* Registers a new access method.
*/
ObjectAddress
diff --git a/src/backend/commands/async.c b/src/backend/commands/async.c
--- a/src/backend/commands/async.c
+++ b/src/backend/commands/async.c
@@ -313,7 +313,7 @@ typedef enum
typedef struct
{
ListenActionKind action;
- char channel[FLEXIBLE_ARRAY_MEMBER]; /* nul-terminated
string */
+ char channel[FLEXIBLE_ARRAY_MEMBER]; /* null-terminated
string */
} ListenAction;
static List *pendingActions = NIL; /* list of ListenAction */
diff --git a/src/backend/commands/copy.c b/src/backend/commands/copy.c
--- a/src/backend/commands/copy.c
+++ b/src/backend/commands/copy.c
@@ -2405,7 +2405,7 @@ CopyFrom(CopyState cstate)
* earlier scan or command. This ensures that if this subtransaction
* aborts then the frozen rows won't be visible after xact cleanup. Note
* that the stronger test of exactly which subtransaction created it is
- * crucial for correctness of this optimisation.
+ * crucial for correctness of this optimization.
*/
if (cstate->freeze)
{
@@ -2972,7 +2972,7 @@ BeginCopyFrom(ParseState *pstate,
* the special case of when the default
expression is the
* nextval() of a sequence which in this
specific case is
* known to be safe for use with the
multi-insert
- * optimisation. Hence we use this special case
function
+ * optimization. Hence we use this special case
function
* checker rather than the standard check for
* contain_volatile_functions().
*/
diff --git a/src/backend/commands/dbcommands.c
b/src/backend/commands/dbcommands.c
--- a/src/backend/commands/dbcommands.c
+++ b/src/backend/commands/dbcommands.c
@@ -685,7 +685,7 @@ createdb(ParseState *pstate, const Creat
/*
* Force synchronous commit, thus minimizing the window between
- * creation of the database files and commital of the
transaction. If
+ * creation of the database files and committal of the
transaction. If
* we crash before committing, we'll have a DB that's taking up
disk
* space but is not in pg_database, which is not good.
*/
@@ -955,7 +955,7 @@ dropdb(const char *dbname, bool missing_
/*
* Force synchronous commit, thus minimizing the window between removal
of
- * the database files and commital of the transaction. If we crash
before
+ * the database files and committal of the transaction. If we crash
before
* committing, we'll have a DB that's gone on disk but still there
* according to pg_database, which is not good.
*/
@@ -1309,7 +1309,7 @@ movedb(const char *dbname, const char *t
/*
* Force synchronous commit, thus minimizing the window between
- * copying the database files and commital of the transaction.
If we
+ * copying the database files and committal of the transaction.
If we
* crash before committing, we'll leave an orphaned set of
files on
* disk, which is not fatal but not good either.
*/
diff --git a/src/backend/commands/explain.c b/src/backend/commands/explain.c
--- a/src/backend/commands/explain.c
+++ b/src/backend/commands/explain.c
@@ -3401,7 +3401,7 @@ ExplainYAMLLineStarting(ExplainState *es
}
/*
- * YAML is a superset of JSON; unfortuantely, the YAML quoting rules are
+ * YAML is a superset of JSON; unfortunately, the YAML quoting rules are
* ridiculously complicated -- as documented in sections 5.3 and 7.3.3 of
* http://yaml.org/spec/1.2/spec.html -- so we chose to just quote everything.
* Empty strings, strings with leading or trailing whitespace, and strings
diff --git a/src/backend/commands/functioncmds.c
b/src/backend/commands/functioncmds.c
--- a/src/backend/commands/functioncmds.c
+++ b/src/backend/commands/functioncmds.c
@@ -1040,7 +1040,7 @@ CreateFunction(ParseState *pstate, Creat
}
else
{
- /* store SQL NULL instead of emtpy array */
+ /* store SQL NULL instead of empty array */
trftypes = NULL;
}
@@ -1441,7 +1441,7 @@ CreateCast(CreateCastStmt *stmt)
(errcode(ERRCODE_WRONG_OBJECT_TYPE),
errmsg("cast will be ignored because the
target data type is a domain")));
- /* Detemine the cast method */
+ /* Determine the cast method */
if (stmt->func != NULL)
castmethod = COERCION_METHOD_FUNCTION;
else if (stmt->inout)
diff --git a/src/backend/commands/indexcmds.c b/src/backend/commands/indexcmds.c
--- a/src/backend/commands/indexcmds.c
+++ b/src/backend/commands/indexcmds.c
@@ -99,7 +99,7 @@ static void RangeVarCallbackForReindexIn
* Errors arising from the attribute list still apply.
*
* Most column type changes that can skip a table rewrite do not invalidate
- * indexes. We ackowledge this when all operator classes, collations and
+ * indexes. We acknowledge this when all operator classes, collations and
* exclusion operators match. Though we could further permit intra-opfamily
* changes for btree and hash indexes, that adds subtle complexity with no
* concrete benefit for core types.
@@ -965,7 +965,7 @@ CheckMutability(Expr *expr)
* indxpath.c could do something with. However, that seems overly
* restrictive. One useful application of partial indexes is to apply
* a UNIQUE constraint across a subset of a table, and in that scenario
- * any evaluatable predicate will work. So accept any predicate here
+ * any evaluable predicate will work. So accept any predicate here
* (except ones requiring a plan), and let indxpath.c fend for itself.
*/
static void
diff --git a/src/backend/commands/publicationcmds.c
b/src/backend/commands/publicationcmds.c
--- a/src/backend/commands/publicationcmds.c
+++ b/src/backend/commands/publicationcmds.c
@@ -525,7 +525,7 @@ OpenTableList(List *tables)
myrelid = RelationGetRelid(rel);
/*
* filter out duplicates when user specifies "foo, foo"
- * Note that this algrithm is know to not be very effective
(O(N^2))
+ * Note that this algorithm is know to not be very effective
(O(N^2))
* but given that it only works on list of tables given to us
by user
* it's deemed acceptable.
*/
diff --git a/src/backend/commands/subscriptioncmds.c
b/src/backend/commands/subscriptioncmds.c
--- a/src/backend/commands/subscriptioncmds.c
+++ b/src/backend/commands/subscriptioncmds.c
@@ -474,7 +474,7 @@ DropSubscription(DropSubscriptionStmt *s
InvokeObjectDropHook(SubscriptionRelationId, subid, 0);
/*
- * Lock the subscription so noboby else can do anything with it
+ * Lock the subscription so nobody else can do anything with it
* (including the replication workers).
*/
LockSharedObject(SubscriptionRelationId, subid, 0, AccessExclusiveLock);
diff --git a/src/backend/commands/tablecmds.c b/src/backend/commands/tablecmds.c
--- a/src/backend/commands/tablecmds.c
+++ b/src/backend/commands/tablecmds.c
@@ -3167,7 +3167,7 @@ AlterTableGetLockLevel(List *cmds)
break;
/*
- * Changing foreign table options may affect
optimisation.
+ * Changing foreign table options may affect
optimization.
*/
case AT_GenericOptions:
case AT_AlterColumnGenericOptions:
@@ -6630,7 +6630,7 @@ ATAddCheckConstraint(List **wqueue, Alte
/*
* Check if ONLY was specified with ALTER TABLE. If so, allow the
- * contraint creation only if there are no children currently. Error
out
+ * constraint creation only if there are no children currently. Error
out
* otherwise.
*/
if (!recurse && children != NIL)
diff --git a/src/backend/executor/execMain.c b/src/backend/executor/execMain.c
--- a/src/backend/executor/execMain.c
+++ b/src/backend/executor/execMain.c
@@ -1261,7 +1261,7 @@ InitResultRelInfo(ResultRelInfo *resultR
resultRelInfo->ri_projectReturning = NULL;
/*
- * If partition_root has been specified, that means we are builiding the
+ * If partition_root has been specified, that means we are building the
* ResultRelationInfo for one of its leaf partitions. In that case, we
* need *not* initialize the leaf partition's constraint, but rather the
* the partition_root's (if any). We must do that explicitly like this,
diff --git a/src/backend/executor/execParallel.c
b/src/backend/executor/execParallel.c
--- a/src/backend/executor/execParallel.c
+++ b/src/backend/executor/execParallel.c
@@ -533,7 +533,7 @@ ExecParallelRetrieveInstrumentation(Plan
int plan_node_id = planstate->plan->plan_node_id;
MemoryContext oldcontext;
- /* Find the instumentation for this node. */
+ /* Find the instrumentation for this node. */
for (i = 0; i < instrumentation->num_plan_nodes; ++i)
if (instrumentation->plan_node_id[i] == plan_node_id)
break;
diff --git a/src/backend/executor/execReplication.c
b/src/backend/executor/execReplication.c
--- a/src/backend/executor/execReplication.c
+++ b/src/backend/executor/execReplication.c
@@ -391,7 +391,7 @@ ExecSimpleRelationInsert(EState *estate,
if (rel->rd_att->constr)
ExecConstraints(resultRelInfo, slot, slot, estate);
- /* Store the slot into tuple that we can insett. */
+ /* Store the slot into tuple that we can inspect. */
tuple = ExecMaterializeSlot(slot);
/* OK, store the tuple and create index entries for it */
diff --git a/src/backend/executor/nodeAgg.c b/src/backend/executor/nodeAgg.c
--- a/src/backend/executor/nodeAgg.c
+++ b/src/backend/executor/nodeAgg.c
@@ -304,7 +304,7 @@ typedef struct AggStatePerTransData
/*
* Slots for holding the evaluated input arguments. These are set up
* during ExecInitAgg() and then used for each input row requiring
- * procesessing besides what's done in AggState->evalproj.
+ * processing besides what's done in AggState->evalproj.
*/
TupleTableSlot *sortslot; /* current input tuple */
TupleTableSlot *uniqslot; /* used for multi-column DISTINCT */
diff --git a/src/backend/executor/nodeWindowAgg.c
b/src/backend/executor/nodeWindowAgg.c
--- a/src/backend/executor/nodeWindowAgg.c
+++ b/src/backend/executor/nodeWindowAgg.c
@@ -354,7 +354,7 @@ advance_windowaggregate(WindowAggState *
/*
* We must track the number of rows included in transValue, since to
- * remove the last input, advance_windowaggregate_base() musn't call the
+ * remove the last input, advance_windowaggregate_base() mustn't call
the
* inverse transition function, but simply reset transValue back to its
* initial value.
*/
diff --git a/src/backend/foreign/foreign.c b/src/backend/foreign/foreign.c
--- a/src/backend/foreign/foreign.c
+++ b/src/backend/foreign/foreign.c
@@ -724,7 +724,7 @@ GetExistingLocalJoinPath(RelOptInfo *joi
Path *path = (Path *) lfirst(lc);
JoinPath *joinpath = NULL;
- /* Skip parameterised paths. */
+ /* Skip parameterized paths. */
if (path->param_info != NULL)
continue;
diff --git a/src/backend/libpq/hba.c b/src/backend/libpq/hba.c
--- a/src/backend/libpq/hba.c
+++ b/src/backend/libpq/hba.c
@@ -109,7 +109,7 @@ static MemoryContext parsed_hba_context
*
* NOTE: the IdentLine structs can contain pre-compiled regular expressions
* that live outside the memory context. Before destroying or resetting the
- * memory context, they need to be expliticly free'd.
+ * memory context, they need to be explicitly free'd.
*/
static List *parsed_ident_lines = NIL;
static MemoryContext parsed_ident_context = NULL;
diff --git a/src/backend/optimizer/geqo/geqo_erx.c
b/src/backend/optimizer/geqo/geqo_erx.c
--- a/src/backend/optimizer/geqo/geqo_erx.c
+++ b/src/backend/optimizer/geqo/geqo_erx.c
@@ -111,7 +111,7 @@ gimme_edge_table(PlannerInfo *root, Gene
for (index1 = 0; index1 < num_gene; index1++)
{
/*
- * presume the tour is circular, i.e. 1->2, 2->3, 3->1 this
operaton
+ * presume the tour is circular, i.e. 1->2, 2->3, 3->1 this
operation
* maps n back to 1
*/
@@ -314,7 +314,7 @@ gimme_gene(PlannerInfo *root, Edge edge,
/*
* give priority to candidates with fewest remaining unused
edges;
* find out what the minimum number of unused edges is
- * (minimum_edges); if there is more than one cadidate with the
+ * (minimum_edges); if there is more than one candidate with the
* minimum number of unused edges keep count of this number
* (minimum_count);
*/
diff --git a/src/backend/optimizer/path/joinpath.c
b/src/backend/optimizer/path/joinpath.c
--- a/src/backend/optimizer/path/joinpath.c
+++ b/src/backend/optimizer/path/joinpath.c
@@ -1618,7 +1618,7 @@ select_mergejoin_clauses(PlannerInfo *ro
/*
* Insist that each side have a non-redundant eclass. This
* restriction is needed because various bits of the planner
expect
- * that each clause in a merge be associatable with some
pathkey in a
+ * that each clause in a merge be associable with some pathkey
in a
* canonical pathkey list, but redundant eclasses can't appear
in
* canonical sort orderings. (XXX it might be worth relaxing
this,
* but not enough time to address it for 8.3.)
diff --git a/src/backend/optimizer/plan/createplan.c
b/src/backend/optimizer/plan/createplan.c
--- a/src/backend/optimizer/plan/createplan.c
+++ b/src/backend/optimizer/plan/createplan.c
@@ -4467,7 +4467,7 @@ fix_indexqual_operand(Node *node, IndexO
}
}
- /* Ooops... */
+ /* Oops... */
elog(ERROR, "index key does not match expected index column");
return NULL; /* keep compiler quiet */
}
diff --git a/src/backend/optimizer/plan/planmain.c
b/src/backend/optimizer/plan/planmain.c
--- a/src/backend/optimizer/plan/planmain.c
+++ b/src/backend/optimizer/plan/planmain.c
@@ -195,7 +195,7 @@ query_planner(PlannerInfo *root, List *t
/*
* Now distribute "placeholders" to base rels as needed. This has to be
* done after join removal because removal could change whether a
- * placeholder is evaluatable at a base rel.
+ * placeholder is evaluable at a base rel.
*/
add_placeholders_to_base_rels(root);
diff --git a/src/backend/optimizer/util/joininfo.c
b/src/backend/optimizer/util/joininfo.c
--- a/src/backend/optimizer/util/joininfo.c
+++ b/src/backend/optimizer/util/joininfo.c
@@ -24,7 +24,7 @@
* Detect whether there is a joinclause that involves
* the two given relations.
*
- * Note: the joinclause does not have to be evaluatable with only these two
+ * Note: the joinclause does not have to be evaluable with only these two
* relations. This is intentional. For example consider
* SELECT * FROM a, b, c WHERE a.x = (b.y + c.z)
* If a is much larger than the other tables, it may be worthwhile to
diff --git a/src/backend/optimizer/util/restrictinfo.c
b/src/backend/optimizer/util/restrictinfo.c
--- a/src/backend/optimizer/util/restrictinfo.c
+++ b/src/backend/optimizer/util/restrictinfo.c
@@ -515,7 +515,7 @@ join_clause_is_movable_into(RestrictInfo
Relids currentrelids,
Relids
current_and_outer)
{
- /* Clause must be evaluatable given available context */
+ /* Clause must be evaluable given available context */
if (!bms_is_subset(rinfo->clause_relids, current_and_outer))
return false;
diff --git a/src/backend/parser/gram.y b/src/backend/parser/gram.y
--- a/src/backend/parser/gram.y
+++ b/src/backend/parser/gram.y
@@ -11312,7 +11312,7 @@ table_ref: relation_expr opt_alias_claus
n->lateral = true;
n->subquery = $2;
n->alias = $3;
- /* same coment as above */
+ /* same comment as above */
if ($3 == NULL)
{
if (IsA($2, SelectStmt) &&
diff --git a/src/backend/parser/parse_collate.c
b/src/backend/parser/parse_collate.c
--- a/src/backend/parser/parse_collate.c
+++ b/src/backend/parser/parse_collate.c
@@ -805,7 +805,7 @@ merge_collation_state(Oid collation,
else if (collation !=
DEFAULT_COLLATION_OID)
{
/*
- * Ooops, we have a conflict.
We cannot throw error
+ * Oops, we have a conflict.
We cannot throw error
* here, since the conflict
could be resolved by a
* later sibling CollateExpr,
or the parent might not
* care about collation anyway.
Return enough info to
@@ -824,7 +824,7 @@ merge_collation_state(Oid collation,
if (collation != context->collation)
{
/*
- * Ooops, we have a conflict of
explicit COLLATE clauses.
+ * Oops, we have a conflict of explicit
COLLATE clauses.
* Here we choose to throw error
immediately; that is what
* the SQL standard says to do, and
there's no good reason
* to be less strict.
diff --git a/src/backend/parser/parse_param.c b/src/backend/parser/parse_param.c
--- a/src/backend/parser/parse_param.c
+++ b/src/backend/parser/parse_param.c
@@ -210,7 +210,7 @@ variable_coerce_param_hook(ParseState *p
}
else
{
- /* Ooops */
+ /* Oops */
ereport(ERROR,
(errcode(ERRCODE_AMBIGUOUS_PARAMETER),
errmsg("inconsistent types deduced for
parameter $%d",
diff --git a/src/backend/parser/parse_utilcmd.c
b/src/backend/parser/parse_utilcmd.c
--- a/src/backend/parser/parse_utilcmd.c
+++ b/src/backend/parser/parse_utilcmd.c
@@ -3050,7 +3050,7 @@ transformAttachPartition(CreateStmtConte
errmsg("\"%s\" is not partitioned",
RelationGetRelationName(parentRel))));
- /* tranform the values */
+ /* transform the values */
Assert(RelationGetPartitionKey(parentRel) != NULL);
cxt->partbound = transformPartitionBound(cxt->pstate, parentRel,
cmd->bound);
diff --git a/src/backend/postmaster/bgworker.c
b/src/backend/postmaster/bgworker.c
--- a/src/backend/postmaster/bgworker.c
+++ b/src/backend/postmaster/bgworker.c
@@ -321,7 +321,7 @@ BackgroundWorkerStateChange(void)
/*
* Copy strings in a paranoid way. If shared memory is
corrupted, the
- * source data might not even be NUL-terminated.
+ * source data might not even be NULL-terminated.
*/
ascii_safe_strlcpy(rw->rw_worker.bgw_name,
slot->worker.bgw_name,
BGW_MAXLEN);
diff --git a/src/backend/postmaster/bgwriter.c
b/src/backend/postmaster/bgwriter.c
--- a/src/backend/postmaster/bgwriter.c
+++ b/src/backend/postmaster/bgwriter.c
@@ -211,7 +211,7 @@ BackgroundWriterMain(void)
/* Flush any leaked data in the top-level context */
MemoryContextResetAndDeleteChildren(bgwriter_context);
- /* re-initilialize to avoid repeated errors causing problems */
+ /* re-initialize to avoid repeated errors causing problems */
WritebackContextInit(&wb_context, &bgwriter_flush_after);
/* Now we can allow interrupts again */
diff --git a/src/backend/postmaster/postmaster.c
b/src/backend/postmaster/postmaster.c
--- a/src/backend/postmaster/postmaster.c
+++ b/src/backend/postmaster/postmaster.c
@@ -5156,7 +5156,7 @@ RandomCancelKey(int32 *cancel_key)
}
/*
- * Count up number of child processes of specified types (dead_end chidren
+ * Count up number of child processes of specified types (dead_end children
* are always excluded).
*/
static int
diff --git a/src/backend/regex/regc_pg_locale.c
b/src/backend/regex/regc_pg_locale.c
--- a/src/backend/regex/regc_pg_locale.c
+++ b/src/backend/regex/regc_pg_locale.c
@@ -224,7 +224,7 @@ static const unsigned char pg_char_prope
* pg_set_regex_collation: set collation for these functions to obey
*
* This is called when beginning compilation or execution of a regexp.
- * Since there's no need for re-entrancy of regexp operations, it's okay
+ * Since there's no need for reentrancy of regexp operations, it's okay
* to store the results in static variables.
*/
void
diff --git a/src/backend/replication/logical/launcher.c
b/src/backend/replication/logical/launcher.c
--- a/src/backend/replication/logical/launcher.c
+++ b/src/backend/replication/logical/launcher.c
@@ -170,7 +170,7 @@ WaitForReplicationWorkerAttach(LogicalRe
/*
* Worker started and attached to our shmem. This check is safe
- * because only laucher ever starts the workers, so nobody can
steal
+ * because only launcher ever starts the workers, so nobody can
steal
* the worker slot.
*/
if (status == BGWH_STARTED && worker->proc)
@@ -180,7 +180,7 @@ WaitForReplicationWorkerAttach(LogicalRe
return false;
/*
- * We need timeout because we generaly don't get notified via
latch
+ * We need timeout because we generally don't get notified via
latch
* about the worker attach.
*/
rc = WaitLatch(MyLatch,
@@ -533,7 +533,7 @@ AtCommit_ApplyLauncher(void)
/*
* Request wakeup of the launcher on commit of the transaction.
*
- * This is used to send launcher signal to stop sleeping and proccess the
+ * This is used to send launcher signal to stop sleeping and process the
* subscriptions when current transaction commits. Should be used when new
* tuple was added to the pg_subscription catalog.
*/
@@ -638,7 +638,7 @@ ApplyLauncherMain(Datum main_arg)
else
{
/*
- * The wait in previous cycle was interruped in less
than
+ * The wait in previous cycle was interrupted in less
than
* wal_retrieve_retry_interval since last worker was
started,
* this usually means crash of the worker, so we should
retry
* in wal_retrieve_retry_interval again.
diff --git a/src/backend/replication/logical/origin.c
b/src/backend/replication/logical/origin.c
--- a/src/backend/replication/logical/origin.c
+++ b/src/backend/replication/logical/origin.c
@@ -1250,7 +1250,7 @@ pg_replication_origin_session_is_setup(P
* Return the replication progress for origin setup in the current session.
*
* If 'flush' is set to true it is ensured that the returned value corresponds
- * to a local transaction that has been flushed. this is useful if asychronous
+ * to a local transaction that has been flushed. this is useful if asynchronous
* commits are used when replaying replicated transactions.
*/
Datum
@@ -1336,7 +1336,7 @@ pg_replication_origin_advance(PG_FUNCTIO
* Return the replication progress for an individual replication origin.
*
* If 'flush' is set to true it is ensured that the returned value corresponds
- * to a local transaction that has been flushed. this is useful if asychronous
+ * to a local transaction that has been flushed. this is useful if asynchronous
* commits are used when replaying replicated transactions.
*/
Datum
diff --git a/src/backend/replication/logical/proto.c
b/src/backend/replication/logical/proto.c
--- a/src/backend/replication/logical/proto.c
+++ b/src/backend/replication/logical/proto.c
@@ -539,7 +539,7 @@ logicalrep_write_attrs(StringInfo out, R
if (att->attisdropped)
continue;
- /* REPLICA IDENTITY FULL means all colums are sent as part of
key. */
+ /* REPLICA IDENTITY FULL means all columns are sent as part of
key. */
if (replidentfull ||
bms_is_member(att->attnum -
FirstLowInvalidHeapAttributeNumber,
idattrs))
diff --git a/src/backend/replication/logical/reorderbuffer.c
b/src/backend/replication/logical/reorderbuffer.c
--- a/src/backend/replication/logical/reorderbuffer.c
+++ b/src/backend/replication/logical/reorderbuffer.c
@@ -1714,7 +1714,7 @@ ReorderBufferCommit(ReorderBuffer *rb, T
*
* NB: Transactions handled here have to have actively aborted (i.e. have
* produced an abort record). Implicitly aborted transactions are handled via
- * ReorderBufferAbortOld(); transactions we're just not interesteded in, but
+ * ReorderBufferAbortOld(); transactions we're just not interested in, but
* which have committed are handled in ReorderBufferForget().
*
* This function purges this transaction and its contents from memory and
@@ -1782,7 +1782,7 @@ ReorderBufferAbortOld(ReorderBuffer *rb,
* toplevel xid.
*
* This is significantly different to ReorderBufferAbort() because
- * transactions that have committed need to be treated differenly from aborted
+ * transactions that have committed need to be treated differently from aborted
* ones since they may have modified the catalog.
*
* Note that this is only allowed to be called in the moment a transaction
@@ -2660,7 +2660,7 @@ StartupReorderBuffer(void)
/*
* ok, has to be a surviving logical slot, iterate and delete
- * everythign starting with xid-*
+ * everything starting with xid-*
*/
sprintf(path, "pg_replslot/%s", logical_de->d_name);
diff --git a/src/backend/replication/logical/snapbuild.c
b/src/backend/replication/logical/snapbuild.c
--- a/src/backend/replication/logical/snapbuild.c
+++ b/src/backend/replication/logical/snapbuild.c
@@ -614,7 +614,7 @@ SnapBuildGetOrBuildSnapshot(SnapBuild *b
if (builder->snapshot == NULL)
{
builder->snapshot = SnapBuildBuildSnapshot(builder, xid);
- /* inrease refcount for the snapshot builder */
+ /* increase refcount for the snapshot builder */
SnapBuildSnapIncRefcount(builder->snapshot);
}
@@ -678,7 +678,7 @@ SnapBuildProcessChange(SnapBuild *builde
if (builder->snapshot == NULL)
{
builder->snapshot = SnapBuildBuildSnapshot(builder,
xid);
- /* inrease refcount for the snapshot builder */
+ /* increase refcount for the snapshot builder */
SnapBuildSnapIncRefcount(builder->snapshot);
}
@@ -911,7 +911,7 @@ SnapBuildEndTxn(SnapBuild *builder, XLog
{
/*
* None of the originally running transaction is
running anymore,
- * so our incrementaly built snapshot now is consistent.
+ * so our incrementally built snapshot now is
consistent.
*/
ereport(LOG,
(errmsg("logical decoding found consistent
point at %X/%X",
diff --git a/src/backend/replication/logical/worker.c
b/src/backend/replication/logical/worker.c
--- a/src/backend/replication/logical/worker.c
+++ b/src/backend/replication/logical/worker.c
@@ -327,7 +327,7 @@ slot_store_cstrings(TupleTableSlot *slot
/*
* Modify slot with user data provided as C strigs.
* This is somewhat similar to heap_modify_tuple but also calls the type
- * input fuction on the user data as the input is the text representation
+ * input function on the user data as the input is the text representation
* of the types.
*/
static void
diff --git a/src/backend/replication/pgoutput/pgoutput.c
b/src/backend/replication/pgoutput/pgoutput.c
--- a/src/backend/replication/pgoutput/pgoutput.c
+++ b/src/backend/replication/pgoutput/pgoutput.c
@@ -172,7 +172,7 @@ pgoutput_startup(LogicalDecodingContext
&data->protocol_version,
&data->publication_names);
- /* Check if we support requested protol */
+ /* Check if we support requested protocol */
if (data->protocol_version != LOGICALREP_PROTO_VERSION_NUM)
ereport(ERROR,
(errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
@@ -424,7 +424,7 @@ publication_invalidation_cb(Datum arg, i
/*
* Initialize the relation schema sync cache for a decoding session.
*
- * The hash table is destoyed at the end of a decoding session. While
+ * The hash table is destroyed at the end of a decoding session. While
* relcache invalidations still exist and will still be invoked, they
* will just see the null hash table global and take no action.
*/
@@ -540,7 +540,7 @@ rel_sync_cache_relation_cb(Datum arg, Oi
/*
* We can get here if the plugin was used in SQL interface as the
- * RelSchemaSyncCache is detroyed when the decoding finishes, but there
+ * RelSchemaSyncCache is destroyed when the decoding finishes, but there
* is no way to unregister the relcache invalidation callback.
*/
if (RelationSyncCache == NULL)
@@ -580,7 +580,7 @@ rel_sync_cache_publication_cb(Datum arg,
/*
* We can get here if the plugin was used in SQL interface as the
- * RelSchemaSyncCache is detroyed when the decoding finishes, but there
+ * RelSchemaSyncCache is destroyed when the decoding finishes, but there
* is no way to unregister the relcache invalidation callback.
*/
if (RelationSyncCache == NULL)
diff --git a/src/backend/storage/ipc/latch.c b/src/backend/storage/ipc/latch.c
--- a/src/backend/storage/ipc/latch.c
+++ b/src/backend/storage/ipc/latch.c
@@ -860,7 +860,7 @@ WaitEventAdjustWin32(WaitEventSet *set,
* reached. At most nevents occurred events are returned.
*
* If timeout = -1, block until an event occurs; if 0, check sockets for
- * readiness, but don't block; if > 0, block for at most timeout miliseconds.
+ * readiness, but don't block; if > 0, block for at most timeout milliseconds.
*
* Returns the number of events occurred, or 0 if the timeout was reached.
*
diff --git a/src/backend/storage/ipc/procarray.c
b/src/backend/storage/ipc/procarray.c
--- a/src/backend/storage/ipc/procarray.c
+++ b/src/backend/storage/ipc/procarray.c
@@ -280,7 +280,7 @@ ProcArrayAdd(PGPROC *proc)
if (arrayP->numProcs >= arrayP->maxProcs)
{
/*
- * Ooops, no room. (This really shouldn't happen, since there
is a
+ * Oops, no room. (This really shouldn't happen, since there
is a
* fixed supply of PGPROC structs too, and so we should have
failed
* earlier.)
*/
@@ -370,7 +370,7 @@ ProcArrayRemove(PGPROC *proc, Transactio
}
}
- /* Ooops */
+ /* Oops */
LWLockRelease(ProcArrayLock);
elog(LOG, "failed to find proc %p in ProcArray", proc);
diff --git a/src/backend/storage/ipc/shm_mq.c b/src/backend/storage/ipc/shm_mq.c
--- a/src/backend/storage/ipc/shm_mq.c
+++ b/src/backend/storage/ipc/shm_mq.c
@@ -501,7 +501,7 @@ shm_mq_sendv(shm_mq_handle *mqh, shm_mq_
* it will point to a temporary buffer. This mostly avoids data copying in
* the hoped-for case where messages are short compared to the buffer size,
* while still allowing longer messages. In either case, the return value
- * remains valid until the next receive operation is perfomed on the queue.
+ * remains valid until the next receive operation is performed on the queue.
*
* When nowait = false, we'll wait on our process latch when the ring buffer
* is empty and we have not yet received a full message. The sender will
diff --git a/src/backend/storage/ipc/standby.c
b/src/backend/storage/ipc/standby.c
--- a/src/backend/storage/ipc/standby.c
+++ b/src/backend/storage/ipc/standby.c
@@ -967,7 +967,7 @@ LogStandbySnapshot(void)
* similar. We keep them separate because xl_xact_running_xacts is a
* contiguous chunk of memory and never exists fully until it is assembled in
* WAL. The inserted records are marked as not being important for durability,
- * to avoid triggering superflous checkpoint / archiving activity.
+ * to avoid triggering superfluous checkpoint / archiving activity.
*/
static XLogRecPtr
LogCurrentRunningXacts(RunningTransactions CurrRunningXacts)
diff --git a/src/backend/storage/lmgr/lock.c b/src/backend/storage/lmgr/lock.c
--- a/src/backend/storage/lmgr/lock.c
+++ b/src/backend/storage/lmgr/lock.c
@@ -1125,7 +1125,7 @@ SetupLockInTable(LockMethod lockMethodTa
&found);
if (!proclock)
{
- /* Ooops, not enough shmem for the proclock */
+ /* Oops, not enough shmem for the proclock */
if (lock->nRequested == 0)
{
/*
@@ -2778,7 +2778,7 @@ GetLockConflicts(const LOCKTAG *locktag,
vxids = (VirtualTransactionId *)
palloc0(sizeof(VirtualTransactionId) * (MaxBackends +
1));
- /* Compute hash code and partiton lock, and look up conflicting modes.
*/
+ /* Compute hash code and partition lock, and look up conflicting modes.
*/
hashcode = LockTagHashCode(locktag);
partitionLock = LockHashPartitionLock(hashcode);
conflictMask = lockMethodTable->conflictTab[lockmode];
@@ -4046,7 +4046,7 @@ lock_twophase_recover(TransactionId xid,
&found);
if (!proclock)
{
- /* Ooops, not enough shmem for the proclock */
+ /* Oops, not enough shmem for the proclock */
if (lock->nRequested == 0)
{
/*
diff --git a/src/backend/storage/lmgr/lwlock.c
b/src/backend/storage/lmgr/lwlock.c
--- a/src/backend/storage/lmgr/lwlock.c
+++ b/src/backend/storage/lmgr/lwlock.c
@@ -781,7 +781,7 @@ LWLockAttemptLock(LWLock *lock, LWLockMo
return false;
}
else
- return true; /* someobdy else has the lock */
+ return true; /* somebody else has the lock */
}
}
pg_unreachable();
@@ -953,7 +953,7 @@ LWLockWakeup(LWLock *lock)
* that happens before the list unlink happens, the list would
end up
* being corrupted.
*
- * The barrier pairs with the LWLockWaitListLock() when
enqueueing for
+ * The barrier pairs with the LWLockWaitListLock() when
enqueuing for
* another lock.
*/
pg_write_barrier();
@@ -1029,7 +1029,7 @@ LWLockDequeueSelf(LWLock *lock)
/*
* Can't just remove ourselves from the list, but we need to iterate
over
- * all entries as somebody else could have unqueued us.
+ * all entries as somebody else could have dequeued us.
*/
proclist_foreach_modify(iter, &lock->waiters, lwWaitLink)
{
diff --git a/src/backend/storage/lmgr/predicate.c
b/src/backend/storage/lmgr/predicate.c
--- a/src/backend/storage/lmgr/predicate.c
+++ b/src/backend/storage/lmgr/predicate.c
@@ -3193,7 +3193,7 @@ ReleasePredicateLocks(bool isCommit)
/*
* We can't trust XactReadOnly here, because a transaction which started
* as READ WRITE can show as READ ONLY later, e.g., within
- * substransactions. We want to flag a transaction as READ ONLY if it
+ * subtransactions. We want to flag a transaction as READ ONLY if it
* commits without writing so that de facto READ ONLY transactions get
the
* benefit of some RO optimizations, so we will use this local variable
to
* get some cleanup logic right which is based on whether the
transaction
diff --git a/src/backend/storage/smgr/md.c b/src/backend/storage/smgr/md.c
--- a/src/backend/storage/smgr/md.c
+++ b/src/backend/storage/smgr/md.c
@@ -1728,7 +1728,7 @@ static void
else
{
/*
- * It doesn't seem worthwile complicating the code by having a
more
+ * It doesn't seem worthwhile complicating the code by having a
more
* aggressive growth strategy here; the number of segments
doesn't
* grow that fast, and the memory context internally will
sometimes
* avoid doing an actual reallocation.
diff --git a/src/backend/tsearch/dict_thesaurus.c
b/src/backend/tsearch/dict_thesaurus.c
--- a/src/backend/tsearch/dict_thesaurus.c
+++ b/src/backend/tsearch/dict_thesaurus.c
@@ -23,7 +23,7 @@
/*
- * Temporay we use TSLexeme.flags for inner use...
+ * Temporary we use TSLexeme.flags for inner use...
*/
#define DT_USEASIS 0x1000
diff --git a/src/backend/tsearch/spell.c b/src/backend/tsearch/spell.c
--- a/src/backend/tsearch/spell.c
+++ b/src/backend/tsearch/spell.c
@@ -37,7 +37,7 @@
* Spell field. The AffixData field is initialized if AF parameter is not
* defined.
* - NISortAffixes():
- * - builds a list of compond affixes from the affix list and stores it
+ * - builds a list of compound affixes from the affix list and stores it
* in the CompoundAffix.
* - builds prefix trees (Trie) from the affix list for prefixes and
suffixes
* and stores them in Suffix and Prefix fields.
diff --git a/src/backend/tsearch/ts_parse.c b/src/backend/tsearch/ts_parse.c
--- a/src/backend/tsearch/ts_parse.c
+++ b/src/backend/tsearch/ts_parse.c
@@ -179,7 +179,7 @@ LexizeExec(LexizeData *ld, ParsedLex **c
if (ld->curDictId == InvalidOid)
{
/*
- * usial mode: dictionary wants only one word, but we should
keep in
+ * usual mode: dictionary wants only one word, but we should
keep in
* mind that we should go through all stack
*/
@@ -272,7 +272,7 @@ LexizeExec(LexizeData *ld, ParsedLex **c
/*
* We should be sure that current type of
lexeme is recognized
- * by our dictinonary: we just check is it
exist in list of
+ * by our dictionary: we just check is it exist
in list of
* dictionaries ?
*/
for (i = 0; i < map->len && !dictExists; i++)
@@ -627,7 +627,7 @@ generateHeadline(HeadlineParsedText *prs
/* start of a new fragment */
infrag = 1;
numfragments++;
- /* add a fragment delimitor if this is after
the first one */
+ /* add a fragment delimiter if this is after
the first one */
if (numfragments > 1)
{
memcpy(ptr, prs->fragdelim,
prs->fragdelimlen);
diff --git a/src/backend/tsearch/wparser_def.c
b/src/backend/tsearch/wparser_def.c
--- a/src/backend/tsearch/wparser_def.c
+++ b/src/backend/tsearch/wparser_def.c
@@ -2445,7 +2445,7 @@ mark_hl_words(HeadlineParsedText *prs, T
break;
}
if (curlen < min_words && i >= prs->curwords)
- { /* got end of
text and our cover is shoter
+ { /* got end of
text and our cover is shorter
* than
min_words */
for (i = p - 1; i >= 0; i--)
{
diff --git a/src/backend/utils/adt/datetime.c b/src/backend/utils/adt/datetime.c
--- a/src/backend/utils/adt/datetime.c
+++ b/src/backend/utils/adt/datetime.c
@@ -85,7 +85,7 @@ const char *const days[] = {"Sunday", "M
* Note that this table must be strictly alphabetically ordered to allow an
* O(ln(N)) search algorithm to be used.
*
- * The token field must be NUL-terminated; we truncate entries to TOKMAXLEN
+ * The token field must be NULL-terminated; we truncate entries to TOKMAXLEN
* characters to fit.
*
* The static table contains no TZ, DTZ, or DYNTZ entries; rather those
diff --git a/src/backend/utils/adt/formatting.c
b/src/backend/utils/adt/formatting.c
--- a/src/backend/utils/adt/formatting.c
+++ b/src/backend/utils/adt/formatting.c
@@ -2265,7 +2265,7 @@ seq_search(char *name, const char *const
for (last = 0, a = array; *a != NULL; a++)
{
- /* comperate first chars */
+ /* compare first chars */
if (*name != **a)
continue;
diff --git a/src/backend/utils/adt/rangetypes_selfuncs.c
b/src/backend/utils/adt/rangetypes_selfuncs.c
--- a/src/backend/utils/adt/rangetypes_selfuncs.c
+++ b/src/backend/utils/adt/rangetypes_selfuncs.c
@@ -533,7 +533,7 @@ calc_hist_selectivity(TypeCacheEntry *ty
{
/*
* Lower bound no longer matters. Just estimate
the fraction
- * with an upper bound <= const uppert bound
+ * with an upper bound <= const upper bound
*/
hist_selec =
calc_hist_selectivity_scalar(typcache,
&const_upper,
diff --git a/src/backend/utils/adt/ruleutils.c
b/src/backend/utils/adt/ruleutils.c
--- a/src/backend/utils/adt/ruleutils.c
+++ b/src/backend/utils/adt/ruleutils.c
@@ -2687,7 +2687,7 @@ is_input_argument(int nth, const char *a
}
/*
- * Append used transformated types to specified buffer
+ * Append used transformed types to specified buffer
*/
static void
print_function_trftypes(StringInfo buf, HeapTuple proctup)
diff --git a/src/backend/utils/adt/selfuncs.c b/src/backend/utils/adt/selfuncs.c
--- a/src/backend/utils/adt/selfuncs.c
+++ b/src/backend/utils/adt/selfuncs.c
@@ -4329,7 +4329,7 @@ get_restriction_variable(PlannerInfo *ro
return true;
}
- /* Ooops, clause has wrong structure (probably var op var) */
+ /* Oops, clause has wrong structure (probably var op var) */
ReleaseVariableStats(*vardata);
ReleaseVariableStats(rdata);
diff --git a/src/backend/utils/adt/tsrank.c b/src/backend/utils/adt/tsrank.c
--- a/src/backend/utils/adt/tsrank.c
+++ b/src/backend/utils/adt/tsrank.c
@@ -899,7 +899,7 @@ calc_rank_cd(const float4 *arrdata, TSVe
/*
* if doc are big enough then ext.q may be equal to ext.p due
to limit
- * of posional information. In this case we approximate number
of
+ * of positional information. In this case we approximate
number of
* noise word as half cover's length
*/
nNoise = (ext.q - ext.p) - (ext.end - ext.begin);
@@ -908,7 +908,7 @@ calc_rank_cd(const float4 *arrdata, TSVe
Wdoc += Cpos / ((double) (1 + nNoise));
CurExtPos = ((double) (ext.q + ext.p)) / 2.0;
- if (NExtent > 0 && CurExtPos > PrevExtPos /*
prevent devision by
+ if (NExtent > 0 && CurExtPos > PrevExtPos /*
prevent division by
* zero in a case of
multiple lexize */ )
SumDist += 1.0 / (CurExtPos - PrevExtPos);
diff --git a/src/backend/utils/adt/varlena.c b/src/backend/utils/adt/varlena.c
--- a/src/backend/utils/adt/varlena.c
+++ b/src/backend/utils/adt/varlena.c
@@ -2222,7 +2222,7 @@ varstr_abbrev_convert(Datum original, So
goto done;
}
- /* Just like strcoll(), strxfrm() expects a NUL-terminated
string */
+ /* Just like strcoll(), strxfrm() expects a NULL-terminated
string */
memcpy(sss->buf1, authoritative_data, len);
sss->buf1[len] = '\0';
sss->last_len1 = len;
diff --git a/src/backend/utils/adt/windowfuncs.c
b/src/backend/utils/adt/windowfuncs.c
--- a/src/backend/utils/adt/windowfuncs.c
+++ b/src/backend/utils/adt/windowfuncs.c
@@ -342,7 +342,7 @@ window_lag(PG_FUNCTION_ARGS)
/*
* lag_with_offset
- * returns the value of VE evelulated on a row that is OFFSET
+ * returns the value of VE evaluated on a row that is OFFSET
* rows before the current row within a partition,
* per spec.
*/
diff --git a/src/backend/utils/cache/lsyscache.c
b/src/backend/utils/cache/lsyscache.c
--- a/src/backend/utils/cache/lsyscache.c
+++ b/src/backend/utils/cache/lsyscache.c
@@ -2332,7 +2332,7 @@ get_typavgwidth(Oid typid, int32 typmod)
}
/*
- * Ooops, we have no idea ... wild guess time.
+ * Oops, we have no idea ... wild guess time.
*/
return 32;
}
diff --git a/src/backend/utils/cache/plancache.c
b/src/backend/utils/cache/plancache.c
--- a/src/backend/utils/cache/plancache.c
+++ b/src/backend/utils/cache/plancache.c
@@ -621,7 +621,7 @@ RevalidateCachedQuery(CachedPlanSource *
return NIL;
}
- /* Ooops, the race case happened. Release useless locks. */
+ /* Oops, the race case happened. Release useless locks. */
AcquirePlannerLocks(plansource->query_list, false);
}
@@ -845,7 +845,7 @@ CheckCachedPlan(CachedPlanSource *planso
return true;
}
- /* Ooops, the race case happened. Release useless locks. */
+ /* Oops, the race case happened. Release useless locks. */
AcquireExecutorLocks(plan->stmt_list, false);
}
diff --git a/src/backend/utils/cache/relcache.c
b/src/backend/utils/cache/relcache.c
--- a/src/backend/utils/cache/relcache.c
+++ b/src/backend/utils/cache/relcache.c
@@ -1433,7 +1433,7 @@ RelationInitPhysicalAddr(Relation relati
* points to the current file since the older file will be gone
(or
* truncated). The new file will still contain older rows so
lookups
* in them will work correctly. This wouldn't work correctly if
- * rewrites were allowed to change the schema in a
noncompatible way,
+ * rewrites were allowed to change the schema in an
incompatible way,
* but those are prevented both on catalog tables and on user
tables
* declared as additional catalog tables.
*/
diff --git a/src/backend/utils/error/elog.c b/src/backend/utils/error/elog.c
--- a/src/backend/utils/error/elog.c
+++ b/src/backend/utils/error/elog.c
@@ -318,7 +318,7 @@ errstart(int elevel, const char *filenam
*/
if (ErrorContext == NULL)
{
- /* Ooops, hard crash time; very little we can do safely here */
+ /* Oops, hard crash time; very little we can do safely here */
write_stderr("error occurred at %s:%d before error message
processing is available\n",
filename ? filename : "(unknown
file)", lineno);
exit(2);
@@ -331,7 +331,7 @@ errstart(int elevel, const char *filenam
if (recursion_depth++ > 0 && elevel >= ERROR)
{
/*
- * Ooops, error during error processing. Clear ErrorContext as
+ * Oops, error during error processing. Clear ErrorContext as
* discussed at top of file. We will not return to the original
* error's reporter or handler, so we don't need it.
*/
@@ -1302,7 +1302,7 @@ elog_start(const char *filename, int lin
/* Make sure that memory context initialization has finished */
if (ErrorContext == NULL)
{
- /* Ooops, hard crash time; very little we can do safely here */
+ /* Oops, hard crash time; very little we can do safely here */
write_stderr("error occurred at %s:%d before error message
processing is available\n",
filename ? filename : "(unknown
file)", lineno);
exit(2);
diff --git a/src/backend/utils/fmgr/fmgr.c b/src/backend/utils/fmgr/fmgr.c
--- a/src/backend/utils/fmgr/fmgr.c
+++ b/src/backend/utils/fmgr/fmgr.c
@@ -877,7 +877,7 @@ struct fmgr_security_definer_cache
* To execute a call, we temporarily replace the flinfo with the cached
* and looked-up one, while keeping the outer fcinfo (which contains all
* the actual arguments, etc.) intact. This is not re-entrant, but then
- * the fcinfo itself can't be used re-entrantly anyway.
+ * the fcinfo itself can't be used reentrantly anyway.
*/
static Datum
fmgr_security_definer(PG_FUNCTION_ARGS)
diff --git a/src/backend/utils/fmgr/funcapi.c b/src/backend/utils/fmgr/funcapi.c
--- a/src/backend/utils/fmgr/funcapi.c
+++ b/src/backend/utils/fmgr/funcapi.c
@@ -879,7 +879,7 @@ get_func_arg_info(HeapTuple procTup,
/*
* get_func_trftypes
*
- * Returns a number of transformated types used by function.
+ * Returns a number of transformed types used by function.
*/
int
get_func_trftypes(HeapTuple procTup,
diff --git a/src/backend/utils/hash/dynahash.c
b/src/backend/utils/hash/dynahash.c
--- a/src/backend/utils/hash/dynahash.c
+++ b/src/backend/utils/hash/dynahash.c
@@ -72,7 +72,7 @@
* when combined with HASH_DEBUG, these are displayed by hdestroy().
*
* Problems & fixes to [email protected]. WARNING: relies on pre-processor
- * concatenation property, in probably unnecessary code 'optimisation'.
+ * concatenation property, in probably unnecessary code 'optimization'.
*
* Modified [email protected] February 1990
* added multiple table interface
diff --git a/src/backend/utils/hash/hashfn.c b/src/backend/utils/hash/hashfn.c
--- a/src/backend/utils/hash/hashfn.c
+++ b/src/backend/utils/hash/hashfn.c
@@ -26,7 +26,7 @@
/*
- * string_hash: hash function for keys that are NUL-terminated strings.
+ * string_hash: hash function for keys that are NULL-terminated strings.
*
* NOTE: this is the default hash function if none is specified.
*/
diff --git a/src/backend/utils/init/postinit.c
b/src/backend/utils/init/postinit.c
--- a/src/backend/utils/init/postinit.c
+++ b/src/backend/utils/init/postinit.c
@@ -1108,7 +1108,7 @@ process_settings(Oid databaseid, Oid rol
relsetting = heap_open(DbRoleSettingRelationId, AccessShareLock);
- /* read all the settings under the same snapsot for efficiency */
+ /* read all the settings under the same snapshot for efficiency */
snapshot =
RegisterSnapshot(GetCatalogSnapshot(DbRoleSettingRelationId));
/* Later settings are ignored if set earlier. */
diff --git a/src/backend/utils/misc/Makefile b/src/backend/utils/misc/Makefile
--- a/src/backend/utils/misc/Makefile
+++ b/src/backend/utils/misc/Makefile
@@ -19,7 +19,7 @@ OBJS = backend_random.o guc.o help_confi
tzparser.o
# This location might depend on the installation directories. Therefore
-# we can't subsitute it into pg_config.h.
+# we can't substitute it into pg_config.h.
ifdef krb_srvtab
override CPPFLAGS += -DPG_KRB_SRVTAB='"$(krb_srvtab)"'
endif
diff --git a/src/backend/utils/misc/README b/src/backend/utils/misc/README
--- a/src/backend/utils/misc/README
+++ b/src/backend/utils/misc/README
@@ -114,7 +114,7 @@ If a show_hook is provided, it points to
This hook allows variable-specific computation of the value displayed
by SHOW (and other SQL features for showing GUC variable values).
The return value can point to a static buffer, since show functions are
-not used re-entrantly.
+not used reentrantly.
Saving/Restoring GUC Variable Values
diff --git a/src/backend/utils/mmgr/freepage.c
b/src/backend/utils/mmgr/freepage.c
--- a/src/backend/utils/mmgr/freepage.c
+++ b/src/backend/utils/mmgr/freepage.c
@@ -318,7 +318,7 @@ sum_free_pages(FreePageManager *fpm)
/*
* Compute the size of the largest run of pages that the user could
- * succesfully get.
+ * successfully get.
*/
static Size
FreePageManagerLargestContiguous(FreePageManager *fpm)
@@ -360,7 +360,7 @@ FreePageManagerLargestContiguous(FreePag
/*
* Recompute the size of the largest run of pages that the user could
- * succesfully get, if it has been marked dirty.
+ * successfully get, if it has been marked dirty.
*/
static void
FreePageManagerUpdateLargest(FreePageManager *fpm)
@@ -1704,7 +1704,7 @@ FreePageManagerPutInternal(FreePageManag
* The act of allocating pages for use in constructing
our btree
* should never cause any page to become more full, so
the new
* split depth should be no greater than the old one,
and perhaps
- * less if we fortutiously allocated a chunk that freed
up a slot
+ * less if we fortuitously allocated a chunk that freed
up a slot
* on the page we need to update.
*/
Assert(result.split_pages <= fpm->btree_recycle_count);
diff --git a/src/backend/utils/time/tqual.c b/src/backend/utils/time/tqual.c
--- a/src/backend/utils/time/tqual.c
+++ b/src/backend/utils/time/tqual.c
@@ -1625,7 +1625,7 @@ HeapTupleHeaderIsOnlyLocked(HeapTupleHea
}
/*
- * check whether the transaciont id 'xid' is in the pre-sorted array 'xip'.
+ * check whether the transaction id 'xid' is in the pre-sorted array 'xip'.
*/
static bool
TransactionIdInArray(TransactionId xid, TransactionId *xip, Size num)
diff --git a/src/bin/pg_dump/parallel.c b/src/bin/pg_dump/parallel.c
--- a/src/bin/pg_dump/parallel.c
+++ b/src/bin/pg_dump/parallel.c
@@ -94,7 +94,7 @@ struct ParallelSlot
/* These fields are valid if workerStatus == WRKR_WORKING: */
ParallelCompletionPtr callback; /* function to call on
completion */
- void *callback_data; /* passthru data for it */
+ void *callback_data; /* passthrough data for it */
ArchiveHandle *AH; /* Archive data worker is using
*/
diff --git a/src/bin/pg_dump/pg_backup_custom.c
b/src/bin/pg_dump/pg_backup_custom.c
--- a/src/bin/pg_dump/pg_backup_custom.c
+++ b/src/bin/pg_dump/pg_backup_custom.c
@@ -198,7 +198,7 @@ InitArchiveFmt_Custom(ArchiveHandle *AH)
*
* Optional.
*
- * Set up extrac format-related TOC data.
+ * Set up extract format-related TOC data.
*/
static void
_ArchiveEntry(ArchiveHandle *AH, TocEntry *te)
diff --git a/src/bin/pg_dump/pg_dump.c b/src/bin/pg_dump/pg_dump.c
--- a/src/bin/pg_dump/pg_dump.c
+++ b/src/bin/pg_dump/pg_dump.c
@@ -3500,7 +3500,7 @@ getPublicationTables(Archive *fout, Tabl
resetPQExpBuffer(query);
- /* Get the publication memebership for the table. */
+ /* Get the publication membership for the table. */
appendPQExpBuffer(query,
"SELECT pr.tableoid, pr.oid,
p.pubname "
"FROM
pg_catalog.pg_publication_rel pr,"
diff --git a/src/bin/psql/common.c b/src/bin/psql/common.c
--- a/src/bin/psql/common.c
+++ b/src/bin/psql/common.c
@@ -828,7 +828,7 @@ StoreQueryTuple(const PGresult *result)
char *varname;
char *value;
- /* concate prefix and column name */
+ /* concatenate prefix and column name */
varname = psprintf("%s%s", pset.gset_prefix, colname);
if (!PQgetisnull(result, 0, i))
diff --git a/src/bin/psql/describe.c b/src/bin/psql/describe.c
--- a/src/bin/psql/describe.c
+++ b/src/bin/psql/describe.c
@@ -2127,7 +2127,7 @@ describeOneTableDetails(const char *sche
printTableAddFooter(&cont, _("Check
constraints:"));
for (i = 0; i < tuples; i++)
{
- /* untranslated contraint name and def
*/
+ /* untranslated constraint name and def
*/
printfPQExpBuffer(&buf, " \"%s\" %s",
PQgetvalue(result, i, 0),
PQgetvalue(result, i, 1));
@@ -3197,7 +3197,7 @@ listTables(const char *tabtypes, const c
if (verbose)
{
/*
- * As of PostgreSQL 9.0, use pg_table_size() to show a more
acurate
+ * As of PostgreSQL 9.0, use pg_table_size() to show a more
accurate
* size of a table, including FSM, VM and TOAST tables.
*/
if (pset.sversion >= 90000)
@@ -5108,7 +5108,7 @@ describeSubscriptions(const char *patter
gettext_noop("Conninfo"));
}
- /* Only display subscritpions in current database. */
+ /* Only display subscriptions in current database. */
appendPQExpBufferStr(&buf,
"FROM
pg_catalog.pg_subscription\n"
"WHERE subdbid = (SELECT oid\n"
diff --git a/src/fe_utils/string_utils.c b/src/fe_utils/string_utils.c
--- a/src/fe_utils/string_utils.c
+++ b/src/fe_utils/string_utils.c
@@ -173,7 +173,7 @@ fmtQualifiedId(int remoteVersion, const
* returned by PQserverVersion()) as a string. This exists mainly to
* encapsulate knowledge about two-part vs. three-part version numbers.
*
- * For re-entrancy, caller must supply the buffer the string is put in.
+ * For reentrancy, caller must supply the buffer the string is put in.
* Recommended size of the buffer is 32 bytes.
*
* Returns address of 'buf', as a notational convenience.
diff --git a/src/include/access/visibilitymap.h
b/src/include/access/visibilitymap.h
--- a/src/include/access/visibilitymap.h
+++ b/src/include/access/visibilitymap.h
@@ -26,7 +26,7 @@
#define VISIBILITYMAP_ALL_VISIBLE 0x01
#define VISIBILITYMAP_ALL_FROZEN 0x02
#define VISIBILITYMAP_VALID_BITS 0x03 /* OR of all valid
-
* visiblitymap flags bits */
+
* visibilitymap flags bits */
/* Macros for visibilitymap test */
#define VM_ALL_VISIBLE(r, b, v) \
diff --git a/src/include/access/xact.h b/src/include/access/xact.h
--- a/src/include/access/xact.h
+++ b/src/include/access/xact.h
@@ -65,7 +65,7 @@ typedef enum
* apply */
} SyncCommitLevel;
-/* Define the default setting for synchonous_commit */
+/* Define the default setting for synchronous_commit */
#define SYNCHRONOUS_COMMIT_ON SYNCHRONOUS_COMMIT_REMOTE_FLUSH
/* Synchronous commit level */
diff --git a/src/include/c.h b/src/include/c.h
--- a/src/include/c.h
+++ b/src/include/c.h
@@ -989,7 +989,7 @@ typedef NameData *Name;
/* gettext domain name mangling */
/*
- * To better support parallel installations of major PostgeSQL
+ * To better support parallel installations of major PostgreSQL
* versions as well as parallel installations of major library soname
* versions, we mangle the gettext domain name by appending those
* version numbers. The coding rule ought to be that wherever the
diff --git a/src/include/catalog/partition.h b/src/include/catalog/partition.h
--- a/src/include/catalog/partition.h
+++ b/src/include/catalog/partition.h
@@ -41,7 +41,7 @@ typedef struct PartitionDescData *Partit
/*-----------------------
* PartitionDispatch - information about one partitioned table in a partition
- * hiearchy required to route a tuple to one of its partitions
+ * hierarchy required to route a tuple to one of its partitions
*
* reldesc Relation descriptor of the table
* key Partition key information of the table
diff --git a/src/include/catalog/pg_subscription.h
b/src/include/catalog/pg_subscription.h
--- a/src/include/catalog/pg_subscription.h
+++ b/src/include/catalog/pg_subscription.h
@@ -23,7 +23,7 @@
#define SubscriptionRelation_Rowtype_Id 6101
/*
- * Technicaly, the subscriptions live inside the database, so a shared catalog
+ * Technically, the subscriptions live inside the database, so a shared catalog
* seems weird, but the replication launcher process needs to access all of
* them to be able to start the workers, so we have to put them in a shared,
* nailed catalog.
@@ -35,7 +35,7 @@ CATALOG(pg_subscription,6100) BKI_SHARED
Oid subowner; /* Owner of the
subscription */
- bool subenabled; /* True if the subsription is
enabled
+ bool subenabled; /* True if the subscription is
enabled
* (the worker
should be running) */
#ifdef CATALOG_VARLEN /* variable-length fields start here */
@@ -65,7 +65,7 @@ typedef FormData_pg_subscription *Form_p
typedef struct Subscription
{
Oid oid; /* Oid of the subscription */
- Oid dbid; /* Oid of the database which
dubscription is in */
+ Oid dbid; /* Oid of the database which
subscription is in */
char *name; /* Name of the subscription */
Oid owner; /* Oid of the subscription
owner */
bool enabled; /* Indicates if the subscription is
enabled */
diff --git a/src/include/lib/simplehash.h b/src/include/lib/simplehash.h
--- a/src/include/lib/simplehash.h
+++ b/src/include/lib/simplehash.h
@@ -345,7 +345,7 @@ SH_GROW(SH_TYPE *tb, uint32 newsize)
* we need. We neither want tb->members increased, nor do we need to do
* deal with deleted elements, nor do we need to compare keys. So a
* special-cased implementation is lot faster. As resizing can be time
- * consuming and frequent, that's worthwile to optimize.
+ * consuming and frequent, that's worthwhile to optimize.
*
* To be able to simply move entries over, we have to start not at the
* first bucket (i.e olddata[0]), but find the first bucket that's
either
@@ -620,7 +620,7 @@ SH_DELETE(SH_TYPE *tb, SH_KEY_TYPE key)
/*
* Backward shift following elements till either an
empty element
- * or an element at its optimal position is
encounterered.
+ * or an element at its optimal position is encountered.
*
* While that sounds expensive, the average chain
length is short,
* and deletions would otherwise require toombstones.
diff --git a/src/include/replication/syncrep.h
b/src/include/replication/syncrep.h
--- a/src/include/replication/syncrep.h
+++ b/src/include/replication/syncrep.h
@@ -50,7 +50,7 @@ typedef struct SyncRepConfigData
* wait for */
uint8 syncrep_method; /* method to choose sync standbys */
int nmembers; /* number of members in
the following list */
- /* member_names contains nmembers consecutive nul-terminated C strings
*/
+ /* member_names contains nmembers consecutive null-terminated C strings
*/
char member_names[FLEXIBLE_ARRAY_MEMBER];
} SyncRepConfigData;
diff --git a/src/include/storage/s_lock.h b/src/include/storage/s_lock.h
--- a/src/include/storage/s_lock.h
+++ b/src/include/storage/s_lock.h
@@ -842,7 +842,7 @@ typedef LONG slock_t;
#define SPIN_DELAY() spin_delay()
/* If using Visual C++ on Win64, inline assembly is unavailable.
- * Use a _mm_pause instrinsic instead of rep nop.
+ * Use a _mm_pause intrinsic instead of rep nop.
*/
#if defined(_WIN64)
static __forceinline void
diff --git a/src/include/tsearch/dicts/spell.h
b/src/include/tsearch/dicts/spell.h
--- a/src/include/tsearch/dicts/spell.h
+++ b/src/include/tsearch/dicts/spell.h
@@ -147,7 +147,7 @@ typedef struct
} CMPDAffix;
/*
- * Type of encoding affix flags in Hunspel dictionaries
+ * Type of encoding affix flags in Hunspell dictionaries
*/
typedef enum
{
diff --git a/src/include/utils/datetime.h b/src/include/utils/datetime.h
--- a/src/include/utils/datetime.h
+++ b/src/include/utils/datetime.h
@@ -209,7 +209,7 @@ struct tzEntry;
/* keep this struct small; it gets used a lot */
typedef struct
{
- char token[TOKMAXLEN + 1]; /* always NUL-terminated */
+ char token[TOKMAXLEN + 1]; /* always NULL-terminated */
char type; /* see field type codes above */
int32 value; /* meaning depends on type */
} datetkn;
@@ -227,7 +227,7 @@ typedef struct TimeZoneAbbrevTable
typedef struct DynamicZoneAbbrev
{
pg_tz *tz; /* NULL if not yet looked up */
- char zone[FLEXIBLE_ARRAY_MEMBER]; /* NUL-terminated zone
name */
+ char zone[FLEXIBLE_ARRAY_MEMBER]; /* NULL-terminated zone
name */
} DynamicZoneAbbrev;
diff --git a/src/include/utils/jsonapi.h b/src/include/utils/jsonapi.h
--- a/src/include/utils/jsonapi.h
+++ b/src/include/utils/jsonapi.h
@@ -45,7 +45,7 @@ typedef enum
* error reporting routines.
* token_terminator and prev_token_terminator point to the character
* AFTER the end of the token, i.e. where there would be a nul byte
- * if we were using nul-terminated strings.
+ * if we were using null-terminated strings.
*/
typedef struct JsonLexContext
{
@@ -127,7 +127,7 @@ extern JsonLexContext *makeJsonLexContex
/*
* Utility function to check if a string is a valid JSON number.
*
- * str agrument does not need to be nul-terminated.
+ * str argument does not need to be null-terminated.
*/
extern bool IsValidJsonNumber(const char *str, int len);
diff --git a/src/interfaces/ecpg/ecpglib/execute.c
b/src/interfaces/ecpg/ecpglib/execute.c
--- a/src/interfaces/ecpg/ecpglib/execute.c
+++ b/src/interfaces/ecpg/ecpglib/execute.c
@@ -2,7 +2,7 @@
/*
* The aim is to get a simpler interface to the database routines.
- * All the tidieous messing around with tuples is supposed to be hidden
+ * All the tedious messing around with tuples is supposed to be hidden
* by this function.
*/
/* Author: Linus Tolke
diff --git a/src/interfaces/ecpg/pgtypeslib/datetime.c
b/src/interfaces/ecpg/pgtypeslib/datetime.c
--- a/src/interfaces/ecpg/pgtypeslib/datetime.c
+++ b/src/interfaces/ecpg/pgtypeslib/datetime.c
@@ -324,7 +324,7 @@ PGTYPESdate_fmt_asc(date dDate, const ch
*
* function works as follows:
* - first we analyze the parameters
- * - if this is a special case with no delimiters, add delimters
+ * - if this is a special case with no delimiters, add delimiters
* - find the tokens. First we look for numerical values. If we have found
* less than 3 tokens, we check for the months' names and thereafter for
* the abbreviations of the months' names.
diff --git a/src/interfaces/ecpg/pgtypeslib/dt.h
b/src/interfaces/ecpg/pgtypeslib/dt.h
--- a/src/interfaces/ecpg/pgtypeslib/dt.h
+++ b/src/interfaces/ecpg/pgtypeslib/dt.h
@@ -213,7 +213,7 @@ typedef double fsec_t;
/* keep this struct small; it gets used a lot */
typedef struct
{
- char token[TOKMAXLEN + 1]; /* always NUL-terminated */
+ char token[TOKMAXLEN + 1]; /* always NULL-terminated */
char type; /* see field type codes above */
int32 value; /* meaning depends on type */
} datetkn;
diff --git a/src/interfaces/ecpg/pgtypeslib/numeric.c
b/src/interfaces/ecpg/pgtypeslib/numeric.c
--- a/src/interfaces/ecpg/pgtypeslib/numeric.c
+++ b/src/interfaces/ecpg/pgtypeslib/numeric.c
@@ -1368,11 +1368,11 @@ PGTYPESnumeric_cmp(numeric *var1, numeri
{
/* use cmp_abs function to calculate the result */
- /* both are positive: normal comparation with cmp_abs */
+ /* both are positive: normal comparison with cmp_abs */
if (var1->sign == NUMERIC_POS && var2->sign == NUMERIC_POS)
return cmp_abs(var1, var2);
- /* both are negative: return the inverse of the normal comparation */
+ /* both are negative: return the inverse of the normal comparison */
if (var1->sign == NUMERIC_NEG && var2->sign == NUMERIC_NEG)
{
/*
diff --git a/src/interfaces/ecpg/preproc/ecpg.header
b/src/interfaces/ecpg/preproc/ecpg.header
--- a/src/interfaces/ecpg/preproc/ecpg.header
+++ b/src/interfaces/ecpg/preproc/ecpg.header
@@ -207,7 +207,7 @@ create_questionmarks(char *name, bool ar
/* In case we have a struct, we have to print as many "?" as there are
attributes in the struct
* An array is only allowed together with an element argument
- * This is essantially only used for inserts, but using a struct as
input parameter is an error anywhere else
+ * This is essentially only used for inserts, but using a struct as
input parameter is an error anywhere else
* so we don't have to worry here. */
if (p->type->type == ECPGt_struct || (array && p->type->type ==
ECPGt_array && p->type->u.element->type == ECPGt_struct))
diff --git a/src/interfaces/ecpg/preproc/ecpg.trailer
b/src/interfaces/ecpg/preproc/ecpg.trailer
--- a/src/interfaces/ecpg/preproc/ecpg.trailer
+++ b/src/interfaces/ecpg/preproc/ecpg.trailer
@@ -355,7 +355,7 @@ ECPGExecuteImmediateStmt: EXECUTE IMMEDI
$$ = $3;
};
/*
- * variable decalartion outside exec sql declare block
+ * variable declaration outside exec sql declare block
*/
ECPGVarDeclaration: single_vt_declaration;
@@ -707,7 +707,7 @@ struct_union_type_with_symbol: s_struct_
free(forward_name);
forward_name = NULL;
- /* This is essantially a typedef but needs the keyword
struct/union as well.
+ /* This is essentially a typedef but needs the keyword
struct/union as well.
* So we create the typedef for each struct definition
with symbol */
for (ptr = types; ptr != NULL; ptr = ptr->next)
{
@@ -1275,7 +1275,7 @@ descriptor_item: SQL_CARDINALITY { $$
;
/*
- * set/reset the automatic transaction mode, this needs a differnet handling
+ * set/reset the automatic transaction mode, this needs a different handling
* as the other set commands
*/
ECPGSetAutocommit: SET SQL_AUTOCOMMIT '=' on_off { $$ = $4; }
@@ -1287,7 +1287,7 @@ on_off: ON { $$ =
mm_strdup("on"); }
;
/*
- * set the actual connection, this needs a differnet handling as the other
+ * set the actual connection, this needs a different handling as the other
* set commands
*/
ECPGSetConnection: SET CONNECTION TO connection_object { $$ = $4; }
diff --git a/src/interfaces/ecpg/preproc/parse.pl
b/src/interfaces/ecpg/preproc/parse.pl
--- a/src/interfaces/ecpg/preproc/parse.pl
+++ b/src/interfaces/ecpg/preproc/parse.pl
@@ -550,7 +550,7 @@ sub dump_fields
if ($len == 1)
{
- # Straight assignement
+ # Straight assignment
$str = ' $$ = ' . $flds_new[0] . ';';
add_to_buffer('rules', $str);
}
diff --git a/src/interfaces/libpq/fe-auth.c b/src/interfaces/libpq/fe-auth.c
--- a/src/interfaces/libpq/fe-auth.c
+++ b/src/interfaces/libpq/fe-auth.c
@@ -803,7 +803,7 @@ pg_fe_getauthname(PQExpBuffer errorMessa
* be sent in cleartext if it is encrypted on the client side. This is
* good because it ensures the cleartext password won't end up in logs,
* pg_stat displays, etc. We export the function so that clients won't
- * be dependent on low-level details like whether the enceyption is MD5
+ * be dependent on low-level details like whether the encryption is MD5
* or something else.
*
* Arguments are the cleartext password, and the SQL name of the user it
diff --git a/src/interfaces/libpq/fe-connect.c
b/src/interfaces/libpq/fe-connect.c
--- a/src/interfaces/libpq/fe-connect.c
+++ b/src/interfaces/libpq/fe-connect.c
@@ -2136,7 +2136,7 @@ keep_going:
/* We will come back to
} /* loop over
addresses */
/*
- * Ooops, no more addresses. An appropriate
error message is
+ * Oops, no more addresses. An appropriate
error message is
* already set up, so just set the right status.
*/
goto error_return;
diff --git a/src/interfaces/libpq/fe-exec.c b/src/interfaces/libpq/fe-exec.c
--- a/src/interfaces/libpq/fe-exec.c
+++ b/src/interfaces/libpq/fe-exec.c
@@ -2334,7 +2334,7 @@ PQputCopyEnd(PGconn *conn, const char *e
{
if (errormsg)
{
- /* Ooops, no way to do this in 2.0 */
+ /* Oops, no way to do this in 2.0 */
printfPQExpBuffer(&conn->errorMessage,
libpq_gettext("function requires at least protocol version 3.0\n"));
return -1;
@@ -3231,7 +3231,7 @@ PQfreeNotify(PGnotify *notify)
*
* For safety the buffer at "to" must be at least 2*length + 1 bytes long.
* A terminating NUL character is added to the output string, whether the
- * input is NUL-terminated or not.
+ * input is NULL-terminated or not.
*
* Returns the actual length of the output (not counting the terminating NUL).
*/
diff --git a/src/interfaces/libpq/libpq-int.h b/src/interfaces/libpq/libpq-int.h
--- a/src/interfaces/libpq/libpq-int.h
+++ b/src/interfaces/libpq/libpq-int.h
@@ -145,7 +145,7 @@ typedef struct pgMessageField
{
struct pgMessageField *next; /* list link */
char code; /* field code */
- char contents[FLEXIBLE_ARRAY_MEMBER]; /*
value, nul-terminated */
+ char contents[FLEXIBLE_ARRAY_MEMBER]; /*
value, null-terminated */
} PGMessageField;
/* Fields needed for notice handling */
@@ -309,7 +309,7 @@ typedef struct pg_conn_host
char *host; /* host name or address, or
socket path */
pg_conn_host_type type; /* type of host */
char *port; /* port number for this host;
if not NULL,
- * overrrides
the PGConn's pgport */
+ * overrides
the PGConn's pgport */
char *password; /* password for this host, read from the
* password
file. only set if the PGconn's
* pgpass field
is NULL. */
@@ -666,7 +666,7 @@ extern void pq_reset_sigpipe(sigset_t *o
#endif
/*
- * The SSL implementatation provides these functions (fe-secure-openssl.c)
+ * The SSL implementation provides these functions (fe-secure-openssl.c)
*/
extern void pgtls_init_library(bool do_ssl, int do_crypto);
extern int pgtls_init(PGconn *conn);
diff --git a/src/interfaces/libpq/win32.c b/src/interfaces/libpq/win32.c
--- a/src/interfaces/libpq/win32.c
+++ b/src/interfaces/libpq/win32.c
@@ -32,7 +32,7 @@
#include "win32.h"
-/* Declared here to avoid pulling in all includes, which causes name
collissions */
+/* Declared here to avoid pulling in all includes, which causes name
collisions */
#ifdef ENABLE_NLS
extern char *libpq_gettext(const char *msgid) pg_attribute_format_arg(1);
#else
diff --git a/src/pl/plperl/ppport.h b/src/pl/plperl/ppport.h
--- a/src/pl/plperl/ppport.h
+++ b/src/pl/plperl/ppport.h
@@ -79,7 +79,7 @@ to be installed on your system.
If this option is given, a copy of each file will be saved with
the given suffix that contains the suggested changes. This does
not require any external programs. Note that this does not
-automagially add a dot between the original filename and the
+automagically add a dot between the original filename and the
suffix. If you want the dot, you have to include it in the option
argument.
@@ -4364,9 +4364,9 @@ DPPP_(my_vload_module)(U32 flags, SV *na
OP * const modname = newSVOP(OP_CONST, 0, name);
/* 5.005 has a somewhat hacky force_normal that doesn't croak on
- SvREADONLY() if PL_compling is true. Current perls take care in
+ SvREADONLY() if PL_compiling is true. Current perls take care in
ck_require() to correctly turn off SvREADONLY before calling
- force_normal_flags(). This seems a better fix than fudging PL_compling
+ force_normal_flags(). This seems a better fix than fudging PL_compiling
*/
SvREADONLY_off(((SVOP*)modname)->op_sv);
modname->op_private |= OPpCONST_BARE;
@@ -6205,10 +6205,10 @@ DPPP_(my_grok_number)(pTHX_ const char *
/* UVs are at least 32 bits, so the first 9 decimal digits cannot
overflow. */
UV value = *s - '0';
- /* This construction seems to be more optimiser friendly.
+ /* This construction seems to be more optimizer friendly.
(without it gcc does the isDIGIT test and the *s - '0' separately)
With it gcc on arm is managing 6 instructions (6 cycles) per digit.
- In theory the optimiser could deduce how far to unroll the loop
+ In theory the optimizer could deduce how far to unroll the loop
before checking for overflow. */
if (++s < send) {
int digit = *s - '0';
@@ -6606,7 +6606,7 @@ DPPP_(my_grok_oct)(pTHX_ const char *sta
bool overflowed = FALSE;
for (; len-- && *s; s++) {
- /* gcc 2.95 optimiser not smart enough to figure that this subtraction
+ /* gcc 2.95 optimizer not smart enough to figure that this subtraction
out front allows slicker code. */
int digit = *s - '0';
if (digit >= 0 && digit <= 7) {
diff --git a/src/pl/plpgsql/src/pl_exec.c b/src/pl/plpgsql/src/pl_exec.c
--- a/src/pl/plpgsql/src/pl_exec.c
+++ b/src/pl/plpgsql/src/pl_exec.c
@@ -5541,7 +5541,7 @@ exec_eval_simple_expr(PLpgSQL_execstate
exec_check_rw_parameter(expr, expr->rwparam);
if (expr->expr_simple_expr == NULL)
{
- /* Ooops, release refcount and fail */
+ /* Oops, release refcount and fail */
ReleaseCachedPlan(cplan, true);
return false;
}
diff --git a/src/pl/plpgsql/src/plpgsql.h b/src/pl/plpgsql/src/plpgsql.h
--- a/src/pl/plpgsql/src/plpgsql.h
+++ b/src/pl/plpgsql/src/plpgsql.h
@@ -357,7 +357,7 @@ typedef struct PLpgSQL_nsitem
*/
int itemno;
struct PLpgSQL_nsitem *prev;
- char name[FLEXIBLE_ARRAY_MEMBER]; /* nul-terminated
string */
+ char name[FLEXIBLE_ARRAY_MEMBER]; /* null-terminated
string */
} PLpgSQL_nsitem;
/*
diff --git a/src/pl/plpython/plpy_elog.c b/src/pl/plpython/plpy_elog.c
--- a/src/pl/plpython/plpy_elog.c
+++ b/src/pl/plpython/plpy_elog.c
@@ -303,7 +303,7 @@ PLy_traceback(PyObject *e, PyObject *v,
long plain_lineno;
/*
- * The second frame points at the internal function,
but to mimick
+ * The second frame points at the internal function,
but to mimic
* Python error reporting we want to say <module>.
*/
if (*tb_depth == 1)
diff --git a/src/pl/plpython/plpy_plpymodule.c
b/src/pl/plpython/plpy_plpymodule.c
--- a/src/pl/plpython/plpy_plpymodule.c
+++ b/src/pl/plpython/plpy_plpymodule.c
@@ -463,7 +463,7 @@ PLy_output(volatile int level, PyObject
if (strcmp(keyword, "message") == 0)
{
- /* the message should not be overwriten */
+ /* the message should not be overwritten */
if (PyTuple_Size(args) != 0)
{
PLy_exception_set(PyExc_TypeError,
"Argument 'message' given by name and position");
diff --git a/src/pl/plpython/plpy_procedure.c b/src/pl/plpython/plpy_procedure.c
--- a/src/pl/plpython/plpy_procedure.c
+++ b/src/pl/plpython/plpy_procedure.c
@@ -122,7 +122,7 @@ PLy_procedure_get(Oid fn_oid, Oid fn_rel
}
PG_CATCH();
{
- /* Do not leave an uninitialised entry in the cache */
+ /* Do not leave an uninitialized entry in the cache */
if (use_cache)
hash_search(PLy_procedure_cache, &key, HASH_REMOVE,
NULL);
PG_RE_THROW();
diff --git a/src/pl/plpython/plpy_typeio.c b/src/pl/plpython/plpy_typeio.c
--- a/src/pl/plpython/plpy_typeio.c
+++ b/src/pl/plpython/plpy_typeio.c
@@ -833,7 +833,7 @@ PLyObject_ToComposite(PLyObToDatum *arg,
/*
* This will set up the dummy PLyTypeInfo's output conversion routines,
- * since we left is_rowtype as 2. A future optimisation could be caching
+ * since we left is_rowtype as 2. A future optimization could be caching
* that info instead of looking it up every time a tuple is returned
from
* the function.
*/
diff --git a/src/pl/plpython/plpy_typeio.h b/src/pl/plpython/plpy_typeio.h
--- a/src/pl/plpython/plpy_typeio.h
+++ b/src/pl/plpython/plpy_typeio.h
@@ -43,7 +43,7 @@ typedef union PLyTypeInput
} PLyTypeInput;
/*
- * Conversion from Python object to a Postgresql Datum.
+ * Conversion from Python object to a PostgreSQL Datum.
*
* The 'inarray' argument to the conversion function is true, if the
* converted value was in an array (Python list). It is used to give a
@@ -78,7 +78,7 @@ typedef union PLyTypeOutput
PLyObToTuple r;
} PLyTypeOutput;
-/* all we need to move Postgresql data to Python objects,
+/* all we need to move PostgreSQL data to Python objects,
* and vice versa
*/
typedef struct PLyTypeInfo
diff --git a/src/test/isolation/specs/receipt-report.spec
b/src/test/isolation/specs/receipt-report.spec
--- a/src/test/isolation/specs/receipt-report.spec
+++ b/src/test/isolation/specs/receipt-report.spec
@@ -7,7 +7,7 @@
# be changed and a report of the closed day's receipts subsequently
# run which will miss a receipt from the date which has been closed.
#
-# There are only six permuations which must cause a serialization failure.
+# There are only six permutations which must cause a serialization failure.
# Failure cases are where s1 overlaps both s2 and s3, but s2 commits before
# s3 executes its first SELECT.
#
diff --git a/src/test/isolation/specs/two-ids.spec
b/src/test/isolation/specs/two-ids.spec
--- a/src/test/isolation/specs/two-ids.spec
+++ b/src/test/isolation/specs/two-ids.spec
@@ -2,7 +2,7 @@
#
# Small, simple test showing read-only anomalies.
#
-# There are only four permuations which must cause a serialization failure.
+# There are only four permutations which must cause a serialization failure.
# Required failure cases are where s2 overlaps both s1 and s3, but s1
# commits before s3 executes its first SELECT.
#
diff --git a/src/test/regress/sql/alter_table.sql
b/src/test/regress/sql/alter_table.sql
--- a/src/test/regress/sql/alter_table.sql
+++ b/src/test/regress/sql/alter_table.sql
@@ -255,7 +255,7 @@ INSERT INTO tmp3 values (5,50);
-- Try (and fail) to add constraint due to invalid source columns
ALTER TABLE tmp3 add constraint tmpconstr foreign key(c) references tmp2 match
full;
--- Try (and fail) to add constraint due to invalide destination columns
explicitly given
+-- Try (and fail) to add constraint due to invalid destination columns
explicitly given
ALTER TABLE tmp3 add constraint tmpconstr foreign key(a) references tmp2(b)
match full;
-- Try (and fail) to add constraint due to invalid data
@@ -1829,7 +1829,7 @@ CREATE UNLOGGED TABLE unlogged3(f1 SERIA
ALTER TABLE unlogged3 SET LOGGED; -- skip self-referencing foreign key
ALTER TABLE unlogged2 SET LOGGED; -- fails because a foreign key to an
unlogged table exists
ALTER TABLE unlogged1 SET LOGGED;
--- check relpersistence of an unlogged table after changing to permament
+-- check relpersistence of an unlogged table after changing to permanent
SELECT relname, relkind, relpersistence FROM pg_class WHERE relname ~
'^unlogged1'
UNION ALL
SELECT 'toast table', t.relkind, t.relpersistence FROM pg_class r JOIN
pg_class t ON t.oid = r.reltoastrelid WHERE r.relname ~ '^unlogged1'
@@ -1917,7 +1917,7 @@ ALTER TABLE partitioned ALTER COLUMN b T
-- cannot drop NOT NULL on columns in the range partition key
ALTER TABLE partitioned ALTER COLUMN a DROP NOT NULL;
--- partitioned table cannot partiticipate in regular inheritance
+-- partitioned table cannot participate in regular inheritance
CREATE TABLE foo (
a int,
b int
diff --git a/src/test/regress/sql/create_table.sql
b/src/test/regress/sql/create_table.sql
--- a/src/test/regress/sql/create_table.sql
+++ b/src/test/regress/sql/create_table.sql
@@ -418,7 +418,7 @@ SELECT attname, attnotnull FROM pg_attri
-- prevent a function referenced in partition key from being dropped
DROP FUNCTION plusone(int);
--- partitioned table cannot partiticipate in regular inheritance
+-- partitioned table cannot participate in regular inheritance
CREATE TABLE partitioned2 (
a int
) PARTITION BY LIST ((a+1));
diff --git a/src/test/regress/sql/errors.sql b/src/test/regress/sql/errors.sql
--- a/src/test/regress/sql/errors.sql
+++ b/src/test/regress/sql/errors.sql
@@ -2,7 +2,7 @@
-- ERRORS
--
--- bad in postquel, but ok in postsql
+-- bad in postquel, but ok in PostgreSQL
select 1;
diff --git a/src/test/regress/sql/groupingsets.sql
b/src/test/regress/sql/groupingsets.sql
--- a/src/test/regress/sql/groupingsets.sql
+++ b/src/test/regress/sql/groupingsets.sql
@@ -140,7 +140,7 @@ select *
from (values (1),(2)) v(x),
lateral (select a, b, sum(v.x) from gstest_data(v.x) group by rollup
(a,b)) s;
--- min max optimisation should still work with GROUP BY ()
+-- min max optimization should still work with GROUP BY ()
explain (costs off)
select min(unique1) from tenk1 GROUP BY ();
diff --git a/src/test/regress/sql/indirect_toast.sql
b/src/test/regress/sql/indirect_toast.sql
--- a/src/test/regress/sql/indirect_toast.sql
+++ b/src/test/regress/sql/indirect_toast.sql
@@ -11,7 +11,7 @@ SELECT descr, substring(make_tuple_indir
-- modification without changing varlenas
UPDATE toasttest SET cnt = cnt +1 RETURNING substring(toasttest::text, 1, 200);
--- modification without modifying asigned value
+-- modification without modifying assigned value
UPDATE toasttest SET cnt = cnt +1, f1 = f1 RETURNING
substring(toasttest::text, 1, 200);
-- modification modifying, but effectively not changing
@@ -20,7 +20,7 @@ UPDATE toasttest SET cnt = cnt +1, f1 =
UPDATE toasttest SET cnt = cnt +1, f1 = '-'||f1||'-' RETURNING
substring(toasttest::text, 1, 200);
SELECT substring(toasttest::text, 1, 200) FROM toasttest;
--- check we didn't screw with main/toast tuple visiblity
+-- check we didn't screw with main/toast tuple visibility
VACUUM FREEZE toasttest;
SELECT substring(toasttest::text, 1, 200) FROM toasttest;
@@ -42,7 +42,7 @@ CREATE TRIGGER toasttest_update_indirect
-- modification without changing varlenas
UPDATE toasttest SET cnt = cnt +1 RETURNING substring(toasttest::text, 1, 200);
--- modification without modifying asigned value
+-- modification without modifying assigned value
UPDATE toasttest SET cnt = cnt +1, f1 = f1 RETURNING
substring(toasttest::text, 1, 200);
-- modification modifying, but effectively not changing
@@ -53,7 +53,7 @@ UPDATE toasttest SET cnt = cnt +1, f1 =
INSERT INTO toasttest(descr, f1, f2) VALUES('one-toasted,one-null, via
indirect', repeat('1234567890',30000), NULL);
SELECT substring(toasttest::text, 1, 200) FROM toasttest;
--- check we didn't screw with main/toast tuple visiblity
+-- check we didn't screw with main/toast tuple visibility
VACUUM FREEZE toasttest;
SELECT substring(toasttest::text, 1, 200) FROM toasttest;
diff --git a/src/test/regress/sql/init_privs.sql
b/src/test/regress/sql/init_privs.sql
--- a/src/test/regress/sql/init_privs.sql
+++ b/src/test/regress/sql/init_privs.sql
@@ -1,4 +1,4 @@
--- Test iniital privileges
+-- Test initial privileges
-- There should always be some initial privileges, set up by initdb
SELECT count(*) > 0 FROM pg_init_privs;
diff --git a/src/test/regress/sql/insert_conflict.sql
b/src/test/regress/sql/insert_conflict.sql
--- a/src/test/regress/sql/insert_conflict.sql
+++ b/src/test/regress/sql/insert_conflict.sql
@@ -138,7 +138,7 @@ insert into insertconflicttest values (1
drop index comp_key_index;
--
--- Partial index tests, no inference predicate specificied
+-- Partial index tests, no inference predicate specified
--
create unique index part_comp_key_index on insertconflicttest(key, fruit)
where key < 5;
create unique index expr_part_comp_key_index on insertconflicttest(key,
lower(fruit)) where key < 5;
diff --git a/src/test/regress/sql/join.sql b/src/test/regress/sql/join.sql
--- a/src/test/regress/sql/join.sql
+++ b/src/test/regress/sql/join.sql
@@ -1456,7 +1456,7 @@ select * from
--
select t1.uunique1 from
- tenk1 t1 join tenk2 t2 on t1.two = t2.two; -- error, prefer "t1" suggestipn
+ tenk1 t1 join tenk2 t2 on t1.two = t2.two; -- error, prefer "t1" suggestion
select t2.uunique1 from
tenk1 t1 join tenk2 t2 on t1.two = t2.two; -- error, prefer "t2" suggestion
select uunique1 from
diff --git a/src/test/regress/sql/matview.sql b/src/test/regress/sql/matview.sql
--- a/src/test/regress/sql/matview.sql
+++ b/src/test/regress/sql/matview.sql
@@ -92,7 +92,7 @@ SELECT * FROM mvtest_tvvm;
-- test diemv when the mv does not exist
DROP MATERIALIZED VIEW IF EXISTS no_such_mv;
--- make sure invalid comination of options is prohibited
+-- make sure invalid combination of options is prohibited
REFRESH MATERIALIZED VIEW CONCURRENTLY mvtest_tvmm WITH NO DATA;
-- no tuple locks on materialized views
diff --git a/src/test/regress/sql/plpgsql.sql b/src/test/regress/sql/plpgsql.sql
--- a/src/test/regress/sql/plpgsql.sql
+++ b/src/test/regress/sql/plpgsql.sql
@@ -1350,7 +1350,7 @@ select * from WSlot order by slotname;
--
-- Install the central phone system and create the phone numbers.
--- They are weired on insert to the patchfields. Again the
+-- They are wired on insert to the patchfields. Again the
-- triggers automatically tell the PSlots to update their
-- backlink field.
--
diff --git a/src/test/regress/sql/replica_identity.sql
b/src/test/regress/sql/replica_identity.sql
--- a/src/test/regress/sql/replica_identity.sql
+++ b/src/test/regress/sql/replica_identity.sql
@@ -56,7 +56,7 @@ SELECT relreplident FROM pg_class WHERE
-- succeed, oid unique index
ALTER TABLE test_replica_identity REPLICA IDENTITY USING INDEX
test_replica_identity_oid_idx;
--- succeed, nondeferrable unique constraint over nonullable cols
+-- succeed, nondeferrable unique constraint over nonnullable cols
ALTER TABLE test_replica_identity REPLICA IDENTITY USING INDEX
test_replica_identity_unique_nondefer;
-- succeed unique index over nonnullable cols
diff --git a/src/test/regress/sql/rolenames.sql
b/src/test/regress/sql/rolenames.sql
--- a/src/test/regress/sql/rolenames.sql
+++ b/src/test/regress/sql/rolenames.sql
@@ -176,7 +176,7 @@ ALTER USER PUBLIC SET application_name t
ALTER USER NONE SET application_name to 'BOMB'; -- error
ALTER USER nonexistent SET application_name to 'BOMB'; -- error
--- CREAETE SCHEMA
+-- CREATE SCHEMA
set client_min_messages to error;
CREATE SCHEMA newschema1 AUTHORIZATION CURRENT_USER;
CREATE SCHEMA newschema2 AUTHORIZATION "current_user";
diff --git a/src/test/regress/sql/rules.sql b/src/test/regress/sql/rules.sql
--- a/src/test/regress/sql/rules.sql
+++ b/src/test/regress/sql/rules.sql
@@ -522,7 +522,7 @@ CREATE TABLE shoe_data (
shoename char(10), -- primary key
sh_avail integer, -- available # of pairs
slcolor char(10), -- preferred shoelace color
- slminlen float, -- miminum shoelace length
+ slminlen float, -- minimum shoelace length
slmaxlen float, -- maximum shoelace length
slunit char(8) -- length unit
);
diff --git a/src/test/regress/sql/tsdicts.sql b/src/test/regress/sql/tsdicts.sql
--- a/src/test/regress/sql/tsdicts.sql
+++ b/src/test/regress/sql/tsdicts.sql
@@ -96,7 +96,7 @@ SELECT ts_lexize('hunspell_num', 'footba
SELECT ts_lexize('hunspell_num', 'ballyklubber');
SELECT ts_lexize('hunspell_num', 'footballyklubber');
--- Synonim dictionary
+-- Synonym dictionary
CREATE TEXT SEARCH DICTIONARY synonym (
Template=synonym,
Synonyms=synonym_sample
diff --git a/src/test/ssl/ServerSetup.pm b/src/test/ssl/ServerSetup.pm
--- a/src/test/ssl/ServerSetup.pm
+++ b/src/test/ssl/ServerSetup.pm
@@ -7,7 +7,7 @@
# - ssl/root+client_ca.crt as the CA root for validating client certs.
# - reject non-SSL connections
# - a database called trustdb that lets anyone in
-# - another database called certdb that uses certificate authentiction, ie.
+# - another database called certdb that uses certificate authentication, ie.
# the client must present a valid certificate signed by the client CA
# - two users, called ssltestuser and anotheruser.
#
--
Sent via pgsql-hackers mailing list ([email protected])
To make changes to your subscription:
http://www.postgresql.org/mailpref/pgsql-hackers