Fix misc typos in comments.

Spotted mostly by Fabien Coelho.

Discussion: https://www.postgresql.org/message-id/alpine.DEB.2.21.1901230947050.16643@lancre
This commit is contained in:
Heikki Linnakangas 2019-01-23 13:39:00 +02:00
parent 1699e6dd1f
commit 95931133a9
15 changed files with 16 additions and 16 deletions

View File

@ -132,7 +132,7 @@ pgp_parse_pkt_hdr(PullFilter *src, uint8 *tag, int *len_p, int allow_ctx)
int res;
uint8 *p;
/* EOF is normal here, thus we dont use GETBYTE */
/* EOF is normal here, thus we don't use GETBYTE */
res = pullf_read(src, 1, &p);
if (res < 0)
return res;

View File

@ -2771,7 +2771,7 @@ estimate_path_cost_size(PlannerInfo *root,
run_cost += cpu_tuple_cost * numGroups;
run_cost += ptarget->cost.per_tuple * numGroups;
/* Accout for the eval cost of HAVING quals, if any */
/* Account for the eval cost of HAVING quals, if any */
if (root->parse->havingQual)
{
QualCost remote_cost;

View File

@ -442,7 +442,7 @@ restartScanEntry:
/*
* Lock the entry leaf page. This is more coarse-grained than
* necessary, because it will conflict with any insertions that
* land on the same leaf page, not only the exacty key we searched
* land on the same leaf page, not only the exact key we searched
* for. But locking an individual tuple would require updating
* that lock whenever it moves because of insertions or vacuums,
* which seems too complicated.

View File

@ -1873,7 +1873,7 @@ CheckOpSlotCompatibility(ExprEvalStep *op, TupleTableSlot *slot)
/*
* Should probably fixed at some point, but for now it's easier to allow
* buffer and heap tuples to be used interchangably.
* buffer and heap tuples to be used interchangeably.
*/
if (slot->tts_ops == &TTSOpsBufferHeapTuple &&
op->d.fetch.kind == &TTSOpsHeapTuple)

View File

@ -1049,7 +1049,7 @@ ExecParallelRetrieveJitInstrumentation(PlanState *planstate,
MemoryContextAllocZero(planstate->state->es_query_cxt, sizeof(JitInstrumentation));
combined = planstate->state->es_jit_worker_instr;
/* Accummulate all the workers' instrumentations. */
/* Accumulate all the workers' instrumentations. */
for (n = 0; n < shared_jit->num_workers; ++n)
InstrJitAgg(combined, &shared_jit->jit_instr[n]);

View File

@ -1476,7 +1476,7 @@ ExecStoreAllNullTuple(TupleTableSlot *slot)
* possible.
*
* If materialize is true, the contents of the slots will be made independent
* from the underlying storage (i.e. all buffer pins are release, memory is
* from the underlying storage (i.e. all buffer pins are released, memory is
* allocated in the slot's context).
*
* If shouldFree is not-NULL it'll be set to true if the returned tuple has

View File

@ -2677,7 +2677,7 @@ alter_identity_column_option:
;
PartitionBoundSpec:
/* a HASH partition*/
/* a HASH partition */
FOR VALUES WITH '(' hash_partbound ')'
{
ListCell *lc;

View File

@ -2243,7 +2243,7 @@ check_log_duration(char *msec_str, bool was_logged)
/*
* Do not log if log_statement_sample_rate = 0. Log a sample if
* log_statement_sample_rate <= 1 and avoid unecessary random() call
* log_statement_sample_rate <= 1 and avoid unnecessary random() call
* if log_statement_sample_rate = 1. But don't compute any of this
* unless needed.
*/

View File

@ -8,7 +8,7 @@
* When a tuple is updated or deleted, our standard visibility rules
* consider that it is *still valid* so long as we are in the same command,
* ie, until the next CommandCounterIncrement() or transaction commit.
* (See acces/heap/heapam_visibility.c, and note that system catalogs are
* (See access/heap/heapam_visibility.c, and note that system catalogs are
* generally scanned under the most current snapshot available, rather than
* the transaction snapshot.) At the command boundary, the old tuple stops
* being valid and the new version, if any, becomes valid. Therefore,

View File

@ -151,7 +151,7 @@ typedef struct RecordCacheEntry
/*
* To deal with non-anonymous record types that are exchanged by backends
* involved in a parallel query, we also need a shared verion of the above.
* involved in a parallel query, we also need a shared version of the above.
*/
struct SharedRecordTypmodRegistry
{

View File

@ -407,12 +407,12 @@ typedef struct
} CState;
/*
* Cache cell for zipfian_random call
* Cache cell for random_zipfian call
*/
typedef struct
{
/* cell keys */
double s; /* s - parameter of zipfan_random function */
double s; /* s - parameter of random_zipfian function */
int64 n; /* number of elements in range (max - min + 1) */
double harmonicn; /* generalizedHarmonicNumber(n, s) */

View File

@ -600,7 +600,7 @@ exec_command_conninfo(PsqlScanState scan_state, bool active_branch)
/*
* If the host is an absolute path, the connection is via socket
* unless overriden by hostaddr
* unless overridden by hostaddr
*/
if (is_absolute_path(host))
{

View File

@ -1136,7 +1136,7 @@ typedef struct ModifyTableState
* nplans how many plans are in the array
* whichplan which plan is being executed (0 .. n-1), or a
* special negative value. See nodeAppend.c.
* pruningstate details required to allow partitions to be
* prune_state details required to allow partitions to be
* eliminated from the scan, or NULL if not possible.
* valid_subplans for runtime pruning, valid appendplans indexes to
* scan.

View File

@ -523,7 +523,7 @@ extern pqsigfunc pqsignal_no_restart(int signo, pqsigfunc func);
/* port/quotes.c */
extern char *escape_single_quotes_ascii(const char *src);
/* port/wait_error.c */
/* common/wait_error.c */
extern char *wait_result_to_str(int exit_status);
extern bool wait_result_is_signal(int exit_status, int signum);
extern bool wait_result_is_any_signal(int exit_status, bool include_command_not_found);

View File

@ -244,7 +244,7 @@ PLy_traceback(PyObject *e, PyObject *v, PyObject *tb,
*tb_depth = 0;
initStringInfo(&tbstr);
/* Mimick Python traceback reporting as close as possible. */
/* Mimic Python traceback reporting as close as possible. */
appendStringInfoString(&tbstr, "Traceback (most recent call last):");
while (tb != NULL && tb != Py_None)
{