Remove whitespace from end of lines

This commit is contained in:
Peter Eisentraut 2013-05-30 21:05:07 -04:00
parent d7eb6f46de
commit 8b5a3998a1
12 changed files with 33 additions and 33 deletions

View File

@ -632,7 +632,7 @@ HINT: Stop the postmaster and use a standalone backend to VACUUM in "mydb".
<varname>autovacuum_max_workers</> databases to be processed,
the next database will be processed as soon as the first worker finishes.
Each worker process will check each table within its database and
execute <command>VACUUM</> and/or <command>ANALYZE</> as needed.
execute <command>VACUUM</> and/or <command>ANALYZE</> as needed.
<varname>log_autovacuum_min_duration</varname> can be used to monitor
autovacuum activity.
</para>

View File

@ -336,7 +336,7 @@ CREATE [ [ GLOBAL | LOCAL ] { TEMPORARY | TEMP } | UNLOGGED ] TABLE [ IF NOT EXI
<para>
Not-null constraints are always copied to the new table.
<literal>CHECK</literal> constraints will be copied only if
<literal>INCLUDING CONSTRAINTS</literal> is specified.
<literal>INCLUDING CONSTRAINTS</literal> is specified.
Indexes, <literal>PRIMARY KEY</>, and <literal>UNIQUE</> constraints
on the original table will be created on the new table only if the
<literal>INCLUDING INDEXES</literal> clause is specified.

View File

@ -33,7 +33,7 @@ NOTIFY <replaceable class="PARAMETER">channel</replaceable> [ , <replaceable cla
with an optional <quote>payload</> string to each client application that
has previously executed
<command>LISTEN <replaceable class="parameter">channel</></command>
for the specified channel name in the current database.
for the specified channel name in the current database.
Notifications are visible to all users.
</para>

View File

@ -324,7 +324,7 @@ PostgreSQL documentation
For a consistent backup, the database server needs to support synchronized snapshots,
a feature that was introduced in <productname>PostgreSQL</productname> 9.2. With this
feature, database clients can ensure they see the same data set even though they use
different connections. <command>pg_dump -j</command> uses multiple database
different connections. <command>pg_dump -j</command> uses multiple database
connections; it connects to the database once with the master process and
once again for each worker job. Without the synchronized snapshot feature, the
different worker jobs wouldn't be guaranteed to see the same data in each connection,

View File

@ -434,7 +434,7 @@ PL_TESTDB = pl_regression
CONTRIB_TESTDB = contrib_regression
ifneq ($(MODULE_big),)
CONTRIB_TESTDB_MODULE = contrib_regression_$(MODULE_big)
else
else
ifneq ($(MODULES),)
CONTRIB_TESTDB_MODULE = contrib_regression_$(MODULES)
else

View File

@ -788,10 +788,10 @@ CREATE OR REPLACE FUNCTION
pg_start_backup(label text, fast boolean DEFAULT false)
RETURNS text STRICT VOLATILE LANGUAGE internal AS 'pg_start_backup';
CREATE OR REPLACE FUNCTION
CREATE OR REPLACE FUNCTION
json_populate_record(base anyelement, from_json json, use_json_as_text boolean DEFAULT false)
RETURNS anyelement LANGUAGE internal STABLE AS 'json_populate_record';
CREATE OR REPLACE FUNCTION
CREATE OR REPLACE FUNCTION
json_populate_recordset(base anyelement, from_json json, use_json_as_text boolean DEFAULT false)
RETURNS SETOF anyelement LANGUAGE internal STABLE ROWS 100 AS 'json_populate_recordset';

View File

@ -466,15 +466,15 @@ INSERT INTO test_json VALUES
('scalar','"a scalar"'),
('array','["zero", "one","two",null,"four","five"]'),
('object','{"field1":"val1","field2":"val2","field3":null}');
SELECT test_json -> 'x'
SELECT test_json -> 'x'
FROM test_json
WHERE json_type = 'scalar';
ERROR: cannot extract element from a scalar
SELECT test_json -> 'x'
SELECT test_json -> 'x'
FROM test_json
WHERE json_type = 'array';
ERROR: cannot extract field from a non-object
SELECT test_json -> 'x'
SELECT test_json -> 'x'
FROM test_json
WHERE json_type = 'object';
?column?
@ -490,7 +490,7 @@ WHERE json_type = 'object';
"val2"
(1 row)
SELECT test_json->>'field2'
SELECT test_json->>'field2'
FROM test_json
WHERE json_type = 'object';
?column?
@ -498,11 +498,11 @@ WHERE json_type = 'object';
val2
(1 row)
SELECT test_json -> 2
SELECT test_json -> 2
FROM test_json
WHERE json_type = 'scalar';
ERROR: cannot extract element from a scalar
SELECT test_json -> 2
SELECT test_json -> 2
FROM test_json
WHERE json_type = 'array';
?column?

View File

@ -139,15 +139,15 @@ INSERT INTO test_json VALUES
('array','["zero", "one","two",null,"four","five"]'),
('object','{"field1":"val1","field2":"val2","field3":null}');
SELECT test_json -> 'x'
SELECT test_json -> 'x'
FROM test_json
WHERE json_type = 'scalar';
SELECT test_json -> 'x'
SELECT test_json -> 'x'
FROM test_json
WHERE json_type = 'array';
SELECT test_json -> 'x'
SELECT test_json -> 'x'
FROM test_json
WHERE json_type = 'object';
@ -155,15 +155,15 @@ SELECT test_json->'field2'
FROM test_json
WHERE json_type = 'object';
SELECT test_json->>'field2'
SELECT test_json->>'field2'
FROM test_json
WHERE json_type = 'object';
SELECT test_json -> 2
SELECT test_json -> 2
FROM test_json
WHERE json_type = 'scalar';
SELECT test_json -> 2
SELECT test_json -> 2
FROM test_json
WHERE json_type = 'array';

View File

@ -523,8 +523,8 @@ sub mkvcbuild
my $mf = Project::read_file(
'src\backend\utils\mb\conversion_procs\\' . $sub . '\Makefile');
my $p = $solution->AddProject($sub, 'dll', 'conversion procs');
$p->AddFile('src\backend\utils\mb\conversion_procs\\'
. $sub . '\\'
$p->AddFile('src\backend\utils\mb\conversion_procs\\'
. $sub . '\\'
. $sub
. '.c');
if ($mf =~ m{^SRCS\s*\+=\s*(.*)$}m)

View File

@ -1,12 +1,12 @@
--add-whitespace
--backup-and-modify-in-place
--add-whitespace
--backup-and-modify-in-place
--delete-old-whitespace
--entab-leading-whitespace=4
--keep-old-blank-lines=2
--entab-leading-whitespace=4
--keep-old-blank-lines=2
--maximum-line-length=78
--nospace-for-semicolon
--nospace-for-semicolon
--opening-brace-on-new-line
--output-line-ending=unix
--paren-tightness=2
--vertical-tightness=2
--paren-tightness=2
--vertical-tightness=2
--vertical-tightness-closing=2

View File

@ -310,7 +310,7 @@ sub post_indent
$source =~ s!
(\n$ident[^(\n]*)\n # e.g. static void
(
$ident\(\n? # func_name(
$ident\(\n? # func_name(
(.*,([ \t]*$comment)?\n)* # args b4 final ln
.*\);([ \t]*$comment)?$ # final line
)

View File

@ -6,9 +6,9 @@ or the environment.
In its simplest form, if all the required objects are installed, simply run
it without any parameters at the top of the source tree you want to process.
pgindent
pgindent
If you don't have all the requirements installed, pgindent will fetch and build
If you don't have all the requirements installed, pgindent will fetch and build
them for you, if you're in a PostgreSQL source tree:
@ -23,7 +23,7 @@ command line option --indent:
Similarly, the entab program can be specified using the PGENTAB environment
variable, or using the --entab command line option.
pgindent also needs a file containing a list of typedefs. This can be
pgindent also needs a file containing a list of typedefs. This can be
specified using the PGTYPEDEFS environment variable, or via the command line
--typedefs option. If neither is used, it will look for it within the
current source tree, or in /usr/local/etc/typedefs.list.
@ -40,6 +40,6 @@ src/tools/pgindent/exclude_file_patterns.
Any non-option arguments are taken as the names of files to be indented. In this
case only these files will be changed, and nothing else will be touched. If the
first non-option argument is not a .c or .h file, it is treated as the name
of a typedefs file for legacy reasons, but this use is deprecated - use the
of a typedefs file for legacy reasons, but this use is deprecated - use the
--typedefs option instead.