4504 lines
120 KiB
C
4504 lines
120 KiB
C
/*-------------------------------------------------------------------------
|
|
*
|
|
* pg_dump.c
|
|
* pg_dump is an utility for dumping out a postgres database
|
|
* into a script file.
|
|
*
|
|
* pg_dump will read the system catalogs in a database and
|
|
* dump out a script that reproduces
|
|
* the schema of the database in terms of
|
|
* user-defined types
|
|
* user-defined functions
|
|
* tables
|
|
* indices
|
|
* aggregates
|
|
* operators
|
|
* ACL - grant/revoke
|
|
*
|
|
* the output script is SQL that is understood by PostgreSQL
|
|
*
|
|
* Portions Copyright (c) 1996-2001, PostgreSQL Global Development Group
|
|
* Portions Copyright (c) 1994, Regents of the University of California
|
|
*
|
|
*
|
|
* IDENTIFICATION
|
|
* $Header: /cvsroot/pgsql/src/bin/pg_dump/pg_dump.c,v 1.191 2001/02/10 02:31:27 tgl Exp $
|
|
*
|
|
* Modifications - 6/10/96 - dave@bensoft.com - version 1.13.dhb
|
|
*
|
|
* Applied 'insert string' patch from "Marc G. Fournier" <scrappy@ki.net>
|
|
* Added '-t table' option
|
|
* Added '-a' option
|
|
* Added '-da' option
|
|
*
|
|
* Modifications - 6/12/96 - dave@bensoft.com - version 1.13.dhb.2
|
|
*
|
|
* - Fixed dumpTable output to output lengths for char and varchar types!
|
|
* - Added single. quote to twin single quote expansion for 'insert' string
|
|
* mode.
|
|
*
|
|
* Modifications - 7/26/96 - asussman@vidya.com
|
|
*
|
|
* - Fixed ouput lengths for char and varchar type where the length is variable (-1)
|
|
*
|
|
* Modifications - 6/1/97 - igor@sba.miami.edu
|
|
* - Added functions to free allocated memory used for retrieving
|
|
* indices,tables,inheritance,types,functions and aggregates.
|
|
* No more leaks reported by Purify.
|
|
*
|
|
*
|
|
* Modifications - 1/26/98 - pjlobo@euitt.upm.es
|
|
* - Added support for password authentication
|
|
*
|
|
* Modifications - 28-Jun-2000 - Philip Warner pjw@rhyme.com.au
|
|
* - Used custom IO routines to allow for more
|
|
* output formats and simple rearrangement of order.
|
|
* - Discouraged operations more appropriate to the 'restore'
|
|
* operation. (eg. -c "clear schema" - now always dumps
|
|
* commands, but pg_restore can be told not to output them).
|
|
* - Added RI warnings to the 'as insert strings' output mode
|
|
* - Added a small number of comments
|
|
* - Added a -Z option for compression level on compressed formats
|
|
* - Restored '-f' in usage output
|
|
*
|
|
*
|
|
* Modifications - 17-Jul-2000 - Philip Warner pjw@rhyme.com.au
|
|
* - Support for BLOB output.
|
|
* - Sort archive by OID, put some items at end (out of OID order)
|
|
*
|
|
* Modifications - 28-Jul-2000 - pjw@rhyme.com.au (1.45)
|
|
*
|
|
* Added --create, --no-owner, --superuser, --no-reconnect (pg_dump & pg_restore)
|
|
* Added code to dump 'Create Schema' statement (pg_dump)
|
|
* Don't bother to disable/enable triggers if we don't have a superuser (pg_restore)
|
|
* Cleaned up code for reconnecting to database.
|
|
* Force a reconnect as superuser before enabling/disabling triggers.
|
|
*
|
|
* Modifications - 31-Jul-2000 - pjw@rhyme.com.au (1.46, 1.47)
|
|
* Added & Removed --throttle (pg_dump)
|
|
* Fixed minor bug in language dumping code: expbuffres were not being reset.
|
|
* Fixed version number initialization in _allocAH (pg_backup_archiver.c)
|
|
*
|
|
* Modifications - 14-Sep-2000 - pjw@rhyme.com.au
|
|
* Use symbols for tests on relkind (ie. use RELKIND_VIEW, not 'v')
|
|
* Support for relkind = RELKIND_VIEW.
|
|
* Fix bug in support for -b option (== --blobs).
|
|
* Dump views as views (using 'create view').
|
|
* Remove 'isViewRule' since we check the relkind when getting tables.
|
|
* Now uses temp table 'pgdump_oid' rather than 'pg_dump_oid' (errors otherwise).
|
|
*
|
|
* Modifications - 02-Oct-2000 - pjw@rhyme.com.au
|
|
*
|
|
* - Be more paranoid when getting views: call get_viewdef in separate statement
|
|
* so we can be more informative in error messages.
|
|
* - Support for 'isstrict' procedure attribute.
|
|
* - Disable --blobs and --table since (a) it's a pain to get ONLY the blobs for the
|
|
* table with the currently implementation, and (b) it's not clear how to restore
|
|
* a partial BLOB backup (given the current OID-based BLOB implementation).
|
|
*
|
|
* Modifications - 04-Jan-2000 - pjw@rhyme.com.au
|
|
*
|
|
* - Check ntuples == 1 for various SELECT statements.
|
|
* - Fix handling of --tables=* (multiple tables never worked properly, AFAICT)
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
|
|
/*
|
|
* Although this is not a backend module, we must include postgres.h anyway
|
|
* so that we can include a bunch of backend include files. pg_dump has
|
|
* never pretended to be very independent of the backend anyhow ...
|
|
*/
|
|
#include "postgres.h"
|
|
|
|
#include <unistd.h> /* for getopt() */
|
|
#include <ctype.h>
|
|
|
|
#include "pg_backup.h"
|
|
|
|
#ifdef HAVE_GETOPT_H
|
|
#include <getopt.h>
|
|
#endif
|
|
#ifdef HAVE_TERMIOS_H
|
|
#include <termios.h>
|
|
#endif
|
|
|
|
#include "access/attnum.h"
|
|
#include "access/htup.h"
|
|
#include "catalog/pg_class.h"
|
|
#include "catalog/pg_trigger.h"
|
|
#include "catalog/pg_type.h"
|
|
|
|
#include "libpq-fe.h"
|
|
#include "libpq/libpq-fs.h"
|
|
#ifndef HAVE_STRDUP
|
|
#include "strdup.h"
|
|
#endif
|
|
|
|
#include "pg_dump.h"
|
|
#include "pg_backup.h"
|
|
|
|
#define atooid(x) ((Oid) strtoul((x), NULL, 10))
|
|
|
|
static void dumpComment(Archive *outfile, const char *target, const char *oid);
|
|
static void dumpSequence(Archive *fout, TableInfo tbinfo);
|
|
static void dumpACL(Archive *fout, TableInfo tbinfo);
|
|
static void dumpTriggers(Archive *fout, const char *tablename,
|
|
TableInfo *tblinfo, int numTables);
|
|
static void dumpRules(Archive *fout, const char *tablename,
|
|
TableInfo *tblinfo, int numTables);
|
|
static void formatStringLiteral(PQExpBuffer buf, const char *str);
|
|
static void clearTableInfo(TableInfo *, int);
|
|
static void dumpOneFunc(Archive *fout, FuncInfo *finfo, int i,
|
|
TypeInfo *tinfo, int numTypes);
|
|
static Oid findLastBuiltinOid(const char*);
|
|
static void setMaxOid(Archive *fout);
|
|
|
|
static void AddAcl(char *aclbuf, const char *keyword);
|
|
static char *GetPrivileges(const char *s);
|
|
|
|
static int dumpBlobs(Archive *AH, char*, void*);
|
|
static int dumpDatabase(Archive *AH);
|
|
static PQExpBuffer getPKconstraint(TableInfo *tblInfo, IndInfo *indInfo);
|
|
|
|
extern char *optarg;
|
|
extern int optind,
|
|
opterr;
|
|
|
|
/* global decls */
|
|
bool g_verbose; /* User wants verbose narration of our
|
|
* activities. */
|
|
Oid g_last_builtin_oid; /* value of the last builtin oid */
|
|
Archive *g_fout; /* the script file */
|
|
PGconn *g_conn; /* the database connection */
|
|
|
|
bool force_quotes; /* User wants to suppress double-quotes */
|
|
bool dumpData; /* dump data using proper insert strings */
|
|
bool attrNames; /* put attr names into insert strings */
|
|
bool schemaOnly;
|
|
bool dataOnly;
|
|
bool aclsSkip;
|
|
|
|
char g_opaque_type[10]; /* name for the opaque type */
|
|
|
|
/* placeholders for the delimiters for comments */
|
|
char g_comment_start[10];
|
|
char g_comment_end[10];
|
|
|
|
|
|
typedef struct _dumpContext {
|
|
TableInfo *tblinfo;
|
|
int tblidx;
|
|
bool oids;
|
|
} DumpContext;
|
|
|
|
static void
|
|
help(const char *progname)
|
|
{
|
|
printf("%s dumps a database as a text file.\n\n", progname);
|
|
puts("Usage:");
|
|
printf(" %s [options] dbname\n\n", progname);
|
|
puts("Options:");
|
|
|
|
#ifdef HAVE_GETOPT_LONG
|
|
puts(
|
|
" -a, --data-only dump out only the data, not the schema\n"
|
|
" -b, --blobs dump out blob data\n"
|
|
" -c, --clean clean (drop) schema prior to create\n"
|
|
" -C, --create output commands to create database\n"
|
|
" -d, --inserts dump data as INSERT, rather than COPY, commands\n"
|
|
" -D, --attribute-inserts dump data as INSERT commands with attribute names\n"
|
|
" -f, --file=FILENAME specify output file name\n"
|
|
" -F, --format {c|f|p} output file format (custom, files, plain text)\n"
|
|
" -h, --host=HOSTNAME server host name\n"
|
|
" -i, --ignore-version proceed when database version != pg_dump version\n"
|
|
" -n, --no-quotes suppress most quotes around identifiers\n"
|
|
" -N, --quotes enable most quotes around identifiers\n"
|
|
" -o, --oids dump object ids (oids)\n"
|
|
" -O, --no-owner do not output \\connect commands in plain text\n"
|
|
" format\n"
|
|
" -p, --port=PORT server port number\n"
|
|
" -R, --no-reconnect disable ALL reconnections to the database in\n"
|
|
" plain text format\n"
|
|
" -s, --schema-only dump out only the schema, no data\n"
|
|
" -S, --superuser=NAME specify the superuser user name to use in plain\n"
|
|
" text format\n"
|
|
" -t, --table=TABLE dump for this table only (* for all)\n"
|
|
" -u, --password use password authentication\n"
|
|
" -v, --verbose verbose\n"
|
|
" -x, --no-acl do not dump ACL's (grant/revoke)\n"
|
|
" -Z, --compress {0-9} compression level for compressed formats\n"
|
|
);
|
|
#else
|
|
puts(
|
|
" -a dump out only the data, no schema\n"
|
|
" -b dump out blob data\n"
|
|
" -c clean (drop) schema prior to create\n"
|
|
" -C output commands to create database\n"
|
|
" -d dump data as INSERT, rather than COPY, commands\n"
|
|
" -D dump data as INSERT commands with attribute names\n"
|
|
" -f FILENAME specify output file name\n"
|
|
" -F {c|f|p} output file format (custom, files, plain text)\n"
|
|
" -h HOSTNAME server host name\n"
|
|
" -i proceed when database version != pg_dump version\n"
|
|
" -n suppress most quotes around identifiers\n"
|
|
" -N enable most quotes around identifiers\n"
|
|
" -o dump object ids (oids)\n"
|
|
" -O do not output \\connect commands in plain text\n"
|
|
" format\n"
|
|
" -p PORT server port number\n"
|
|
" -R disable ALL reconnections to the database in\n"
|
|
" plain text format\n"
|
|
" -s dump out only the schema, no data\n"
|
|
" -S NAME specify the superuser user name to use in plain\n"
|
|
" text format\n"
|
|
" -t TABLE dump for this table only (* for all)\n"
|
|
" -u use password authentication\n"
|
|
" -v verbose\n"
|
|
" -x do not dump ACL's (grant/revoke)\n"
|
|
" -Z {0-9} compression level for compressed formats\n"
|
|
);
|
|
#endif
|
|
puts("If no database name is not supplied, then the PGDATABASE environment\nvariable value is used.\n");
|
|
puts("Report bugs to <pgsql-bugs@postgresql.org>.");
|
|
}
|
|
|
|
|
|
static void
|
|
version(void)
|
|
{
|
|
puts("pg_dump (PostgreSQL) " PG_VERSION);
|
|
puts("Portions Copyright (c) 1996-2001, PostgreSQL Global Development Group");
|
|
puts("Portions Copyright (c) 1996 Regents of the University of California");
|
|
puts("Read the file COPYRIGHT to see the usage and distribution terms.");
|
|
}
|
|
|
|
|
|
static void
|
|
exit_nicely(PGconn *conn)
|
|
{
|
|
PQfinish(conn);
|
|
exit(1);
|
|
}
|
|
|
|
|
|
#define COPYBUFSIZ 8192
|
|
|
|
/*
|
|
* Dump a table's contents for loading using the COPY command
|
|
* - this routine is called by the Archiver when it wants the table
|
|
* to be dumped.
|
|
*/
|
|
|
|
static int
|
|
dumpClasses_nodumpData(Archive *fout, char* oid, void *dctxv)
|
|
{
|
|
const DumpContext *dctx = (DumpContext*)dctxv;
|
|
const char *classname = dctx->tblinfo[dctx->tblidx].relname;
|
|
const bool oids = dctx->oids;
|
|
|
|
PGresult *res;
|
|
char query[255];
|
|
int ret;
|
|
bool copydone;
|
|
char copybuf[COPYBUFSIZ];
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s dumping out the contents of table %s\n", g_comment_start, classname);
|
|
|
|
if (oids == true)
|
|
{
|
|
/*
|
|
* archprintf(fout, "COPY %s WITH OIDS FROM stdin;\n",
|
|
* fmtId(classname, force_quotes));
|
|
*
|
|
* - Not used as of V1.3 (needs to be in ArchiveEntry call)
|
|
*
|
|
*/
|
|
|
|
sprintf(query, "COPY %s WITH OIDS TO stdout;\n",
|
|
fmtId(classname, force_quotes));
|
|
}
|
|
else
|
|
{
|
|
/*
|
|
*archprintf(fout, "COPY %s FROM stdin;\n", fmtId(classname, force_quotes));
|
|
*
|
|
* - Not used as of V1.3 (needs to be in ArchiveEntry call)
|
|
*
|
|
*/
|
|
|
|
sprintf(query, "COPY %s TO stdout;\n", fmtId(classname, force_quotes));
|
|
}
|
|
res = PQexec(g_conn, query);
|
|
if (!res ||
|
|
PQresultStatus(res) == PGRES_FATAL_ERROR)
|
|
{
|
|
fprintf(stderr, "SQL query to dump the contents of Table '%s' "
|
|
"did not execute. Explanation from backend: '%s'.\n"
|
|
"The query was: '%s'.\n",
|
|
classname, PQerrorMessage(g_conn), query);
|
|
exit_nicely(g_conn);
|
|
}
|
|
else
|
|
{
|
|
if (PQresultStatus(res) != PGRES_COPY_OUT)
|
|
{
|
|
fprintf(stderr, "SQL query to dump the contents of Table '%s' "
|
|
"executed abnormally.\n"
|
|
"PQexec() returned status %d when %d was expected.\n"
|
|
"The query was: '%s'.\n",
|
|
classname, PQresultStatus(res), PGRES_COPY_OUT, query);
|
|
exit_nicely(g_conn);
|
|
}
|
|
else
|
|
{
|
|
copydone = false;
|
|
|
|
while (!copydone)
|
|
{
|
|
ret = PQgetline(g_conn, copybuf, COPYBUFSIZ);
|
|
|
|
if (copybuf[0] == '\\' &&
|
|
copybuf[1] == '.' &&
|
|
copybuf[2] == '\0')
|
|
{
|
|
copydone = true; /* don't print this... */
|
|
}
|
|
else
|
|
{
|
|
archputs(copybuf, fout);
|
|
switch (ret)
|
|
{
|
|
case EOF:
|
|
copydone = true;
|
|
/* FALLTHROUGH */
|
|
case 0:
|
|
archputc('\n', fout);
|
|
break;
|
|
case 1:
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* THROTTLE:
|
|
*
|
|
* There was considerable discussion in late July, 2000 regarding slowing down
|
|
* pg_dump when backing up large tables. Users with both slow & fast (muti-processor)
|
|
* machines experienced performance degradation when doing a backup.
|
|
*
|
|
* Initial attempts based on sleeping for a number of ms for each ms of work were deemed
|
|
* too complex, then a simple 'sleep in each loop' implementation was suggested. The latter
|
|
* failed because the loop was too tight. Finally, the following was implemented:
|
|
*
|
|
* If throttle is non-zero, then
|
|
* See how long since the last sleep.
|
|
* Work out how long to sleep (based on ratio).
|
|
* If sleep is more than 100ms, then
|
|
* sleep
|
|
* reset timer
|
|
* EndIf
|
|
* EndIf
|
|
*
|
|
* where the throttle value was the number of ms to sleep per ms of work. The calculation was
|
|
* done in each loop.
|
|
*
|
|
* Most of the hard work is done in the backend, and this solution still did not work
|
|
* particularly well: on slow machines, the ratio was 50:1, and on medium paced machines, 1:1,
|
|
* and on fast multi-processor machines, it had little or no effect, for reasons that were unclear.
|
|
*
|
|
* Further discussion ensued, and the proposal was dropped.
|
|
*
|
|
* For those people who want this feature, it can be implemented using gettimeofday in each
|
|
* loop, calculating the time since last sleep, multiplying that by the sleep ratio, then
|
|
* if the result is more than a preset 'minimum sleep time' (say 100ms), call the 'select'
|
|
* function to sleep for a subsecond period ie.
|
|
*
|
|
* select(0, NULL, NULL, NULL, &tvi);
|
|
*
|
|
* This will return after the interval specified in the structure tvi. Fianally, call
|
|
* gettimeofday again to save the 'last sleep time'.
|
|
*/
|
|
}
|
|
archprintf(fout, "\\.\n");
|
|
}
|
|
ret = PQendcopy(g_conn);
|
|
if (ret != 0)
|
|
{
|
|
fprintf(stderr, "SQL query to dump the contents of Table '%s' "
|
|
"did not execute correctly. After we read all the "
|
|
"table contents from the backend, PQendcopy() failed. "
|
|
"Explanation from backend: '%s'.\n"
|
|
"The query was: '%s'.\n",
|
|
classname, PQerrorMessage(g_conn), query);
|
|
PQclear(res);
|
|
exit_nicely(g_conn);
|
|
}
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int
|
|
dumpClasses_dumpData(Archive *fout, char* oid, void *dctxv)
|
|
{
|
|
const DumpContext *dctx = (DumpContext*)dctxv;
|
|
const char *classname = dctx->tblinfo[dctx->tblidx].relname;
|
|
|
|
PGresult *res;
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
int tuple;
|
|
int field;
|
|
|
|
appendPQExpBuffer(q, "SELECT * FROM %s", fmtId(classname, force_quotes));
|
|
res = PQexec(g_conn, q->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpClasses(): command failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
for (tuple = 0; tuple < PQntuples(res); tuple++)
|
|
{
|
|
archprintf(fout, "INSERT INTO %s ", fmtId(classname, force_quotes));
|
|
if (attrNames == true)
|
|
{
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "(");
|
|
for (field = 0; field < PQnfields(res); field++)
|
|
{
|
|
if (field > 0)
|
|
appendPQExpBuffer(q, ",");
|
|
appendPQExpBuffer(q, fmtId(PQfname(res, field), force_quotes));
|
|
}
|
|
appendPQExpBuffer(q, ") ");
|
|
archprintf(fout, "%s", q->data);
|
|
}
|
|
archprintf(fout, "VALUES (");
|
|
for (field = 0; field < PQnfields(res); field++)
|
|
{
|
|
if (field > 0)
|
|
archprintf(fout, ",");
|
|
if (PQgetisnull(res, tuple, field))
|
|
{
|
|
archprintf(fout, "NULL");
|
|
continue;
|
|
}
|
|
switch (PQftype(res, field))
|
|
{
|
|
case INT2OID:
|
|
case INT4OID:
|
|
case OIDOID: /* int types */
|
|
case FLOAT4OID:
|
|
case FLOAT8OID:/* float types */
|
|
/* These types are printed without quotes */
|
|
archprintf(fout, "%s",
|
|
PQgetvalue(res, tuple, field));
|
|
break;
|
|
case ZPBITOID:
|
|
case VARBITOID:
|
|
archprintf(fout, "B'%s'",
|
|
PQgetvalue(res, tuple, field));
|
|
break;
|
|
default:
|
|
|
|
/*
|
|
* All other types are printed as string literals,
|
|
* with appropriate escaping of special characters.
|
|
*/
|
|
resetPQExpBuffer(q);
|
|
formatStringLiteral(q, PQgetvalue(res, tuple, field));
|
|
archprintf(fout, "%s", q->data);
|
|
break;
|
|
}
|
|
}
|
|
archprintf(fout, ");\n");
|
|
}
|
|
PQclear(res);
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* Convert a string value to an SQL string literal,
|
|
* with appropriate escaping of special characters.
|
|
* Quote mark ' goes to '' per SQL standard, other
|
|
* stuff goes to \ sequences.
|
|
* The literal is appended to the given PQExpBuffer.
|
|
*/
|
|
static void
|
|
formatStringLiteral(PQExpBuffer buf, const char *str)
|
|
{
|
|
appendPQExpBufferChar(buf, '\'');
|
|
while (*str)
|
|
{
|
|
char ch = *str++;
|
|
|
|
if (ch == '\\' || ch == '\'')
|
|
{
|
|
appendPQExpBufferChar(buf, ch); /* double these */
|
|
appendPQExpBufferChar(buf, ch);
|
|
}
|
|
else if ((unsigned char) ch < (unsigned char) ' ' &&
|
|
ch != '\n' && ch != '\t')
|
|
{
|
|
/* generate octal escape for control chars other than whitespace */
|
|
appendPQExpBufferChar(buf, '\\');
|
|
appendPQExpBufferChar(buf, ((ch >> 6) & 3) + '0');
|
|
appendPQExpBufferChar(buf, ((ch >> 3) & 7) + '0');
|
|
appendPQExpBufferChar(buf, (ch & 7) + '0');
|
|
}
|
|
else
|
|
appendPQExpBufferChar(buf, ch);
|
|
}
|
|
appendPQExpBufferChar(buf, '\'');
|
|
}
|
|
|
|
/*
|
|
* DumpClasses -
|
|
* dump the contents of all the classes.
|
|
*/
|
|
static void
|
|
dumpClasses(const TableInfo *tblinfo, const int numTables, Archive *fout,
|
|
const char *onlytable, const bool oids, const bool force_quotes)
|
|
{
|
|
|
|
int i;
|
|
char *all_only;
|
|
DataDumperPtr dumpFn;
|
|
DumpContext *dumpCtx;
|
|
char *oidsPart;
|
|
char copyBuf[512];
|
|
char *copyStmt;
|
|
|
|
if (onlytable == NULL || (strlen(onlytable) == 0) )
|
|
all_only = "all";
|
|
else
|
|
all_only = "only";
|
|
|
|
if (oids == true)
|
|
oidsPart = "WITH OIDS ";
|
|
else
|
|
oidsPart = "";
|
|
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s preparing to dump out the contents of %s %d table%s/sequence%s %s\n",
|
|
g_comment_start, all_only,
|
|
(onlytable == NULL || (strlen(onlytable) == 0)) ? numTables : 1,
|
|
(onlytable == NULL || (strlen(onlytable) == 0)) ? "s" : "",
|
|
(onlytable == NULL || (strlen(onlytable) == 0)) ? "s" : "",
|
|
g_comment_end);
|
|
|
|
/* Dump SEQUENCEs first (if dataOnly) */
|
|
if (dataOnly)
|
|
{
|
|
for (i = 0; i < numTables; i++)
|
|
{
|
|
if (!(tblinfo[i].sequence))
|
|
continue;
|
|
if (!onlytable || (strcmp(tblinfo[i].relname, onlytable) == 0) || (strlen(onlytable) == 0) )
|
|
{
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s dumping out schema of sequence '%s' %s\n",
|
|
g_comment_start, tblinfo[i].relname, g_comment_end);
|
|
/* becomeUser(fout, tblinfo[i].usename); */
|
|
dumpSequence(fout, tblinfo[i]);
|
|
}
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < numTables; i++)
|
|
{
|
|
const char *classname = tblinfo[i].relname;
|
|
|
|
/* Skip VIEW relations */
|
|
if (tblinfo[i].viewdef != NULL)
|
|
continue;
|
|
|
|
if (tblinfo[i].sequence)/* already dumped */
|
|
continue;
|
|
|
|
if (!onlytable || (strcmp(classname, onlytable) == 0) || (strlen(onlytable) == 0))
|
|
{
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s preparing to dump out the contents of Table '%s' %s\n",
|
|
g_comment_start, classname, g_comment_end);
|
|
|
|
/* becomeUser(fout, tblinfo[i].usename); */
|
|
|
|
dumpCtx = (DumpContext*)malloc(sizeof(DumpContext));
|
|
dumpCtx->tblinfo = (TableInfo*)tblinfo;
|
|
dumpCtx->tblidx = i;
|
|
dumpCtx->oids = oids;
|
|
|
|
if (!dumpData) /* Dump/restore using COPY */
|
|
{
|
|
dumpFn = dumpClasses_nodumpData;
|
|
/* dumpClasses_nodumpData(fout, classname, oids); */
|
|
sprintf(copyBuf, "COPY %s %s FROM stdin;\n", fmtId(tblinfo[i].relname, force_quotes),
|
|
oidsPart);
|
|
copyStmt = copyBuf;
|
|
}
|
|
else /* Restore using INSERT */
|
|
{
|
|
dumpFn = dumpClasses_dumpData;
|
|
/* dumpClasses_dumpData(fout, classname); */
|
|
copyStmt = NULL;
|
|
}
|
|
|
|
ArchiveEntry(fout, tblinfo[i].oid, fmtId(tblinfo[i].relname, false),
|
|
"TABLE DATA", NULL, "", "", copyStmt, tblinfo[i].usename,
|
|
dumpFn, dumpCtx);
|
|
}
|
|
}
|
|
}
|
|
|
|
int
|
|
main(int argc, char **argv)
|
|
{
|
|
int c;
|
|
const char *progname;
|
|
const char *filename = NULL;
|
|
const char *format = "p";
|
|
const char *dbname = NULL;
|
|
const char *pghost = NULL;
|
|
const char *pgport = NULL;
|
|
char *tablename = NULL;
|
|
bool oids = false;
|
|
TableInfo *tblinfo;
|
|
int numTables;
|
|
bool use_password = false;
|
|
int compressLevel = -1;
|
|
bool ignore_version = false;
|
|
int plainText = 0;
|
|
int outputClean = 0;
|
|
int outputCreate = 0;
|
|
int outputBlobs = 0;
|
|
int outputNoOwner = 0;
|
|
int outputNoReconnect = 0;
|
|
char *outputSuperuser = NULL;
|
|
|
|
RestoreOptions *ropt;
|
|
|
|
#ifdef HAVE_GETOPT_LONG
|
|
static struct option long_options[] = {
|
|
{"data-only", no_argument, NULL, 'a'},
|
|
{"blobs", no_argument, NULL, 'b' },
|
|
{"clean", no_argument, NULL, 'c'},
|
|
{"create", no_argument, NULL, 'C'},
|
|
{"file", required_argument, NULL, 'f'},
|
|
{"format", required_argument, NULL, 'F'},
|
|
{"inserts", no_argument, NULL, 'd'},
|
|
{"attribute-inserts", no_argument, NULL, 'D'},
|
|
{"host", required_argument, NULL, 'h'},
|
|
{"ignore-version", no_argument, NULL, 'i'},
|
|
{"no-reconnect", no_argument, NULL, 'R'},
|
|
{"no-quotes", no_argument, NULL, 'n'},
|
|
{"quotes", no_argument, NULL, 'N'},
|
|
{"oids", no_argument, NULL, 'o'},
|
|
{"no-owner", no_argument, NULL, 'O'},
|
|
{"port", required_argument, NULL, 'p'},
|
|
{"schema-only", no_argument, NULL, 's'},
|
|
{"superuser", required_argument, NULL, 'S'},
|
|
{"table", required_argument, NULL, 't'},
|
|
{"password", no_argument, NULL, 'u'},
|
|
{"verbose", no_argument, NULL, 'v'},
|
|
{"no-acl", no_argument, NULL, 'x'},
|
|
{"compress", required_argument, NULL, 'Z'},
|
|
{"help", no_argument, NULL, '?'},
|
|
{"version", no_argument, NULL, 'V'}
|
|
};
|
|
int optindex;
|
|
|
|
#endif
|
|
|
|
g_verbose = false;
|
|
force_quotes = true;
|
|
|
|
strcpy(g_comment_start, "-- ");
|
|
g_comment_end[0] = '\0';
|
|
strcpy(g_opaque_type, "opaque");
|
|
|
|
dataOnly = schemaOnly = dumpData = attrNames = false;
|
|
|
|
if (!strrchr(argv[0], SEP_CHAR))
|
|
progname = argv[0];
|
|
else
|
|
progname = strrchr(argv[0], SEP_CHAR) + 1;
|
|
|
|
/* Set defaulty options based on progname */
|
|
if (strcmp(progname, "pg_backup") == 0)
|
|
{
|
|
format = "c";
|
|
outputBlobs = true;
|
|
}
|
|
|
|
if (argc > 1)
|
|
{
|
|
if (strcmp(argv[1], "--help")==0 || strcmp(argv[1], "-?")==0)
|
|
{
|
|
help(progname);
|
|
exit(0);
|
|
}
|
|
if (strcmp(argv[1], "--version")==0 || strcmp(argv[1], "-V")==0)
|
|
{
|
|
version();
|
|
exit(0);
|
|
}
|
|
}
|
|
|
|
#ifdef HAVE_GETOPT_LONG
|
|
while ((c = getopt_long(argc, argv, "abcCdDf:F:h:inNoOp:RsS:t:uvxzZ:V?", long_options, &optindex)) != -1)
|
|
#else
|
|
while ((c = getopt(argc, argv, "abcCdDf:F:h:inNoOp:RsS:t:uvxzZ:V?-")) != -1)
|
|
#endif
|
|
|
|
{
|
|
switch (c)
|
|
{
|
|
case 'a': /* Dump data only */
|
|
dataOnly = true;
|
|
break;
|
|
|
|
case 'b': /* Dump blobs */
|
|
outputBlobs = true;
|
|
break;
|
|
|
|
case 'c': /* clean (i.e., drop) schema prior to
|
|
* create */
|
|
outputClean = 1;
|
|
break;
|
|
|
|
case 'C': /* Create DB */
|
|
|
|
outputCreate = 1;
|
|
break;
|
|
|
|
case 'd': /* dump data as proper insert strings */
|
|
dumpData = true;
|
|
break;
|
|
|
|
case 'D': /* dump data as proper insert strings with
|
|
* attr names */
|
|
dumpData = true;
|
|
attrNames = true;
|
|
break;
|
|
|
|
case 'f':
|
|
filename = optarg;
|
|
break;
|
|
|
|
case 'F':
|
|
format = optarg;
|
|
break;
|
|
|
|
case 'h': /* server host */
|
|
pghost = optarg;
|
|
break;
|
|
|
|
case 'i': /* ignore database version mismatch */
|
|
ignore_version = true;
|
|
break;
|
|
|
|
case 'n': /* Do not force double-quotes on
|
|
* identifiers */
|
|
force_quotes = false;
|
|
break;
|
|
|
|
case 'N': /* Force double-quotes on identifiers */
|
|
force_quotes = true;
|
|
break;
|
|
|
|
case 'o': /* Dump oids */
|
|
oids = true;
|
|
break;
|
|
|
|
|
|
case 'O': /* Don't reconnect to match owner */
|
|
outputNoOwner = 1;
|
|
break;
|
|
|
|
case 'p': /* server port */
|
|
pgport = optarg;
|
|
break;
|
|
|
|
case 'R': /* No reconnect */
|
|
outputNoReconnect = 1;
|
|
break;
|
|
|
|
case 's': /* dump schema only */
|
|
schemaOnly = true;
|
|
break;
|
|
|
|
case 'S': /* Username for superuser in plain text output */
|
|
outputSuperuser = strdup(optarg);
|
|
break;
|
|
|
|
case 't': /* Dump data for this table only */
|
|
{
|
|
int i;
|
|
|
|
tablename = strdup(optarg);
|
|
|
|
/*
|
|
* quoted string? Then strip quotes and preserve
|
|
* case...
|
|
*/
|
|
if (tablename[0] == '"')
|
|
{
|
|
strcpy(tablename, &tablename[1]);
|
|
if (*(tablename + strlen(tablename) - 1) == '"')
|
|
*(tablename + strlen(tablename) - 1) = '\0';
|
|
}
|
|
/* otherwise, convert table name to lowercase... */
|
|
else
|
|
{
|
|
for (i = 0; tablename[i]; i++)
|
|
if (isupper((unsigned char) tablename[i]))
|
|
tablename[i] = tolower((unsigned char) tablename[i]);
|
|
|
|
/* '*' is a special case meaning ALL tables, but only if unquoted */
|
|
if (strcmp(tablename,"*") == 0)
|
|
tablename[0] = '\0';
|
|
|
|
}
|
|
}
|
|
break;
|
|
|
|
case 'u':
|
|
use_password = true;
|
|
break;
|
|
|
|
case 'v': /* verbose */
|
|
g_verbose = true;
|
|
break;
|
|
|
|
case 'x': /* skip ACL dump */
|
|
aclsSkip = true;
|
|
break;
|
|
|
|
case 'Z': /* Compression Level */
|
|
compressLevel = atoi(optarg);
|
|
break;
|
|
|
|
case 'V':
|
|
version();
|
|
exit(0);
|
|
break;
|
|
|
|
#ifndef HAVE_GETOPT_LONG
|
|
case '-':
|
|
fprintf(stderr, "%s was compiled without support for long options.\n"
|
|
"Use --help for help on invocation options.\n", progname);
|
|
exit(1);
|
|
break;
|
|
#endif
|
|
default:
|
|
fprintf(stderr, "Try '%s --help' for more information.\n", progname);
|
|
exit(1);
|
|
}
|
|
}
|
|
|
|
if (optind < (argc - 1)) {
|
|
fprintf(stderr,
|
|
"%s: extra parameters found on command line after '%s' (first is '%s').\n"
|
|
"Please respecify command.\nUse --help for help on invocation options.\n",
|
|
progname, argv[optind], argv[optind+1]);
|
|
exit(1);
|
|
}
|
|
|
|
if (dataOnly && schemaOnly)
|
|
{
|
|
fprintf(stderr,
|
|
"%s: 'Schema Only' and 'Data Only' are incompatible options.\n",
|
|
progname);
|
|
exit(1);
|
|
}
|
|
|
|
if (outputBlobs && tablename != NULL && strlen(tablename) > 0 )
|
|
{
|
|
fprintf(stderr,
|
|
"%s: BLOB output is not supported for a single table. Use all tables or a full dump instead.\n",
|
|
progname);
|
|
exit(1);
|
|
}
|
|
|
|
if (dumpData == true && oids == true)
|
|
{
|
|
fprintf(stderr,
|
|
"%s: INSERT's can not set oids, so INSERT and OID options can not be used together.\n",
|
|
progname);
|
|
exit(1);
|
|
}
|
|
|
|
if (outputBlobs == true && (format[0] == 'p' || format[0] == 'P') )
|
|
{
|
|
fprintf(stderr,
|
|
"%s: BLOB output is not supported for plain text dump files. Use a different output format.\n",
|
|
progname);
|
|
exit(1);
|
|
}
|
|
|
|
/* open the output file */
|
|
switch (format[0]) {
|
|
|
|
case 'c':
|
|
case 'C':
|
|
g_fout = CreateArchive(filename, archCustom, compressLevel);
|
|
break;
|
|
|
|
case 'f':
|
|
case 'F':
|
|
g_fout = CreateArchive(filename, archFiles, compressLevel);
|
|
break;
|
|
|
|
case 'p':
|
|
case 'P':
|
|
plainText = 1;
|
|
g_fout = CreateArchive(filename, archNull, 0);
|
|
break;
|
|
|
|
case 't':
|
|
case 'T':
|
|
g_fout = CreateArchive(filename, archTar, compressLevel);
|
|
break;
|
|
|
|
default:
|
|
fprintf(stderr,
|
|
"%s: invalid output format '%s' specified\n", progname, format);
|
|
exit(1);
|
|
}
|
|
|
|
if (g_fout == NULL)
|
|
{
|
|
fprintf(stderr,
|
|
"%s: could not open output file named %s for writing\n",
|
|
progname, filename);
|
|
exit(1);
|
|
}
|
|
|
|
/* Let the archiver know how noisy to be */
|
|
g_fout->verbose = g_verbose;
|
|
|
|
dbname = argv[optind];
|
|
|
|
/* Open the database using the Archiver, so it knows about it. Errors mean death */
|
|
g_conn = ConnectDatabase(g_fout, dbname, pghost, pgport, use_password, ignore_version);
|
|
|
|
/*
|
|
* Start serializable transaction to dump consistent data
|
|
*/
|
|
{
|
|
PGresult *res;
|
|
|
|
res = PQexec(g_conn, "begin");
|
|
if (!res || PQresultStatus(res) != PGRES_COMMAND_OK)
|
|
exit_horribly(g_fout, "BEGIN command failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
|
|
PQclear(res);
|
|
res = PQexec(g_conn, "set transaction isolation level serializable");
|
|
if (!res || PQresultStatus(res) != PGRES_COMMAND_OK)
|
|
exit_horribly(g_fout, "SET TRANSACTION command failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
|
|
PQclear(res);
|
|
}
|
|
|
|
g_last_builtin_oid = findLastBuiltinOid(dbname);
|
|
|
|
/* Dump the database definition */
|
|
if (!dataOnly)
|
|
dumpDatabase(g_fout);
|
|
|
|
if (oids == true)
|
|
setMaxOid(g_fout);
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s last builtin oid is %u %s\n",
|
|
g_comment_start, g_last_builtin_oid, g_comment_end);
|
|
tblinfo = dumpSchema(g_fout, &numTables, tablename, aclsSkip, oids, schemaOnly, dataOnly);
|
|
|
|
if (!schemaOnly)
|
|
{
|
|
dumpClasses(tblinfo, numTables, g_fout, tablename, oids, force_quotes);
|
|
}
|
|
|
|
if (outputBlobs)
|
|
ArchiveEntry(g_fout, "0", "BLOBS", "BLOBS", NULL, "", "", "", "", dumpBlobs, 0);
|
|
|
|
if (!dataOnly) /* dump indexes and triggers at the end
|
|
* for performance */
|
|
{
|
|
dumpTriggers(g_fout, tablename, tblinfo, numTables);
|
|
dumpRules(g_fout, tablename, tblinfo, numTables);
|
|
}
|
|
|
|
/* Now sort the output nicely */
|
|
SortTocByOID(g_fout);
|
|
MoveToStart(g_fout, "DATABASE");
|
|
MoveToEnd(g_fout, "TABLE DATA");
|
|
MoveToEnd(g_fout, "BLOBS");
|
|
MoveToEnd(g_fout, "INDEX");
|
|
MoveToEnd(g_fout, "CONSTRAINT");
|
|
MoveToEnd(g_fout, "TRIGGER");
|
|
MoveToEnd(g_fout, "RULE");
|
|
MoveToEnd(g_fout, "SEQUENCE SET");
|
|
|
|
if (plainText)
|
|
{
|
|
ropt = NewRestoreOptions();
|
|
ropt->filename = (char*)filename;
|
|
ropt->dropSchema = outputClean;
|
|
ropt->aclsSkip = aclsSkip;
|
|
ropt->superuser = outputSuperuser;
|
|
ropt->create = outputCreate;
|
|
ropt->noOwner = outputNoOwner;
|
|
ropt->noReconnect = outputNoReconnect;
|
|
|
|
if (outputSuperuser)
|
|
ropt->superuser = outputSuperuser;
|
|
else
|
|
ropt->superuser = PQuser(g_conn);
|
|
|
|
if (compressLevel == -1)
|
|
ropt->compression = 0;
|
|
else
|
|
ropt->compression = compressLevel;
|
|
|
|
RestoreArchive(g_fout, ropt);
|
|
}
|
|
|
|
CloseArchive(g_fout);
|
|
|
|
clearTableInfo(tblinfo, numTables);
|
|
PQfinish(g_conn);
|
|
exit(0);
|
|
}
|
|
|
|
/*
|
|
* dumpDatabase:
|
|
* dump the database definition
|
|
*
|
|
*/
|
|
static int
|
|
dumpDatabase(Archive *AH)
|
|
{
|
|
PQExpBuffer dbQry = createPQExpBuffer();
|
|
PQExpBuffer delQry = createPQExpBuffer();
|
|
PQExpBuffer creaQry = createPQExpBuffer();
|
|
PGresult *res;
|
|
int ntups;
|
|
int i_dba;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s saving database definition\n", g_comment_start);
|
|
|
|
/* Get the dba */
|
|
appendPQExpBuffer(dbQry, "select (select usename from pg_user where datdba = usesysid) as dba from pg_database"
|
|
" where datname = ");
|
|
formatStringLiteral(dbQry, PQdb(g_conn));
|
|
|
|
res = PQexec(g_conn, dbQry->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getDatabase(): SELECT failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
if (ntups != 1)
|
|
{
|
|
fprintf(stderr, "getDatabase(): SELECT returned %d databases.\n", ntups);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
appendPQExpBuffer(creaQry, "Create Database \"%s\";\n", PQdb(g_conn));
|
|
appendPQExpBuffer(delQry, "Drop Database \"%s\";\n", PQdb(g_conn));
|
|
i_dba = PQfnumber(res, "dba");
|
|
|
|
ArchiveEntry(AH, "0" /* OID */, PQdb(g_conn) /* Name */, "DATABASE", NULL,
|
|
creaQry->data /* Create */, delQry->data /*Del*/,
|
|
"" /* Copy */, PQgetvalue(res, 0, i_dba) /*Owner*/,
|
|
NULL /* Dumper */, NULL /* Dumper Arg */);
|
|
|
|
PQclear(res);
|
|
|
|
return 1;
|
|
}
|
|
|
|
|
|
/*
|
|
* dumpBlobs:
|
|
* dump all blobs
|
|
*
|
|
*/
|
|
|
|
#define loBufSize 16384
|
|
#define loFetchSize 1000
|
|
|
|
static int
|
|
dumpBlobs(Archive *AH, char* junkOid, void *junkVal)
|
|
{
|
|
PQExpBuffer oidQry = createPQExpBuffer();
|
|
PQExpBuffer oidFetchQry = createPQExpBuffer();
|
|
PGresult *res;
|
|
int i;
|
|
int loFd;
|
|
char buf[loBufSize];
|
|
int cnt;
|
|
int blobOid;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s saving BLOBs\n", g_comment_start);
|
|
|
|
/* Cursor to get all BLOB tables */
|
|
appendPQExpBuffer(oidQry, "Declare blobOid Cursor for SELECT DISTINCT loid FROM pg_largeobject");
|
|
|
|
res = PQexec(g_conn, oidQry->data);
|
|
if (!res || PQresultStatus(res) != PGRES_COMMAND_OK)
|
|
{
|
|
fprintf(stderr, "dumpBlobs(): Declare Cursor failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Fetch for cursor */
|
|
appendPQExpBuffer(oidFetchQry, "Fetch %d in blobOid", loFetchSize);
|
|
|
|
do {
|
|
/* Do a fetch */
|
|
PQclear(res);
|
|
res = PQexec(g_conn, oidFetchQry->data);
|
|
|
|
if (!res || PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpBlobs(): Fetch Cursor failed. Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Process the tuples, if any */
|
|
for (i = 0; i < PQntuples(res); i++)
|
|
{
|
|
blobOid = atoi(PQgetvalue(res, i, 0));
|
|
/* Open the BLOB */
|
|
loFd = lo_open(g_conn, blobOid, INV_READ);
|
|
if (loFd == -1)
|
|
{
|
|
fprintf(stderr, "dumpBlobs(): Could not open large object. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
StartBlob(AH, blobOid);
|
|
|
|
/* Now read it in chunks, sending data to archive */
|
|
do {
|
|
cnt = lo_read(g_conn, loFd, buf, loBufSize);
|
|
if (cnt < 0) {
|
|
fprintf(stderr, "dumpBlobs(): Error reading large object. "
|
|
" Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
WriteData(AH, buf, cnt);
|
|
|
|
} while (cnt > 0);
|
|
|
|
lo_close(g_conn, loFd);
|
|
|
|
EndBlob(AH, blobOid);
|
|
|
|
}
|
|
} while (PQntuples(res) > 0);
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*
|
|
* getTypes:
|
|
* read all base types in the system catalogs and return them in the
|
|
* TypeInfo* structure
|
|
*
|
|
* numTypes is set to the number of types read in
|
|
*
|
|
*/
|
|
TypeInfo *
|
|
getTypes(int *numTypes)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
TypeInfo *tinfo;
|
|
|
|
int i_oid;
|
|
int i_typowner;
|
|
int i_typname;
|
|
int i_typlen;
|
|
int i_typprtlen;
|
|
int i_typinput;
|
|
int i_typoutput;
|
|
int i_typreceive;
|
|
int i_typsend;
|
|
int i_typelem;
|
|
int i_typdelim;
|
|
int i_typdefault;
|
|
int i_typrelid;
|
|
int i_typbyval;
|
|
int i_usename;
|
|
int i_typedefn;
|
|
|
|
/* find all base types */
|
|
|
|
/*
|
|
* we include even the built-in types because those may be used as
|
|
* array elements by user-defined types
|
|
*/
|
|
|
|
/*
|
|
* we filter out the built-in types when we dump out the types
|
|
*/
|
|
|
|
appendPQExpBuffer(query, "SELECT pg_type.oid, typowner, typname, typlen, typprtlen, "
|
|
"typinput, typoutput, typreceive, typsend, typelem, typdelim, "
|
|
"typdefault, typrelid, typbyval, "
|
|
"(select usename from pg_user where typowner = usesysid) as usename, "
|
|
"format_type(pg_type.oid, NULL) as typedefn "
|
|
"from pg_type" );
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTypes(): SELECT failed. Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
tinfo = (TypeInfo *) malloc(ntups * sizeof(TypeInfo));
|
|
|
|
i_oid = PQfnumber(res, "oid");
|
|
i_typowner = PQfnumber(res, "typowner");
|
|
i_typname = PQfnumber(res, "typname");
|
|
i_typlen = PQfnumber(res, "typlen");
|
|
i_typprtlen = PQfnumber(res, "typprtlen");
|
|
i_typinput = PQfnumber(res, "typinput");
|
|
i_typoutput = PQfnumber(res, "typoutput");
|
|
i_typreceive = PQfnumber(res, "typreceive");
|
|
i_typsend = PQfnumber(res, "typsend");
|
|
i_typelem = PQfnumber(res, "typelem");
|
|
i_typdelim = PQfnumber(res, "typdelim");
|
|
i_typdefault = PQfnumber(res, "typdefault");
|
|
i_typrelid = PQfnumber(res, "typrelid");
|
|
i_typbyval = PQfnumber(res, "typbyval");
|
|
i_usename = PQfnumber(res, "usename");
|
|
i_typedefn = PQfnumber(res, "typedefn");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
tinfo[i].oid = strdup(PQgetvalue(res, i, i_oid));
|
|
tinfo[i].typowner = strdup(PQgetvalue(res, i, i_typowner));
|
|
tinfo[i].typname = strdup(PQgetvalue(res, i, i_typname));
|
|
tinfo[i].typlen = strdup(PQgetvalue(res, i, i_typlen));
|
|
tinfo[i].typprtlen = strdup(PQgetvalue(res, i, i_typprtlen));
|
|
tinfo[i].typinput = strdup(PQgetvalue(res, i, i_typinput));
|
|
tinfo[i].typoutput = strdup(PQgetvalue(res, i, i_typoutput));
|
|
tinfo[i].typreceive = strdup(PQgetvalue(res, i, i_typreceive));
|
|
tinfo[i].typsend = strdup(PQgetvalue(res, i, i_typsend));
|
|
tinfo[i].typelem = strdup(PQgetvalue(res, i, i_typelem));
|
|
tinfo[i].typdelim = strdup(PQgetvalue(res, i, i_typdelim));
|
|
tinfo[i].typdefault = strdup(PQgetvalue(res, i, i_typdefault));
|
|
tinfo[i].typrelid = strdup(PQgetvalue(res, i, i_typrelid));
|
|
tinfo[i].usename = strdup(PQgetvalue(res, i, i_usename));
|
|
tinfo[i].typedefn = strdup(PQgetvalue(res, i, i_typedefn));
|
|
|
|
if (strlen(tinfo[i].usename) == 0)
|
|
fprintf(stderr, "WARNING: owner of type '%s' appears to be invalid\n",tinfo[i].typname);
|
|
|
|
if (strcmp(PQgetvalue(res, i, i_typbyval), "f") == 0)
|
|
tinfo[i].passedbyvalue = 0;
|
|
else
|
|
tinfo[i].passedbyvalue = 1;
|
|
|
|
/*
|
|
* check for user-defined array types, omit system generated ones
|
|
*/
|
|
if ((strcmp(tinfo[i].typelem, "0") != 0) &&
|
|
tinfo[i].typname[0] != '_')
|
|
tinfo[i].isArray = 1;
|
|
else
|
|
tinfo[i].isArray = 0;
|
|
}
|
|
|
|
*numTypes = ntups;
|
|
|
|
PQclear(res);
|
|
|
|
return tinfo;
|
|
}
|
|
|
|
/*
|
|
* getOperators:
|
|
* read all operators in the system catalogs and return them in the
|
|
* OprInfo* structure
|
|
*
|
|
* numOprs is set to the number of operators read in
|
|
*
|
|
*
|
|
*/
|
|
OprInfo *
|
|
getOperators(int *numOprs)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
|
|
OprInfo *oprinfo;
|
|
|
|
int i_oid;
|
|
int i_oprname;
|
|
int i_oprkind;
|
|
int i_oprcode;
|
|
int i_oprleft;
|
|
int i_oprright;
|
|
int i_oprcom;
|
|
int i_oprnegate;
|
|
int i_oprrest;
|
|
int i_oprjoin;
|
|
int i_oprcanhash;
|
|
int i_oprlsortop;
|
|
int i_oprrsortop;
|
|
int i_usename;
|
|
|
|
/*
|
|
* find all operators, including builtin operators, filter out
|
|
* system-defined operators at dump-out time
|
|
*/
|
|
|
|
appendPQExpBuffer(query, "SELECT pg_operator.oid, oprname, oprkind, oprcode, "
|
|
"oprleft, oprright, oprcom, oprnegate, oprrest, oprjoin, "
|
|
"oprcanhash, oprlsortop, oprrsortop, "
|
|
"(select usename from pg_user where oprowner = usesysid) as usename "
|
|
"from pg_operator");
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getOperators(): SELECT failed. Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
*numOprs = ntups;
|
|
|
|
oprinfo = (OprInfo *) malloc(ntups * sizeof(OprInfo));
|
|
|
|
i_oid = PQfnumber(res, "oid");
|
|
i_oprname = PQfnumber(res, "oprname");
|
|
i_oprkind = PQfnumber(res, "oprkind");
|
|
i_oprcode = PQfnumber(res, "oprcode");
|
|
i_oprleft = PQfnumber(res, "oprleft");
|
|
i_oprright = PQfnumber(res, "oprright");
|
|
i_oprcom = PQfnumber(res, "oprcom");
|
|
i_oprnegate = PQfnumber(res, "oprnegate");
|
|
i_oprrest = PQfnumber(res, "oprrest");
|
|
i_oprjoin = PQfnumber(res, "oprjoin");
|
|
i_oprcanhash = PQfnumber(res, "oprcanhash");
|
|
i_oprlsortop = PQfnumber(res, "oprlsortop");
|
|
i_oprrsortop = PQfnumber(res, "oprrsortop");
|
|
i_usename = PQfnumber(res, "usename");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
oprinfo[i].oid = strdup(PQgetvalue(res, i, i_oid));
|
|
oprinfo[i].oprname = strdup(PQgetvalue(res, i, i_oprname));
|
|
oprinfo[i].oprkind = strdup(PQgetvalue(res, i, i_oprkind));
|
|
oprinfo[i].oprcode = strdup(PQgetvalue(res, i, i_oprcode));
|
|
oprinfo[i].oprleft = strdup(PQgetvalue(res, i, i_oprleft));
|
|
oprinfo[i].oprright = strdup(PQgetvalue(res, i, i_oprright));
|
|
oprinfo[i].oprcom = strdup(PQgetvalue(res, i, i_oprcom));
|
|
oprinfo[i].oprnegate = strdup(PQgetvalue(res, i, i_oprnegate));
|
|
oprinfo[i].oprrest = strdup(PQgetvalue(res, i, i_oprrest));
|
|
oprinfo[i].oprjoin = strdup(PQgetvalue(res, i, i_oprjoin));
|
|
oprinfo[i].oprcanhash = strdup(PQgetvalue(res, i, i_oprcanhash));
|
|
oprinfo[i].oprlsortop = strdup(PQgetvalue(res, i, i_oprlsortop));
|
|
oprinfo[i].oprrsortop = strdup(PQgetvalue(res, i, i_oprrsortop));
|
|
oprinfo[i].usename = strdup(PQgetvalue(res, i, i_usename));
|
|
|
|
if (strlen(oprinfo[i].usename) == 0)
|
|
fprintf(stderr, "WARNING: owner of operator '%s' appears to be invalid\n",
|
|
oprinfo[i].oprname);
|
|
|
|
}
|
|
|
|
PQclear(res);
|
|
|
|
return oprinfo;
|
|
}
|
|
|
|
void
|
|
clearTypeInfo(TypeInfo *tp, int numTypes)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < numTypes; ++i)
|
|
{
|
|
if (tp[i].oid)
|
|
free(tp[i].oid);
|
|
if (tp[i].typowner)
|
|
free(tp[i].typowner);
|
|
if (tp[i].typname)
|
|
free(tp[i].typname);
|
|
if (tp[i].typlen)
|
|
free(tp[i].typlen);
|
|
if (tp[i].typprtlen)
|
|
free(tp[i].typprtlen);
|
|
if (tp[i].typinput)
|
|
free(tp[i].typinput);
|
|
if (tp[i].typoutput)
|
|
free(tp[i].typoutput);
|
|
if (tp[i].typreceive)
|
|
free(tp[i].typreceive);
|
|
if (tp[i].typsend)
|
|
free(tp[i].typsend);
|
|
if (tp[i].typelem)
|
|
free(tp[i].typelem);
|
|
if (tp[i].typdelim)
|
|
free(tp[i].typdelim);
|
|
if (tp[i].typdefault)
|
|
free(tp[i].typdefault);
|
|
if (tp[i].typrelid)
|
|
free(tp[i].typrelid);
|
|
if (tp[i].usename)
|
|
free(tp[i].usename);
|
|
}
|
|
free(tp);
|
|
}
|
|
|
|
void
|
|
clearFuncInfo(FuncInfo *fun, int numFuncs)
|
|
{
|
|
int i,
|
|
a;
|
|
|
|
if (!fun)
|
|
return;
|
|
for (i = 0; i < numFuncs; ++i)
|
|
{
|
|
if (fun[i].oid)
|
|
free(fun[i].oid);
|
|
if (fun[i].proname)
|
|
free(fun[i].proname);
|
|
if (fun[i].usename)
|
|
free(fun[i].usename);
|
|
for (a = 0; a < FUNC_MAX_ARGS; ++a)
|
|
if (fun[i].argtypes[a])
|
|
free(fun[i].argtypes[a]);
|
|
if (fun[i].prorettype)
|
|
free(fun[i].prorettype);
|
|
if (fun[i].prosrc)
|
|
free(fun[i].prosrc);
|
|
if (fun[i].probin)
|
|
free(fun[i].probin);
|
|
}
|
|
free(fun);
|
|
}
|
|
|
|
static void
|
|
clearTableInfo(TableInfo *tblinfo, int numTables)
|
|
{
|
|
int i,
|
|
j;
|
|
|
|
for (i = 0; i < numTables; ++i)
|
|
{
|
|
|
|
if (tblinfo[i].oid)
|
|
free(tblinfo[i].oid);
|
|
if (tblinfo[i].relacl)
|
|
free(tblinfo[i].relacl);
|
|
if (tblinfo[i].usename)
|
|
free(tblinfo[i].usename);
|
|
|
|
if (tblinfo[i].relname)
|
|
free(tblinfo[i].relname);
|
|
|
|
if (tblinfo[i].sequence)
|
|
continue;
|
|
|
|
/* Process Attributes */
|
|
for (j = 0; j < tblinfo[i].numatts; j++)
|
|
{
|
|
if (tblinfo[i].attoids[j])
|
|
free(tblinfo[i].attoids[j]);
|
|
if (tblinfo[i].attnames[j])
|
|
free(tblinfo[i].attnames[j]);
|
|
if (tblinfo[i].typnames[j])
|
|
free(tblinfo[i].typnames[j]);
|
|
}
|
|
|
|
if (tblinfo[i].triggers) {
|
|
for (j = 0; j < tblinfo[i].ntrig ; j++)
|
|
{
|
|
if (tblinfo[i].triggers[j].tgsrc)
|
|
free(tblinfo[i].triggers[j].tgsrc);
|
|
if (tblinfo[i].triggers[j].oid)
|
|
free(tblinfo[i].triggers[j].oid);
|
|
if (tblinfo[i].triggers[j].tgname)
|
|
free(tblinfo[i].triggers[j].tgname);
|
|
if (tblinfo[i].triggers[j].tgdel)
|
|
free(tblinfo[i].triggers[j].tgdel);
|
|
}
|
|
free(tblinfo[i].triggers);
|
|
}
|
|
|
|
if (tblinfo[i].atttypmod)
|
|
free((int *) tblinfo[i].atttypmod);
|
|
if (tblinfo[i].inhAttrs)
|
|
free((int *) tblinfo[i].inhAttrs);
|
|
if (tblinfo[i].attnames)
|
|
free(tblinfo[i].attnames);
|
|
if (tblinfo[i].atttypedefns)
|
|
free(tblinfo[i].atttypedefns);
|
|
if (tblinfo[i].typnames)
|
|
free(tblinfo[i].typnames);
|
|
if (tblinfo[i].notnull)
|
|
free(tblinfo[i].notnull);
|
|
if (tblinfo[i].primary_key_name)
|
|
free(tblinfo[i].primary_key_name);
|
|
}
|
|
free(tblinfo);
|
|
}
|
|
|
|
void
|
|
clearInhInfo(InhInfo *inh, int numInherits)
|
|
{
|
|
int i;
|
|
|
|
if (!inh)
|
|
return;
|
|
for (i = 0; i < numInherits; ++i)
|
|
{
|
|
if (inh[i].inhrelid)
|
|
free(inh[i].inhrelid);
|
|
if (inh[i].inhparent)
|
|
free(inh[i].inhparent);
|
|
}
|
|
free(inh);
|
|
}
|
|
|
|
void
|
|
clearOprInfo(OprInfo *opr, int numOprs)
|
|
{
|
|
int i;
|
|
|
|
if (!opr)
|
|
return;
|
|
for (i = 0; i < numOprs; ++i)
|
|
{
|
|
if (opr[i].oid)
|
|
free(opr[i].oid);
|
|
if (opr[i].oprname)
|
|
free(opr[i].oprname);
|
|
if (opr[i].oprkind)
|
|
free(opr[i].oprkind);
|
|
if (opr[i].oprcode)
|
|
free(opr[i].oprcode);
|
|
if (opr[i].oprleft)
|
|
free(opr[i].oprleft);
|
|
if (opr[i].oprright)
|
|
free(opr[i].oprright);
|
|
if (opr[i].oprcom)
|
|
free(opr[i].oprcom);
|
|
if (opr[i].oprnegate)
|
|
free(opr[i].oprnegate);
|
|
if (opr[i].oprrest)
|
|
free(opr[i].oprrest);
|
|
if (opr[i].oprjoin)
|
|
free(opr[i].oprjoin);
|
|
if (opr[i].oprcanhash)
|
|
free(opr[i].oprcanhash);
|
|
if (opr[i].oprlsortop)
|
|
free(opr[i].oprlsortop);
|
|
if (opr[i].oprrsortop)
|
|
free(opr[i].oprrsortop);
|
|
if (opr[i].usename)
|
|
free(opr[i].usename);
|
|
}
|
|
free(opr);
|
|
}
|
|
|
|
void
|
|
clearIndInfo(IndInfo *ind, int numIndices)
|
|
{
|
|
int i,
|
|
a;
|
|
|
|
if (!ind)
|
|
return;
|
|
for (i = 0; i < numIndices; ++i)
|
|
{
|
|
if (ind[i].indoid)
|
|
free(ind[i].indoid);
|
|
if (ind[i].indexrelname)
|
|
free(ind[i].indexrelname);
|
|
if (ind[i].indrelname)
|
|
free(ind[i].indrelname);
|
|
if (ind[i].indamname)
|
|
free(ind[i].indamname);
|
|
if (ind[i].indproc)
|
|
free(ind[i].indproc);
|
|
if (ind[i].indisunique)
|
|
free(ind[i].indisunique);
|
|
if (ind[i].indisprimary)
|
|
free(ind[i].indisprimary);
|
|
for (a = 0; a < INDEX_MAX_KEYS; ++a)
|
|
{
|
|
if (ind[i].indkey[a])
|
|
free(ind[i].indkey[a]);
|
|
if (ind[i].indclass[a])
|
|
free(ind[i].indclass[a]);
|
|
}
|
|
}
|
|
free(ind);
|
|
}
|
|
|
|
void
|
|
clearAggInfo(AggInfo *agginfo, int numArgs)
|
|
{
|
|
int i;
|
|
|
|
if (!agginfo)
|
|
return;
|
|
for (i = 0; i < numArgs; ++i)
|
|
{
|
|
if (agginfo[i].oid)
|
|
free(agginfo[i].oid);
|
|
if (agginfo[i].aggname)
|
|
free(agginfo[i].aggname);
|
|
if (agginfo[i].aggtransfn)
|
|
free(agginfo[i].aggtransfn);
|
|
if (agginfo[i].aggfinalfn)
|
|
free(agginfo[i].aggfinalfn);
|
|
if (agginfo[i].aggtranstype)
|
|
free(agginfo[i].aggtranstype);
|
|
if (agginfo[i].aggbasetype)
|
|
free(agginfo[i].aggbasetype);
|
|
if (agginfo[i].agginitval)
|
|
free(agginfo[i].agginitval);
|
|
if (agginfo[i].usename)
|
|
free(agginfo[i].usename);
|
|
}
|
|
free(agginfo);
|
|
}
|
|
|
|
/*
|
|
* getAggregates:
|
|
* read all the user-defined aggregates in the system catalogs and
|
|
* return them in the AggInfo* structure
|
|
*
|
|
* numAggs is set to the number of aggregates read in
|
|
*
|
|
*
|
|
*/
|
|
AggInfo *
|
|
getAggregates(int *numAggs)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
AggInfo *agginfo;
|
|
|
|
int i_oid;
|
|
int i_aggname;
|
|
int i_aggtransfn;
|
|
int i_aggfinalfn;
|
|
int i_aggtranstype;
|
|
int i_aggbasetype;
|
|
int i_agginitval;
|
|
int i_usename;
|
|
|
|
/* find all user-defined aggregates */
|
|
|
|
appendPQExpBuffer(query,
|
|
"SELECT pg_aggregate.oid, aggname, aggtransfn, "
|
|
"aggfinalfn, aggtranstype, aggbasetype, "
|
|
"agginitval, "
|
|
"(select usename from pg_user where aggowner = usesysid) as usename "
|
|
"from pg_aggregate" );
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getAggregates(): SELECT failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
*numAggs = ntups;
|
|
|
|
agginfo = (AggInfo *) malloc(ntups * sizeof(AggInfo));
|
|
|
|
i_oid = PQfnumber(res, "oid");
|
|
i_aggname = PQfnumber(res, "aggname");
|
|
i_aggtransfn = PQfnumber(res, "aggtransfn");
|
|
i_aggfinalfn = PQfnumber(res, "aggfinalfn");
|
|
i_aggtranstype = PQfnumber(res, "aggtranstype");
|
|
i_aggbasetype = PQfnumber(res, "aggbasetype");
|
|
i_agginitval = PQfnumber(res, "agginitval");
|
|
i_usename = PQfnumber(res, "usename");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
agginfo[i].oid = strdup(PQgetvalue(res, i, i_oid));
|
|
agginfo[i].aggname = strdup(PQgetvalue(res, i, i_aggname));
|
|
agginfo[i].aggtransfn = strdup(PQgetvalue(res, i, i_aggtransfn));
|
|
agginfo[i].aggfinalfn = strdup(PQgetvalue(res, i, i_aggfinalfn));
|
|
agginfo[i].aggtranstype = strdup(PQgetvalue(res, i, i_aggtranstype));
|
|
agginfo[i].aggbasetype = strdup(PQgetvalue(res, i, i_aggbasetype));
|
|
agginfo[i].agginitval = strdup(PQgetvalue(res, i, i_agginitval));
|
|
agginfo[i].usename = strdup(PQgetvalue(res, i, i_usename));
|
|
if (strlen(agginfo[i].usename) == 0)
|
|
fprintf(stderr, "WARNING: owner of aggregate '%s' appears to be invalid\n",
|
|
agginfo[i].aggname);
|
|
|
|
}
|
|
|
|
PQclear(res);
|
|
|
|
return agginfo;
|
|
}
|
|
|
|
/*
|
|
* getFuncs:
|
|
* read all the user-defined functions in the system catalogs and
|
|
* return them in the FuncInfo* structure
|
|
*
|
|
* numFuncs is set to the number of functions read in
|
|
*
|
|
*
|
|
*/
|
|
FuncInfo *
|
|
getFuncs(int *numFuncs)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
FuncInfo *finfo;
|
|
|
|
int i_oid;
|
|
int i_proname;
|
|
int i_prolang;
|
|
int i_pronargs;
|
|
int i_proargtypes;
|
|
int i_prorettype;
|
|
int i_proretset;
|
|
int i_prosrc;
|
|
int i_probin;
|
|
int i_iscachable;
|
|
int i_isstrict;
|
|
int i_usename;
|
|
|
|
/* find all user-defined funcs */
|
|
|
|
appendPQExpBuffer(query,
|
|
"SELECT pg_proc.oid, proname, prolang, pronargs, prorettype, "
|
|
"proretset, proargtypes, prosrc, probin, "
|
|
"(select usename from pg_user where proowner = usesysid) as usename, "
|
|
"proiscachable, proisstrict "
|
|
"from pg_proc "
|
|
"where pg_proc.oid > '%u'::oid",
|
|
g_last_builtin_oid);
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getFuncs(): SELECT failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
*numFuncs = ntups;
|
|
|
|
finfo = (FuncInfo *) malloc(ntups * sizeof(FuncInfo));
|
|
|
|
memset((char *) finfo, 0, ntups * sizeof(FuncInfo));
|
|
|
|
i_oid = PQfnumber(res, "oid");
|
|
i_proname = PQfnumber(res, "proname");
|
|
i_prolang = PQfnumber(res, "prolang");
|
|
i_pronargs = PQfnumber(res, "pronargs");
|
|
i_proargtypes = PQfnumber(res, "proargtypes");
|
|
i_prorettype = PQfnumber(res, "prorettype");
|
|
i_proretset = PQfnumber(res, "proretset");
|
|
i_prosrc = PQfnumber(res, "prosrc");
|
|
i_probin = PQfnumber(res, "probin");
|
|
i_iscachable = PQfnumber(res, "proiscachable");
|
|
i_isstrict = PQfnumber(res, "proisstrict");
|
|
i_usename = PQfnumber(res, "usename");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
finfo[i].oid = strdup(PQgetvalue(res, i, i_oid));
|
|
finfo[i].proname = strdup(PQgetvalue(res, i, i_proname));
|
|
|
|
finfo[i].prosrc = strdup(PQgetvalue(res, i, i_prosrc));
|
|
finfo[i].probin = strdup(PQgetvalue(res, i, i_probin));
|
|
|
|
finfo[i].prorettype = strdup(PQgetvalue(res, i, i_prorettype));
|
|
finfo[i].retset = (strcmp(PQgetvalue(res, i, i_proretset), "t") == 0);
|
|
finfo[i].nargs = atoi(PQgetvalue(res, i, i_pronargs));
|
|
finfo[i].lang = atoi(PQgetvalue(res, i, i_prolang));
|
|
finfo[i].usename = strdup(PQgetvalue(res, i, i_usename));
|
|
finfo[i].iscachable = (strcmp(PQgetvalue(res, i, i_iscachable),"t") == 0);
|
|
finfo[i].isstrict = (strcmp(PQgetvalue(res, i, i_isstrict),"t") == 0);
|
|
|
|
if (strlen(finfo[i].usename) == 0)
|
|
fprintf(stderr, "WARNING: owner of function '%s' appears to be invalid\n",
|
|
finfo[i].proname);
|
|
|
|
if (finfo[i].nargs < 0 || finfo[i].nargs > FUNC_MAX_ARGS)
|
|
{
|
|
fprintf(stderr, "failed sanity check: %s has %d args\n",
|
|
finfo[i].proname, finfo[i].nargs);
|
|
exit(1);
|
|
}
|
|
parseNumericArray(PQgetvalue(res, i, i_proargtypes),
|
|
finfo[i].argtypes,
|
|
finfo[i].nargs);
|
|
finfo[i].dumped = 0;
|
|
}
|
|
|
|
PQclear(res);
|
|
|
|
return finfo;
|
|
|
|
}
|
|
|
|
/*
|
|
* getTables
|
|
* read all the user-defined tables (no indices, no catalogs)
|
|
* in the system catalogs return them in the TableInfo* structure
|
|
*
|
|
* numTables is set to the number of tables read in
|
|
*
|
|
*
|
|
*/
|
|
TableInfo *
|
|
getTables(int *numTables, FuncInfo *finfo, int numFuncs)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
PQExpBuffer delqry = createPQExpBuffer();
|
|
TableInfo *tblinfo;
|
|
|
|
int i_reloid;
|
|
int i_relname;
|
|
int i_relkind;
|
|
int i_relacl;
|
|
int i_usename;
|
|
int i_relchecks;
|
|
int i_reltriggers;
|
|
int i_relhasindex;
|
|
|
|
char relkindview[2];
|
|
|
|
relkindview[0] = RELKIND_VIEW;
|
|
relkindview[1] = '\0';
|
|
|
|
/*
|
|
* find all the user-defined tables (no indices and no catalogs),
|
|
* ordering by oid is important so that we always process the parent
|
|
* tables before the child tables when traversing the tblinfo*
|
|
*
|
|
* we ignore tables that are not type 'r' (ordinary relation) or 'S'
|
|
* (sequence) or 'v' (view).
|
|
*/
|
|
|
|
appendPQExpBuffer(query,
|
|
"SELECT pg_class.oid, relname, relkind, relacl, "
|
|
"(select usename from pg_user where relowner = usesysid) as usename, "
|
|
"relchecks, reltriggers, relhasindex "
|
|
"from pg_class "
|
|
"where relname !~ '^pg_' "
|
|
"and relkind in ('%c', '%c', '%c') "
|
|
"order by oid",
|
|
RELKIND_RELATION, RELKIND_SEQUENCE, RELKIND_VIEW);
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
*numTables = ntups;
|
|
|
|
tblinfo = (TableInfo *) malloc(ntups * sizeof(TableInfo));
|
|
|
|
i_reloid = PQfnumber(res, "oid");
|
|
i_relname = PQfnumber(res, "relname");
|
|
i_relkind = PQfnumber(res, "relkind");
|
|
i_relacl = PQfnumber(res, "relacl");
|
|
i_usename = PQfnumber(res, "usename");
|
|
i_relchecks = PQfnumber(res, "relchecks");
|
|
i_reltriggers = PQfnumber(res, "reltriggers");
|
|
i_relhasindex = PQfnumber(res, "relhasindex");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
tblinfo[i].oid = strdup(PQgetvalue(res, i, i_reloid));
|
|
tblinfo[i].relname = strdup(PQgetvalue(res, i, i_relname));
|
|
tblinfo[i].relacl = strdup(PQgetvalue(res, i, i_relacl));
|
|
tblinfo[i].sequence = (strcmp(PQgetvalue(res, i, i_relkind), "S") == 0);
|
|
tblinfo[i].usename = strdup(PQgetvalue(res, i, i_usename));
|
|
tblinfo[i].ncheck = atoi(PQgetvalue(res, i, i_relchecks));
|
|
tblinfo[i].ntrig = atoi(PQgetvalue(res, i, i_reltriggers));
|
|
|
|
if (strlen(tblinfo[i].usename) == 0)
|
|
fprintf(stderr, "WARNING: owner of table '%s' appears to be invalid\n",
|
|
tblinfo[i].relname);
|
|
|
|
/* Get view definition */
|
|
if (strcmp(PQgetvalue(res, i, i_relkind), relkindview) == 0)
|
|
{
|
|
PGresult *res2;
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT pg_get_viewdef(");
|
|
formatStringLiteral(query, tblinfo[i].relname);
|
|
appendPQExpBuffer(query, ") as viewdef");
|
|
res2 = PQexec(g_conn, query->data);
|
|
if (!res2 || PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for VIEW DEFINITION) failed. "
|
|
"Explanation from backend: %s",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
if (PQntuples(res2) != 1)
|
|
{
|
|
if (PQntuples(res2) < 1)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for VIEW %s) returned no definitions",
|
|
tblinfo[i].relname);
|
|
} else {
|
|
fprintf(stderr, "getTables(): SELECT (for VIEW %s) returned more than 1 definition",
|
|
tblinfo[i].relname);
|
|
}
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
tblinfo[i].viewdef = strdup(PQgetvalue(res2, 0, 0));
|
|
|
|
if (strlen(tblinfo[i].viewdef) == 0)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for VIEW %s) returned empty definition",
|
|
tblinfo[i].relname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
}
|
|
else
|
|
tblinfo[i].viewdef = NULL;
|
|
|
|
/*
|
|
* Exclude inherited CHECKs from CHECK constraints total. If a
|
|
* constraint matches by name and condition with a constraint
|
|
* belonging to a parent class, we assume it was inherited.
|
|
*/
|
|
if (tblinfo[i].ncheck > 0)
|
|
{
|
|
PGresult *res2;
|
|
int ntups2;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s excluding inherited CHECK constraints "
|
|
"for relation: '%s' %s\n",
|
|
g_comment_start,
|
|
tblinfo[i].relname,
|
|
g_comment_end);
|
|
|
|
/* XXXX: Use LOJ maybe - need to compare with subsequent query for non-inherited */
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT rcname from pg_relcheck, pg_inherits as i "
|
|
"where rcrelid = '%s'::oid "
|
|
" and rcrelid = i.inhrelid"
|
|
" and exists "
|
|
" (select * from pg_relcheck as c "
|
|
" where c.rcname = pg_relcheck.rcname "
|
|
" and c.rcsrc = pg_relcheck.rcsrc "
|
|
" and c.rcrelid = i.inhparent) ",
|
|
tblinfo[i].oid);
|
|
res2 = PQexec(g_conn, query->data);
|
|
if (!res2 ||
|
|
PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for inherited CHECK) failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
ntups2 = PQntuples(res2);
|
|
tblinfo[i].ncheck -= ntups2;
|
|
if (tblinfo[i].ncheck < 0)
|
|
{
|
|
fprintf(stderr, "getTables(): found more inherited CHECKs than total for "
|
|
"relation %s\n",
|
|
tblinfo[i].relname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
PQclear(res2);
|
|
}
|
|
|
|
/* Get non-inherited CHECK constraints, if any */
|
|
if (tblinfo[i].ncheck > 0)
|
|
{
|
|
PGresult *res2;
|
|
int i_rcname,
|
|
i_rcsrc;
|
|
int ntups2;
|
|
int i2;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s finding CHECK constraints for relation: '%s' %s\n",
|
|
g_comment_start,
|
|
tblinfo[i].relname,
|
|
g_comment_end);
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT rcname, rcsrc from pg_relcheck "
|
|
"where rcrelid = '%s'::oid "
|
|
" and not exists "
|
|
" (select * from pg_relcheck as c, pg_inherits as i "
|
|
" where i.inhrelid = pg_relcheck.rcrelid "
|
|
" and c.rcname = pg_relcheck.rcname "
|
|
" and c.rcsrc = pg_relcheck.rcsrc "
|
|
" and c.rcrelid = i.inhparent) ",
|
|
tblinfo[i].oid);
|
|
res2 = PQexec(g_conn, query->data);
|
|
if (!res2 ||
|
|
PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for CHECK) failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
ntups2 = PQntuples(res2);
|
|
if (ntups2 != tblinfo[i].ncheck)
|
|
{
|
|
fprintf(stderr, "getTables(): relation '%s': %d CHECKs were expected, but got %d\n",
|
|
tblinfo[i].relname, tblinfo[i].ncheck, ntups2);
|
|
exit_nicely(g_conn);
|
|
}
|
|
i_rcname = PQfnumber(res2, "rcname");
|
|
i_rcsrc = PQfnumber(res2, "rcsrc");
|
|
tblinfo[i].check_expr = (char **) malloc(ntups2 * sizeof(char *));
|
|
for (i2 = 0; i2 < ntups2; i2++)
|
|
{
|
|
const char *name = PQgetvalue(res2, i2, i_rcname);
|
|
const char *expr = PQgetvalue(res2, i2, i_rcsrc);
|
|
|
|
resetPQExpBuffer(query);
|
|
if (name[0] != '$')
|
|
{
|
|
appendPQExpBuffer(query, "CONSTRAINT %s ",
|
|
fmtId(name, force_quotes));
|
|
}
|
|
appendPQExpBuffer(query, "CHECK (%s)", expr);
|
|
tblinfo[i].check_expr[i2] = strdup(query->data);
|
|
}
|
|
PQclear(res2);
|
|
}
|
|
else
|
|
tblinfo[i].check_expr = NULL;
|
|
|
|
/* Get primary key */
|
|
if (strcmp(PQgetvalue(res, i, i_relhasindex), "t") == 0)
|
|
{
|
|
PGresult *res2;
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query,
|
|
"SELECT Oid FROM pg_index i WHERE i.indisprimary AND i.indrelid = %s ",
|
|
tblinfo[i].oid);
|
|
res2 = PQexec(g_conn, query->data);
|
|
if (!res2 || PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for PRIMARY KEY) failed on table %s. Explanation from backend: %s\n",
|
|
tblinfo[i].relname, PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
if (PQntuples(res2) > 1) {
|
|
fprintf(stderr, "getTables(): SELECT (for PRIMARY KEY) produced more than one row on table %s.\n",
|
|
tblinfo[i].relname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
if (PQntuples(res2) == 1) {
|
|
tblinfo[i].pkIndexOid = strdup(PQgetvalue(res2, 0, 0));
|
|
} else {
|
|
tblinfo[i].pkIndexOid = NULL;
|
|
}
|
|
|
|
}
|
|
else
|
|
tblinfo[i].pkIndexOid = NULL;
|
|
|
|
/* Get primary key name (if primary key exist) */
|
|
if (tblinfo[i].pkIndexOid != NULL)
|
|
{
|
|
PGresult *res2;
|
|
int n;
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query,
|
|
"SELECT c.relname "
|
|
"FROM pg_index i LEFT OUTER JOIN pg_class c ON c.oid = i.indexrelid "
|
|
"WHERE i.indrelid = %s"
|
|
"AND i.indisprimary ",
|
|
tblinfo[i].oid);
|
|
res2 = PQexec(g_conn, query->data);
|
|
if (!res2 || PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for PRIMARY KEY NAME) failed for table %s. Explanation from backend: %s",
|
|
tblinfo[i].relname, PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
n = PQntuples(res2);
|
|
if (n != 1)
|
|
{
|
|
fprintf(stderr,
|
|
"getTables(): SELECT (for PRIMARY KEY NAME) failed for table %s. "
|
|
"This is impossible but object with OID == %s have %d primary keys.\n",
|
|
tblinfo[i].relname,
|
|
tblinfo[i].oid,
|
|
n);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Sanity check on LOJ */
|
|
if (PQgetisnull(res2, 0, 0))
|
|
{
|
|
fprintf(stderr,"getTables(): SELECT (for PRIMARY KEY NAME) on table %s returned NULL value.\n",
|
|
tblinfo[i].relname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
tblinfo[i].primary_key_name =
|
|
strdup(fmtId(PQgetvalue(res2, 0, 0), force_quotes));
|
|
if (tblinfo[i].primary_key_name == NULL)
|
|
{
|
|
perror("strdup");
|
|
exit(1);
|
|
}
|
|
}
|
|
else
|
|
tblinfo[i].primary_key_name = NULL;
|
|
|
|
/* Get Triggers */
|
|
if (tblinfo[i].ntrig > 0)
|
|
{
|
|
PGresult *res2;
|
|
int i_tgoid,
|
|
i_tgname,
|
|
i_tgfoid,
|
|
i_tgtype,
|
|
i_tgnargs,
|
|
i_tgargs,
|
|
i_tgisconstraint,
|
|
i_tgconstrname,
|
|
i_tgdeferrable,
|
|
i_tginitdeferred;
|
|
int ntups2;
|
|
int i2;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s finding Triggers for relation: '%s' %s\n",
|
|
g_comment_start,
|
|
tblinfo[i].relname,
|
|
g_comment_end);
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT tgname, tgfoid, tgtype, tgnargs, tgargs, tgisconstraint, tgconstrname, tgdeferrable, tginitdeferred, oid "
|
|
"from pg_trigger "
|
|
"where tgrelid = '%s'::oid ",
|
|
tblinfo[i].oid);
|
|
res2 = PQexec(g_conn, query->data);
|
|
if (!res2 ||
|
|
PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (for TRIGGER) failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
ntups2 = PQntuples(res2);
|
|
if (ntups2 != tblinfo[i].ntrig)
|
|
{
|
|
fprintf(stderr, "getTables(): relation '%s': %d Triggers were expected, but got %d\n",
|
|
tblinfo[i].relname, tblinfo[i].ntrig, ntups2);
|
|
exit_nicely(g_conn);
|
|
}
|
|
i_tgname = PQfnumber(res2, "tgname");
|
|
i_tgfoid = PQfnumber(res2, "tgfoid");
|
|
i_tgtype = PQfnumber(res2, "tgtype");
|
|
i_tgnargs = PQfnumber(res2, "tgnargs");
|
|
i_tgargs = PQfnumber(res2, "tgargs");
|
|
i_tgoid = PQfnumber(res2, "oid");
|
|
i_tgisconstraint = PQfnumber(res2, "tgisconstraint");
|
|
i_tgconstrname = PQfnumber(res2, "tgconstrname");
|
|
i_tgdeferrable = PQfnumber(res2, "tgdeferrable");
|
|
i_tginitdeferred = PQfnumber(res2, "tginitdeferred");
|
|
|
|
tblinfo[i].triggers = (TrigInfo*) malloc(ntups2 * sizeof(TrigInfo));
|
|
resetPQExpBuffer(query);
|
|
for (i2 = 0; i2 < ntups2; i2++)
|
|
{
|
|
const char *tgfuncoid = PQgetvalue(res2, i2, i_tgfoid);
|
|
char *tgfunc = NULL;
|
|
int2 tgtype = atoi(PQgetvalue(res2, i2, i_tgtype));
|
|
int tgnargs = atoi(PQgetvalue(res2, i2, i_tgnargs));
|
|
const char *tgargs = PQgetvalue(res2, i2, i_tgargs);
|
|
int tgisconstraint;
|
|
int tgdeferrable;
|
|
int tginitdeferred;
|
|
const char *p;
|
|
int findx;
|
|
|
|
if (strcmp(PQgetvalue(res2, i2, i_tgisconstraint), "f") == 0)
|
|
tgisconstraint = 0;
|
|
else
|
|
tgisconstraint = 1;
|
|
|
|
if (strcmp(PQgetvalue(res2, i2, i_tgdeferrable), "f") == 0)
|
|
tgdeferrable = 0;
|
|
else
|
|
tgdeferrable = 1;
|
|
|
|
if (strcmp(PQgetvalue(res2, i2, i_tginitdeferred), "f") == 0)
|
|
tginitdeferred = 0;
|
|
else
|
|
tginitdeferred = 1;
|
|
|
|
for (findx = 0; findx < numFuncs; findx++)
|
|
{
|
|
if (strcmp(finfo[findx].oid, tgfuncoid) == 0 &&
|
|
finfo[findx].nargs == 0 &&
|
|
strcmp(finfo[findx].prorettype, "0") == 0)
|
|
break;
|
|
}
|
|
|
|
if (findx == numFuncs)
|
|
{
|
|
PGresult *r;
|
|
int numFuncs;
|
|
|
|
/*
|
|
* the funcname is an oid which we use to find the
|
|
* name of the pg_proc. We need to do this because
|
|
* getFuncs() only reads in the user-defined funcs not
|
|
* all the funcs. We might not find what we want by
|
|
* looking in FuncInfo*
|
|
*/
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query,
|
|
"SELECT proname from pg_proc "
|
|
"where pg_proc.oid = '%s'::oid",
|
|
tgfuncoid);
|
|
|
|
r = PQexec(g_conn, query->data);
|
|
if (!r || PQresultStatus(r) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTables(): SELECT (funcname) failed for trigger %s. Explanation from backend: '%s'.\n",
|
|
PQgetvalue(res2, i2, i_tgname), PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Sanity: Check we got only one tuple */
|
|
numFuncs = PQntuples(r);
|
|
if (numFuncs != 1) {
|
|
fprintf(stderr, "getTables(): SELECT (funcname) for trigger %s returned %d tuples. Expected 1.\n",
|
|
PQgetvalue(res2, i2, i_tgname), numFuncs);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
tgfunc = strdup(PQgetvalue(r, 0, PQfnumber(r, "proname")));
|
|
PQclear(r);
|
|
}
|
|
else
|
|
tgfunc = strdup(finfo[findx].proname);
|
|
|
|
appendPQExpBuffer(delqry, "DROP TRIGGER %s ",
|
|
fmtId(PQgetvalue(res2, i2, i_tgname),
|
|
force_quotes));
|
|
appendPQExpBuffer(delqry, "ON %s;\n",
|
|
fmtId(tblinfo[i].relname, force_quotes));
|
|
|
|
resetPQExpBuffer(query);
|
|
if (tgisconstraint)
|
|
{
|
|
appendPQExpBuffer(query, "CREATE CONSTRAINT TRIGGER ");
|
|
appendPQExpBuffer(query, fmtId(PQgetvalue(res2, i2, i_tgconstrname), force_quotes));
|
|
}
|
|
else
|
|
{
|
|
appendPQExpBuffer(query, "CREATE TRIGGER ");
|
|
appendPQExpBuffer(query, fmtId(PQgetvalue(res2, i2, i_tgname), force_quotes));
|
|
}
|
|
appendPQExpBufferChar(query, ' ');
|
|
/* Trigger type */
|
|
findx = 0;
|
|
if (TRIGGER_FOR_BEFORE(tgtype))
|
|
appendPQExpBuffer(query, "BEFORE");
|
|
else
|
|
appendPQExpBuffer(query, "AFTER");
|
|
if (TRIGGER_FOR_INSERT(tgtype))
|
|
{
|
|
appendPQExpBuffer(query, " INSERT");
|
|
findx++;
|
|
}
|
|
if (TRIGGER_FOR_DELETE(tgtype))
|
|
{
|
|
if (findx > 0)
|
|
appendPQExpBuffer(query, " OR DELETE");
|
|
else
|
|
appendPQExpBuffer(query, " DELETE");
|
|
findx++;
|
|
}
|
|
if (TRIGGER_FOR_UPDATE(tgtype))
|
|
{
|
|
if (findx > 0)
|
|
appendPQExpBuffer(query, " OR UPDATE");
|
|
else
|
|
appendPQExpBuffer(query, " UPDATE");
|
|
}
|
|
appendPQExpBuffer(query, " ON %s ", fmtId(tblinfo[i].relname, force_quotes));
|
|
|
|
if (tgisconstraint)
|
|
{
|
|
if (!tgdeferrable)
|
|
appendPQExpBuffer(query, " NOT");
|
|
appendPQExpBuffer(query, " DEFERRABLE INITIALLY ");
|
|
if (tginitdeferred)
|
|
appendPQExpBuffer(query, "DEFERRED");
|
|
else
|
|
appendPQExpBuffer(query, "IMMEDIATE");
|
|
|
|
}
|
|
|
|
appendPQExpBuffer(query, " FOR EACH ROW");
|
|
appendPQExpBuffer(query, " EXECUTE PROCEDURE %s (",
|
|
fmtId(tgfunc, force_quotes));
|
|
for (findx = 0; findx < tgnargs; findx++)
|
|
{
|
|
const char *s;
|
|
|
|
for (p = tgargs;;)
|
|
{
|
|
p = strchr(p, '\\');
|
|
if (p == NULL)
|
|
{
|
|
fprintf(stderr, "getTables(): relation '%s': bad argument "
|
|
"string (%s) for trigger '%s'\n",
|
|
tblinfo[i].relname,
|
|
PQgetvalue(res2, i2, i_tgargs),
|
|
PQgetvalue(res2, i2, i_tgname));
|
|
exit_nicely(g_conn);
|
|
}
|
|
p++;
|
|
if (*p == '\\')
|
|
{
|
|
p++;
|
|
continue;
|
|
}
|
|
if (p[0] == '0' && p[1] == '0' && p[2] == '0')
|
|
break;
|
|
}
|
|
p--;
|
|
appendPQExpBufferChar(query, '\'');
|
|
for (s = tgargs; s < p;)
|
|
{
|
|
if (*s == '\'')
|
|
appendPQExpBufferChar(query, '\\');
|
|
appendPQExpBufferChar(query, *s++);
|
|
}
|
|
appendPQExpBufferChar(query, '\'');
|
|
appendPQExpBuffer(query, (findx < tgnargs - 1) ? ", " : "");
|
|
tgargs = p + 4;
|
|
}
|
|
appendPQExpBuffer(query, ");\n");
|
|
|
|
tblinfo[i].triggers[i2].tgsrc = strdup(query->data);
|
|
|
|
/*** Initialize trcomments and troids ***/
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "TRIGGER %s ",
|
|
fmtId(PQgetvalue(res2, i2, i_tgname), force_quotes));
|
|
appendPQExpBuffer(query, "ON %s",
|
|
fmtId(tblinfo[i].relname, force_quotes));
|
|
tblinfo[i].triggers[i2].tgcomment = strdup(query->data);
|
|
tblinfo[i].triggers[i2].oid = strdup(PQgetvalue(res2, i2, i_tgoid));
|
|
tblinfo[i].triggers[i2].tgname = strdup(fmtId(PQgetvalue(res2, i2, i_tgname),false));
|
|
tblinfo[i].triggers[i2].tgdel = strdup(delqry->data);
|
|
|
|
if (tgfunc)
|
|
free(tgfunc);
|
|
}
|
|
PQclear(res2);
|
|
}
|
|
else
|
|
{
|
|
tblinfo[i].triggers = NULL;
|
|
}
|
|
|
|
}
|
|
|
|
PQclear(res);
|
|
|
|
return tblinfo;
|
|
|
|
}
|
|
|
|
/*
|
|
* getInherits
|
|
* read all the inheritance information
|
|
* from the system catalogs return them in the InhInfo* structure
|
|
*
|
|
* numInherits is set to the number of tables read in
|
|
*
|
|
*
|
|
*/
|
|
InhInfo *
|
|
getInherits(int *numInherits)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
InhInfo *inhinfo;
|
|
|
|
int i_inhrelid;
|
|
int i_inhparent;
|
|
|
|
/* find all the inheritance information */
|
|
|
|
appendPQExpBuffer(query, "SELECT inhrelid, inhparent from pg_inherits");
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getInherits(): SELECT failed. Explanation from backend: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
*numInherits = ntups;
|
|
|
|
inhinfo = (InhInfo *) malloc(ntups * sizeof(InhInfo));
|
|
|
|
i_inhrelid = PQfnumber(res, "inhrelid");
|
|
i_inhparent = PQfnumber(res, "inhparent");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
inhinfo[i].inhrelid = strdup(PQgetvalue(res, i, i_inhrelid));
|
|
inhinfo[i].inhparent = strdup(PQgetvalue(res, i, i_inhparent));
|
|
}
|
|
|
|
PQclear(res);
|
|
return inhinfo;
|
|
}
|
|
|
|
/*
|
|
* getTableAttrs -
|
|
* for each table in tblinfo, read its attributes types and names
|
|
*
|
|
* this is implemented in a very inefficient way right now, looping
|
|
* through the tblinfo and doing a join per table to find the attrs and their
|
|
* types
|
|
*
|
|
* modifies tblinfo
|
|
*/
|
|
void
|
|
getTableAttrs(TableInfo *tblinfo, int numTables)
|
|
{
|
|
int i,
|
|
j;
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
int i_attname;
|
|
int i_typname;
|
|
int i_atttypmod;
|
|
int i_attnotnull;
|
|
int i_atthasdef;
|
|
int i_attoid;
|
|
int i_atttypedefn;
|
|
PGresult *res;
|
|
int ntups;
|
|
|
|
for (i = 0; i < numTables; i++)
|
|
{
|
|
|
|
if (tblinfo[i].sequence)
|
|
continue;
|
|
|
|
/* find all the user attributes and their types */
|
|
/* we must read the attribute names in attribute number order! */
|
|
|
|
/*
|
|
* because we will use the attnum to index into the attnames array
|
|
* later
|
|
*/
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s finding the attrs and types for table: '%s' %s\n",
|
|
g_comment_start,
|
|
tblinfo[i].relname,
|
|
g_comment_end);
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "SELECT a.oid as attoid, a.attnum, a.attname, t.typname, a.atttypmod, "
|
|
"a.attnotnull, a.atthasdef, format_type(a.atttypid, a.atttypmod) as atttypedefn "
|
|
"from pg_attribute a LEFT OUTER JOIN pg_type t ON a.atttypid = t.oid "
|
|
"where a.attrelid = '%s'::oid "
|
|
"and a.attnum > 0 order by attnum",
|
|
tblinfo[i].oid);
|
|
res = PQexec(g_conn, q->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTableAttrs(): SELECT failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
i_attoid = PQfnumber(res, "attoid");
|
|
i_attname = PQfnumber(res, "attname");
|
|
i_typname = PQfnumber(res, "typname");
|
|
i_atttypmod = PQfnumber(res, "atttypmod");
|
|
i_attnotnull = PQfnumber(res, "attnotnull");
|
|
i_atthasdef = PQfnumber(res, "atthasdef");
|
|
i_atttypedefn = PQfnumber(res, "atttypedefn");
|
|
|
|
tblinfo[i].numatts = ntups;
|
|
tblinfo[i].attoids = (char **) malloc(ntups * sizeof(char *));
|
|
tblinfo[i].attnames = (char **) malloc(ntups * sizeof(char *));
|
|
tblinfo[i].atttypedefns = (char **) malloc(ntups * sizeof(char *));
|
|
tblinfo[i].typnames = (char **) malloc(ntups * sizeof(char *));
|
|
tblinfo[i].atttypmod = (int *) malloc(ntups * sizeof(int));
|
|
tblinfo[i].inhAttrs = (int *) malloc(ntups * sizeof(int));
|
|
tblinfo[i].notnull = (bool *) malloc(ntups * sizeof(bool));
|
|
tblinfo[i].adef_expr = (char **) malloc(ntups * sizeof(char *));
|
|
tblinfo[i].parentRels = NULL;
|
|
tblinfo[i].numParents = 0;
|
|
for (j = 0; j < ntups; j++)
|
|
{
|
|
|
|
/* Sanity check on LOJ */
|
|
if (PQgetisnull(res, j, i_typname))
|
|
{
|
|
fprintf(stderr, "getTableAttrs(): SELECT produced NULL attribute type name for attr %d on table %s.\n",
|
|
j, tblinfo[i].relname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
tblinfo[i].attoids[j] = strdup(PQgetvalue(res, j, i_attoid));
|
|
tblinfo[i].attnames[j] = strdup(PQgetvalue(res, j, i_attname));
|
|
tblinfo[i].atttypedefns[j] = strdup(PQgetvalue(res, j, i_atttypedefn));
|
|
tblinfo[i].typnames[j] = strdup(PQgetvalue(res, j, i_typname));
|
|
tblinfo[i].atttypmod[j] = atoi(PQgetvalue(res, j, i_atttypmod));
|
|
tblinfo[i].inhAttrs[j] = 0; /* this flag is set in
|
|
* flagInhAttrs() */
|
|
tblinfo[i].notnull[j] = (PQgetvalue(res, j, i_attnotnull)[0] == 't') ? true : false;
|
|
if (PQgetvalue(res, j, i_atthasdef)[0] == 't')
|
|
{
|
|
PGresult *res2;
|
|
int numAttr;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s finding DEFAULT expression for attr: '%s' %s\n",
|
|
g_comment_start,
|
|
tblinfo[i].attnames[j],
|
|
g_comment_end);
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "SELECT adsrc from pg_attrdef "
|
|
"where adrelid = '%s'::oid and adnum = %d ",
|
|
tblinfo[i].oid, j + 1);
|
|
res2 = PQexec(g_conn, q->data);
|
|
if (!res2 ||
|
|
PQresultStatus(res2) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getTableAttrs(): SELECT (for DEFAULT) failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Sanity: Check we got only one tuple */
|
|
numAttr = PQntuples(res2);
|
|
if (numAttr != 1) {
|
|
fprintf(stderr, "getTableAttrs(): SELECT (for DEFAULT) for attr %s returned %d tuples. Expected 1.\n",
|
|
tblinfo[i].attnames[j], numAttr);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
tblinfo[i].adef_expr[j] = strdup(PQgetvalue(res2, 0, PQfnumber(res2, "adsrc")));
|
|
PQclear(res2);
|
|
}
|
|
else
|
|
tblinfo[i].adef_expr[j] = NULL;
|
|
}
|
|
PQclear(res);
|
|
}
|
|
}
|
|
|
|
|
|
/*
|
|
* getIndices
|
|
* read all the user-defined indices information
|
|
* from the system catalogs return them in the InhInfo* structure
|
|
*
|
|
* numIndices is set to the number of indices read in
|
|
*
|
|
*
|
|
*/
|
|
IndInfo *
|
|
getIndices(int *numIndices)
|
|
{
|
|
int i;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
PGresult *res;
|
|
int ntups;
|
|
IndInfo *indinfo;
|
|
|
|
int i_indexrelname;
|
|
int i_indrelname;
|
|
int i_indamname;
|
|
int i_indproc;
|
|
int i_indkey;
|
|
int i_indclass;
|
|
int i_indisunique;
|
|
int i_indoid;
|
|
int i_oid;
|
|
int i_indisprimary;
|
|
|
|
/*
|
|
* find all the user-defined indices. We do not handle partial
|
|
* indices.
|
|
*
|
|
* Notice we skip indices on system classes
|
|
*
|
|
* this is a 4-way join !!
|
|
*
|
|
* XXXX: Use LOJ
|
|
*/
|
|
|
|
appendPQExpBuffer(query,
|
|
"SELECT i.oid, t1.oid as indoid, t1.relname as indexrelname, t2.relname as indrelname, "
|
|
"i.indproc, i.indkey, i.indclass, "
|
|
"a.amname as indamname, i.indisunique, i.indisprimary "
|
|
"from pg_index i, pg_class t1, pg_class t2, pg_am a "
|
|
"WHERE t1.oid = i.indexrelid and t2.oid = i.indrelid "
|
|
"and t1.relam = a.oid and i.indexrelid > '%u'::oid "
|
|
"and t2.relname !~ '^pg_' ",
|
|
g_last_builtin_oid);
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "getIndices(): SELECT failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
ntups = PQntuples(res);
|
|
|
|
*numIndices = ntups;
|
|
|
|
indinfo = (IndInfo *) malloc(ntups * sizeof(IndInfo));
|
|
|
|
memset((char *) indinfo, 0, ntups * sizeof(IndInfo));
|
|
|
|
i_oid = PQfnumber(res, "oid");
|
|
i_indoid = PQfnumber(res, "indoid");
|
|
i_indexrelname = PQfnumber(res, "indexrelname");
|
|
i_indrelname = PQfnumber(res, "indrelname");
|
|
i_indamname = PQfnumber(res, "indamname");
|
|
i_indproc = PQfnumber(res, "indproc");
|
|
i_indkey = PQfnumber(res, "indkey");
|
|
i_indclass = PQfnumber(res, "indclass");
|
|
i_indisunique = PQfnumber(res, "indisunique");
|
|
i_indisprimary = PQfnumber(res, "indisprimary");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
indinfo[i].oid = strdup(PQgetvalue(res, i, i_oid));
|
|
indinfo[i].indoid = strdup(PQgetvalue(res, i, i_indoid));
|
|
indinfo[i].indexrelname = strdup(PQgetvalue(res, i, i_indexrelname));
|
|
indinfo[i].indrelname = strdup(PQgetvalue(res, i, i_indrelname));
|
|
indinfo[i].indamname = strdup(PQgetvalue(res, i, i_indamname));
|
|
indinfo[i].indproc = strdup(PQgetvalue(res, i, i_indproc));
|
|
parseNumericArray(PQgetvalue(res, i, i_indkey),
|
|
indinfo[i].indkey,
|
|
INDEX_MAX_KEYS);
|
|
parseNumericArray(PQgetvalue(res, i, i_indclass),
|
|
indinfo[i].indclass,
|
|
INDEX_MAX_KEYS);
|
|
indinfo[i].indisunique = strdup(PQgetvalue(res, i, i_indisunique));
|
|
indinfo[i].indisprimary = strdup(PQgetvalue(res, i, i_indisprimary));
|
|
}
|
|
PQclear(res);
|
|
return indinfo;
|
|
}
|
|
|
|
/*------------------------------------------------------------------
|
|
* dumpComments --
|
|
*
|
|
* This routine is used to dump any comments associated with the
|
|
* oid handed to this routine. The routine takes a constant character
|
|
* string for the target part of the object and the oid of the object
|
|
* whose comments are to be dumped. It is perfectly acceptable
|
|
* to hand an oid to this routine which has not been commented. In
|
|
* addition, the routine takes the stdio FILE handle to which the
|
|
* output should be written.
|
|
*------------------------------------------------------------------
|
|
*/
|
|
|
|
static void
|
|
dumpComment(Archive *fout, const char *target, const char *oid)
|
|
{
|
|
|
|
PGresult *res;
|
|
PQExpBuffer query;
|
|
int i_description;
|
|
|
|
/*** Build query to find comment ***/
|
|
|
|
query = createPQExpBuffer();
|
|
appendPQExpBuffer(query, "SELECT description FROM pg_description WHERE objoid = ");
|
|
appendPQExpBuffer(query, oid);
|
|
|
|
/*** Execute query ***/
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res || PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "DumpComment: SELECT failed: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/*** If a comment exists, build COMMENT ON statement ***/
|
|
|
|
if (PQntuples(res) != 0)
|
|
{
|
|
i_description = PQfnumber(res, "description");
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "COMMENT ON %s IS ", target);
|
|
formatStringLiteral(query, PQgetvalue(res, 0, i_description));
|
|
appendPQExpBuffer(query, ";\n");
|
|
|
|
ArchiveEntry(fout, oid, target, "COMMENT", NULL, query->data, "" /*Del*/,
|
|
"" /* Copy */, "" /*Owner*/, NULL, NULL);
|
|
|
|
}
|
|
|
|
/*** Clear the statement buffer and return ***/
|
|
|
|
PQclear(res);
|
|
|
|
}
|
|
|
|
/*------------------------------------------------------------------
|
|
* dumpDBComment --
|
|
*
|
|
* This routine is used to dump any comments associated with the
|
|
* database to which we are currently connected. If the user chose
|
|
* to dump the schema of the database, then this is the first
|
|
* statement issued.
|
|
*------------------------------------------------------------------
|
|
*/
|
|
|
|
void
|
|
dumpDBComment(Archive *fout)
|
|
{
|
|
|
|
PGresult *res;
|
|
PQExpBuffer query;
|
|
int i_oid;
|
|
|
|
/*** Build query to find comment ***/
|
|
|
|
query = createPQExpBuffer();
|
|
appendPQExpBuffer(query, "SELECT oid FROM pg_database WHERE datname = ");
|
|
formatStringLiteral(query, PQdb(g_conn));
|
|
|
|
/*** Execute query ***/
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res || PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpDBComment: SELECT failed: '%s'.\n",
|
|
PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/*** If a comment exists, build COMMENT ON statement ***/
|
|
|
|
if (PQntuples(res) != 0)
|
|
{
|
|
i_oid = PQfnumber(res, "oid");
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "DATABASE %s", fmtId(PQdb(g_conn), force_quotes));
|
|
dumpComment(fout, query->data, PQgetvalue(res, 0, i_oid));
|
|
}
|
|
|
|
/*** Clear the statement buffer and return ***/
|
|
|
|
PQclear(res);
|
|
|
|
}
|
|
|
|
/*
|
|
* dumpTypes
|
|
* writes out to fout the queries to recreate all the user-defined types
|
|
*
|
|
*/
|
|
void
|
|
dumpTypes(Archive *fout, FuncInfo *finfo, int numFuncs,
|
|
TypeInfo *tinfo, int numTypes)
|
|
{
|
|
int i;
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
PQExpBuffer delq = createPQExpBuffer();
|
|
int funcInd;
|
|
|
|
for (i = 0; i < numTypes; i++)
|
|
{
|
|
|
|
/* skip all the builtin types */
|
|
if (atooid(tinfo[i].oid) <= g_last_builtin_oid)
|
|
continue;
|
|
|
|
/* skip relation types */
|
|
if (atoi(tinfo[i].typrelid) != 0)
|
|
continue;
|
|
|
|
/* skip all array types that start w/ underscore */
|
|
if ((tinfo[i].typname[0] == '_') &&
|
|
(strcmp(tinfo[i].typinput, "array_in") == 0))
|
|
continue;
|
|
|
|
/*
|
|
* before we create a type, we need to create the input and output
|
|
* functions for it, if they haven't been created already
|
|
*/
|
|
funcInd = findFuncByName(finfo, numFuncs, tinfo[i].typinput);
|
|
if (funcInd != -1)
|
|
dumpOneFunc(fout, finfo, funcInd, tinfo, numTypes);
|
|
|
|
funcInd = findFuncByName(finfo, numFuncs, tinfo[i].typoutput);
|
|
if (funcInd != -1)
|
|
dumpOneFunc(fout, finfo, funcInd, tinfo, numTypes);
|
|
|
|
appendPQExpBuffer(delq, "DROP TYPE %s;\n", fmtId(tinfo[i].typname, force_quotes));
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q,
|
|
"CREATE TYPE %s "
|
|
"( internallength = %s, externallength = %s,",
|
|
fmtId(tinfo[i].typname, force_quotes),
|
|
tinfo[i].typlen,
|
|
tinfo[i].typprtlen);
|
|
/* cannot combine these because fmtId uses static result area */
|
|
appendPQExpBuffer(q, " input = %s,",
|
|
fmtId(tinfo[i].typinput, force_quotes));
|
|
appendPQExpBuffer(q, " output = %s,",
|
|
fmtId(tinfo[i].typoutput, force_quotes));
|
|
appendPQExpBuffer(q, " send = %s,",
|
|
fmtId(tinfo[i].typsend, force_quotes));
|
|
appendPQExpBuffer(q, " receive = %s, default = ",
|
|
fmtId(tinfo[i].typreceive, force_quotes));
|
|
formatStringLiteral(q, tinfo[i].typdefault);
|
|
|
|
if (tinfo[i].isArray)
|
|
{
|
|
char *elemType;
|
|
|
|
elemType = findTypeByOid(tinfo, numTypes, tinfo[i].typelem, zeroAsOpaque);
|
|
if (elemType == NULL)
|
|
{
|
|
fprintf(stderr, "Notice: array type %s - type for elements (oid %s) is not dumped.\n",
|
|
tinfo[i].typname, tinfo[i].typelem);
|
|
resetPQExpBuffer(q);
|
|
resetPQExpBuffer(delq);
|
|
continue;
|
|
}
|
|
|
|
appendPQExpBuffer(q, ", element = %s, delimiter = ", elemType);
|
|
formatStringLiteral(q, tinfo[i].typdelim);
|
|
}
|
|
if (tinfo[i].passedbyvalue)
|
|
appendPQExpBuffer(q, ",passedbyvalue);\n");
|
|
else
|
|
appendPQExpBuffer(q, ");\n");
|
|
|
|
ArchiveEntry(fout, tinfo[i].oid, fmtId(tinfo[i].typname, force_quotes), "TYPE", NULL,
|
|
q->data, delq->data, "", tinfo[i].usename, NULL, NULL);
|
|
|
|
/*** Dump Type Comments ***/
|
|
|
|
resetPQExpBuffer(q);
|
|
resetPQExpBuffer(delq);
|
|
|
|
appendPQExpBuffer(q, "TYPE %s", fmtId(tinfo[i].typname, force_quotes));
|
|
dumpComment(fout, q->data, tinfo[i].oid);
|
|
|
|
resetPQExpBuffer(q);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* dumpProcLangs
|
|
* writes out to fout the queries to recreate user-defined procedural languages
|
|
*
|
|
*/
|
|
void
|
|
dumpProcLangs(Archive *fout, FuncInfo *finfo, int numFuncs,
|
|
TypeInfo *tinfo, int numTypes)
|
|
{
|
|
PGresult *res;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
PQExpBuffer defqry = createPQExpBuffer();
|
|
PQExpBuffer delqry = createPQExpBuffer();
|
|
int ntups;
|
|
int i_oid;
|
|
int i_lanname;
|
|
int i_lanpltrusted;
|
|
int i_lanplcallfoid;
|
|
int i_lancompiler;
|
|
Oid lanoid;
|
|
char *lanname;
|
|
char *lancompiler;
|
|
const char *lanplcallfoid;
|
|
int i,
|
|
fidx;
|
|
|
|
appendPQExpBuffer(query, "SELECT oid, * FROM pg_language "
|
|
"WHERE lanispl "
|
|
"ORDER BY oid");
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpProcLangs(): SELECT failed. Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
ntups = PQntuples(res);
|
|
|
|
i_lanname = PQfnumber(res, "lanname");
|
|
i_lanpltrusted = PQfnumber(res, "lanpltrusted");
|
|
i_lanplcallfoid = PQfnumber(res, "lanplcallfoid");
|
|
i_lancompiler = PQfnumber(res, "lancompiler");
|
|
i_oid = PQfnumber(res, "oid");
|
|
|
|
for (i = 0; i < ntups; i++)
|
|
{
|
|
lanoid = atooid(PQgetvalue(res, i, i_oid));
|
|
if (lanoid <= g_last_builtin_oid)
|
|
continue;
|
|
|
|
lanplcallfoid = PQgetvalue(res, i, i_lanplcallfoid);
|
|
|
|
|
|
for (fidx = 0; fidx < numFuncs; fidx++)
|
|
{
|
|
if (!strcmp(finfo[fidx].oid, lanplcallfoid))
|
|
break;
|
|
}
|
|
if (fidx >= numFuncs)
|
|
{
|
|
fprintf(stderr, "dumpProcLangs(): handler procedure for "
|
|
"language %s not found\n", PQgetvalue(res, i, i_lanname));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
dumpOneFunc(fout, finfo, fidx, tinfo, numTypes);
|
|
|
|
lanname = PQgetvalue(res, i, i_lanname);
|
|
lancompiler = PQgetvalue(res, i, i_lancompiler);
|
|
|
|
appendPQExpBuffer(delqry, "DROP PROCEDURAL LANGUAGE ");
|
|
formatStringLiteral(delqry, lanname);
|
|
appendPQExpBuffer(delqry, ";\n");
|
|
|
|
appendPQExpBuffer(defqry, "CREATE %sPROCEDURAL LANGUAGE ",
|
|
(PQgetvalue(res, i, i_lanpltrusted)[0] == 't') ?
|
|
"TRUSTED " : "");
|
|
formatStringLiteral(defqry, lanname);
|
|
appendPQExpBuffer(defqry, " HANDLER %s LANCOMPILER ",
|
|
fmtId(finfo[fidx].proname, force_quotes));
|
|
formatStringLiteral(defqry, lancompiler);
|
|
appendPQExpBuffer(defqry, ";\n");
|
|
|
|
ArchiveEntry(fout, PQgetvalue(res, i, i_oid), lanname, "PROCEDURAL LANGUAGE",
|
|
NULL, defqry->data, delqry->data, "", "", NULL, NULL);
|
|
|
|
resetPQExpBuffer(defqry);
|
|
resetPQExpBuffer(delqry);
|
|
}
|
|
|
|
PQclear(res);
|
|
|
|
}
|
|
|
|
/*
|
|
* dumpFuncs
|
|
* writes out to fout the queries to recreate all the user-defined functions
|
|
*
|
|
*/
|
|
void
|
|
dumpFuncs(Archive *fout, FuncInfo *finfo, int numFuncs,
|
|
TypeInfo *tinfo, int numTypes)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < numFuncs; i++)
|
|
dumpOneFunc(fout, finfo, i, tinfo, numTypes);
|
|
}
|
|
|
|
/*
|
|
* dumpOneFunc:
|
|
* dump out only one function, the index of which is given in the third
|
|
* argument
|
|
*
|
|
*/
|
|
|
|
static void
|
|
dumpOneFunc(Archive *fout, FuncInfo *finfo, int i,
|
|
TypeInfo *tinfo, int numTypes)
|
|
{
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
PQExpBuffer fn = createPQExpBuffer();
|
|
PQExpBuffer delqry = createPQExpBuffer();
|
|
PQExpBuffer fnlist = createPQExpBuffer();
|
|
int j;
|
|
PQExpBuffer asPart = createPQExpBuffer();
|
|
char func_lang[NAMEDATALEN + 1];
|
|
PGresult *res;
|
|
int nlangs;
|
|
int i_lanname;
|
|
char query[256];
|
|
|
|
char *listSep;
|
|
char *listSepComma = ",";
|
|
char *listSepNone = "";
|
|
char *rettypename;
|
|
|
|
if (finfo[i].dumped)
|
|
return;
|
|
else
|
|
finfo[i].dumped = 1;
|
|
|
|
/* becomeUser(fout, finfo[i].usename); */
|
|
|
|
sprintf(query, "SELECT lanname FROM pg_language WHERE oid = %u",
|
|
finfo[i].lang);
|
|
res = PQexec(g_conn, query);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpOneFunc(): SELECT for procedural language failed. Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
nlangs = PQntuples(res);
|
|
|
|
if (nlangs != 1)
|
|
{
|
|
fprintf(stderr, "dumpOneFunc(): procedural language for function %s not found\n", finfo[i].proname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
i_lanname = PQfnumber(res, "lanname");
|
|
|
|
/*
|
|
* See backend/commands/define.c for details of how the 'AS' clause
|
|
* is used.
|
|
*/
|
|
if (strcmp(finfo[i].probin, "-") != 0)
|
|
{
|
|
appendPQExpBuffer(asPart, "AS ");
|
|
formatStringLiteral(asPart, finfo[i].probin);
|
|
if (strcmp(finfo[i].prosrc, "-") != 0)
|
|
{
|
|
appendPQExpBuffer(asPart, ", ");
|
|
formatStringLiteral(asPart, finfo[i].prosrc);
|
|
}
|
|
}
|
|
else
|
|
{
|
|
if (strcmp(finfo[i].prosrc, "-") != 0)
|
|
{
|
|
appendPQExpBuffer(asPart, "AS ");
|
|
formatStringLiteral(asPart, finfo[i].prosrc);
|
|
}
|
|
}
|
|
|
|
strcpy(func_lang, PQgetvalue(res, 0, i_lanname));
|
|
|
|
PQclear(res);
|
|
|
|
resetPQExpBuffer(fn);
|
|
appendPQExpBuffer(fn, "%s (", fmtId(finfo[i].proname, force_quotes));
|
|
for (j = 0; j < finfo[i].nargs; j++)
|
|
{
|
|
char *typname;
|
|
|
|
typname = findTypeByOid(tinfo, numTypes, finfo[i].argtypes[j], zeroAsOpaque);
|
|
if (typname == NULL)
|
|
{
|
|
fprintf(stderr, "Notice: function \"%s\" is not dumped.\n",
|
|
finfo[i].proname);
|
|
|
|
fprintf(stderr, "Reason: the %d th argument type name (oid %s) not found.\n",
|
|
j, finfo[i].argtypes[j]);
|
|
resetPQExpBuffer(q);
|
|
resetPQExpBuffer(fn);
|
|
resetPQExpBuffer(delqry);
|
|
resetPQExpBuffer(fnlist);
|
|
resetPQExpBuffer(asPart);
|
|
return;
|
|
}
|
|
|
|
appendPQExpBuffer(fn, "%s%s",
|
|
(j > 0) ? "," : "",
|
|
typname);
|
|
appendPQExpBuffer(fnlist, "%s%s",
|
|
(j > 0) ? "," : "",
|
|
typname );
|
|
}
|
|
appendPQExpBuffer(fn, ")");
|
|
|
|
resetPQExpBuffer(delqry);
|
|
appendPQExpBuffer(delqry, "DROP FUNCTION %s;\n", fn->data );
|
|
|
|
rettypename = findTypeByOid(tinfo, numTypes, finfo[i].prorettype, zeroAsOpaque);
|
|
|
|
if (rettypename == NULL)
|
|
{
|
|
fprintf(stderr, "Notice: function \"%s\" is not dumped.\n",
|
|
finfo[i].proname);
|
|
|
|
fprintf(stderr, "Reason: return type name (oid %s) not found.\n",
|
|
finfo[i].prorettype);
|
|
resetPQExpBuffer(q);
|
|
resetPQExpBuffer(fn);
|
|
resetPQExpBuffer(delqry);
|
|
resetPQExpBuffer(fnlist);
|
|
resetPQExpBuffer(asPart);
|
|
return;
|
|
}
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "CREATE FUNCTION %s ", fn->data );
|
|
appendPQExpBuffer(q, "RETURNS %s%s %s LANGUAGE ",
|
|
(finfo[i].retset) ? "SETOF " : "",
|
|
rettypename,
|
|
asPart->data);
|
|
formatStringLiteral(q, func_lang);
|
|
|
|
if (finfo[i].iscachable || finfo[i].isstrict) /* OR in new attrs here */
|
|
{
|
|
appendPQExpBuffer(q, " WITH (");
|
|
listSep = listSepNone;
|
|
|
|
if (finfo[i].iscachable) {
|
|
appendPQExpBuffer(q, "%s iscachable", listSep);
|
|
listSep = listSepComma;
|
|
}
|
|
|
|
if (finfo[i].isstrict) {
|
|
appendPQExpBuffer(q, "%s isstrict", listSep);
|
|
listSep = listSepComma;
|
|
}
|
|
appendPQExpBuffer(q, " )");
|
|
}
|
|
|
|
appendPQExpBuffer(q, ";\n");
|
|
|
|
ArchiveEntry(fout, finfo[i].oid, fn->data, "FUNCTION", NULL, q->data, delqry->data,
|
|
"", finfo[i].usename, NULL, NULL);
|
|
|
|
/*** Dump Function Comments ***/
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "FUNCTION %s ",
|
|
fmtId(finfo[i].proname, force_quotes));
|
|
appendPQExpBuffer(q, "( %s )", fnlist->data);
|
|
dumpComment(fout, q->data, finfo[i].oid);
|
|
|
|
}
|
|
|
|
/*
|
|
* dumpOprs
|
|
* writes out to fout the queries to recreate all the user-defined operators
|
|
*
|
|
*/
|
|
void
|
|
dumpOprs(Archive *fout, OprInfo *oprinfo, int numOperators,
|
|
TypeInfo *tinfo, int numTypes)
|
|
{
|
|
#define OPR_NOTICE(arg) {\
|
|
fprintf(stderr, "Notice: operator \"%s\"(oid %s) is not dumped.\n",oprinfo[i].oprname, oprinfo[i].oid);\
|
|
fprintf(stderr, "Reason: " CppAsString(arg));\
|
|
fprintf (stderr, " (oid %s) not found.\n",oprinfo[i].arg);\
|
|
}
|
|
|
|
int i;
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
PQExpBuffer delq = createPQExpBuffer();
|
|
PQExpBuffer leftarg = createPQExpBuffer();
|
|
PQExpBuffer rightarg = createPQExpBuffer();
|
|
PQExpBuffer commutator = createPQExpBuffer();
|
|
PQExpBuffer negator = createPQExpBuffer();
|
|
PQExpBuffer restrictor = createPQExpBuffer();
|
|
PQExpBuffer join = createPQExpBuffer();
|
|
PQExpBuffer sort1 = createPQExpBuffer();
|
|
PQExpBuffer sort2 = createPQExpBuffer();
|
|
|
|
for (i = 0; i < numOperators; i++)
|
|
{
|
|
char *name;
|
|
|
|
resetPQExpBuffer(leftarg);
|
|
resetPQExpBuffer(rightarg);
|
|
resetPQExpBuffer(commutator);
|
|
resetPQExpBuffer(negator);
|
|
resetPQExpBuffer(restrictor);
|
|
resetPQExpBuffer(join);
|
|
resetPQExpBuffer(sort1);
|
|
resetPQExpBuffer(sort2);
|
|
|
|
/* skip all the builtin oids */
|
|
if (atooid(oprinfo[i].oid) <= g_last_builtin_oid)
|
|
continue;
|
|
|
|
/*
|
|
* some operator are invalid because they were the result of user
|
|
* defining operators before commutators exist
|
|
*/
|
|
if (strcmp(oprinfo[i].oprcode, "-") == 0)
|
|
continue;
|
|
|
|
/*
|
|
* right unary means there's a left arg and left unary means
|
|
* there's a right arg
|
|
*/
|
|
if (strcmp(oprinfo[i].oprkind, "r") == 0 ||
|
|
strcmp(oprinfo[i].oprkind, "b") == 0)
|
|
{
|
|
name = findTypeByOid(tinfo, numTypes,
|
|
oprinfo[i].oprleft, zeroAsOpaque);
|
|
if (name == NULL)
|
|
{
|
|
OPR_NOTICE(oprleft);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(leftarg, ",\n\tLEFTARG = %s ",name);
|
|
}
|
|
|
|
if (strcmp(oprinfo[i].oprkind, "l") == 0 ||
|
|
strcmp(oprinfo[i].oprkind, "b") == 0)
|
|
{
|
|
name = findTypeByOid(tinfo, numTypes,
|
|
oprinfo[i].oprright, zeroAsOpaque);
|
|
if (name == NULL)
|
|
{
|
|
OPR_NOTICE(oprright);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(rightarg, ",\n\tRIGHTARG = %s ", name);
|
|
}
|
|
|
|
if (!(strcmp(oprinfo[i].oprcom, "0") == 0))
|
|
{
|
|
name = findOprByOid(oprinfo, numOperators, oprinfo[i].oprcom);
|
|
if (name == NULL)
|
|
{
|
|
OPR_NOTICE(oprcom);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(commutator, ",\n\tCOMMUTATOR = %s ", name);
|
|
}
|
|
|
|
if (!(strcmp(oprinfo[i].oprnegate, "0") == 0))
|
|
{
|
|
name = findOprByOid(oprinfo, numOperators, oprinfo[i].oprnegate);
|
|
if (name == NULL)
|
|
{
|
|
OPR_NOTICE(oprnegate);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(negator, ",\n\tNEGATOR = %s ", name);
|
|
}
|
|
|
|
if (!(strcmp(oprinfo[i].oprrest, "-") == 0))
|
|
appendPQExpBuffer(restrictor, ",\n\tRESTRICT = %s ", oprinfo[i].oprrest);
|
|
|
|
if (!(strcmp(oprinfo[i].oprjoin, "-") == 0))
|
|
appendPQExpBuffer(join, ",\n\tJOIN = %s ", oprinfo[i].oprjoin);
|
|
|
|
if (!(strcmp(oprinfo[i].oprlsortop, "0") == 0))
|
|
{
|
|
name = findOprByOid(oprinfo, numOperators, oprinfo[i].oprlsortop);
|
|
if (name == NULL)
|
|
{
|
|
OPR_NOTICE(oprlsortop);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(sort1, ",\n\tSORT1 = %s ", name);
|
|
}
|
|
|
|
if (!(strcmp(oprinfo[i].oprrsortop, "0") == 0))
|
|
{
|
|
name = findOprByOid(oprinfo, numOperators, oprinfo[i].oprrsortop);
|
|
if (name == NULL)
|
|
{
|
|
OPR_NOTICE(oprrsortop);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(sort2, ",\n\tSORT2 = %s ", name);
|
|
}
|
|
|
|
resetPQExpBuffer(delq);
|
|
appendPQExpBuffer(delq, "DROP OPERATOR %s (%s", oprinfo[i].oprname,
|
|
findTypeByOid(tinfo, numTypes, oprinfo[i].oprleft, zeroAsOpaque) );
|
|
appendPQExpBuffer(delq, ", %s);\n",
|
|
findTypeByOid(tinfo, numTypes, oprinfo[i].oprright, zeroAsOpaque) );
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q,
|
|
"CREATE OPERATOR %s "
|
|
"(PROCEDURE = %s %s%s%s%s%s%s%s%s%s);\n",
|
|
oprinfo[i].oprname,
|
|
oprinfo[i].oprcode,
|
|
leftarg->data,
|
|
rightarg->data,
|
|
commutator->data,
|
|
negator->data,
|
|
restrictor->data,
|
|
(strcmp(oprinfo[i].oprcanhash, "t") == 0) ? ",\n\tHASHES" : "",
|
|
join->data,
|
|
sort1->data,
|
|
sort2->data);
|
|
|
|
ArchiveEntry(fout, oprinfo[i].oid, oprinfo[i].oprname, "OPERATOR", NULL,
|
|
q->data, delq->data, "", oprinfo[i].usename, NULL, NULL);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* dumpAggs
|
|
* writes out to fout the queries to create all the user-defined aggregates
|
|
*
|
|
*/
|
|
void
|
|
dumpAggs(Archive *fout, AggInfo *agginfo, int numAggs,
|
|
TypeInfo *tinfo, int numTypes)
|
|
{
|
|
#define AGG_NOTICE(arg) {\
|
|
fprintf(stderr, "Notice: aggregate \"%s\"(oid %s) is not dumped.\n",agginfo[i].aggname, agginfo[i].oid);\
|
|
fprintf(stderr, "Reason: " CppAsString(arg) );\
|
|
fprintf (stderr, " (oid %s) not found.\n",agginfo[i].arg);\
|
|
}
|
|
|
|
int i;
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
PQExpBuffer delq = createPQExpBuffer();
|
|
PQExpBuffer aggSig = createPQExpBuffer();
|
|
PQExpBuffer details = createPQExpBuffer();
|
|
|
|
for (i = 0; i < numAggs; i++)
|
|
{
|
|
char *name;
|
|
|
|
resetPQExpBuffer(details);
|
|
|
|
/* skip all the builtin oids */
|
|
if (atooid(agginfo[i].oid) <= g_last_builtin_oid)
|
|
continue;
|
|
|
|
name = findTypeByOid(tinfo, numTypes, agginfo[i].aggbasetype, zeroAsAny + useBaseTypeName);
|
|
if (name == NULL)
|
|
{
|
|
AGG_NOTICE(aggbasetype);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(details, "BASETYPE = %s, ", name);
|
|
|
|
name = findTypeByOid(tinfo, numTypes, agginfo[i].aggtranstype, zeroAsOpaque + useBaseTypeName);
|
|
if (name == NULL)
|
|
{
|
|
AGG_NOTICE(aggtranstype);
|
|
continue;
|
|
}
|
|
appendPQExpBuffer(details,
|
|
"SFUNC = %s, STYPE = %s",
|
|
agginfo[i].aggtransfn, name);
|
|
|
|
if (agginfo[i].agginitval)
|
|
{
|
|
appendPQExpBuffer(details, ", INITCOND = ");
|
|
formatStringLiteral(details, agginfo[i].agginitval);
|
|
}
|
|
|
|
if (!(strcmp(agginfo[i].aggfinalfn, "-") == 0))
|
|
appendPQExpBuffer(details, ", FINALFUNC = %s",
|
|
agginfo[i].aggfinalfn);
|
|
|
|
resetPQExpBuffer(aggSig);
|
|
appendPQExpBuffer(aggSig, "%s %s", agginfo[i].aggname,
|
|
findTypeByOid(tinfo, numTypes, agginfo[i].aggbasetype, zeroAsOpaque + useBaseTypeName));
|
|
|
|
resetPQExpBuffer(delq);
|
|
appendPQExpBuffer(delq, "DROP AGGREGATE %s;\n", aggSig->data);
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "CREATE AGGREGATE %s ( %s );\n",
|
|
agginfo[i].aggname,
|
|
details->data);
|
|
|
|
ArchiveEntry(fout, agginfo[i].oid, aggSig->data, "AGGREGATE", NULL,
|
|
q->data, delq->data, "", agginfo[i].usename, NULL, NULL);
|
|
|
|
/*** Dump Aggregate Comments ***/
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "AGGREGATE %s %s", agginfo[i].aggname,
|
|
findTypeByOid(tinfo, numTypes, agginfo[i].aggbasetype, zeroAsOpaque + useBaseTypeName));
|
|
dumpComment(fout, q->data, agginfo[i].oid);
|
|
|
|
}
|
|
}
|
|
|
|
/*
|
|
* These are some support functions to fix the acl problem of pg_dump
|
|
*
|
|
* Matthew C. Aycock 12/02/97
|
|
*/
|
|
|
|
/* Append a keyword to a keyword list, inserting comma if needed.
|
|
* Caller must make aclbuf big enough for all possible keywords.
|
|
*/
|
|
static void
|
|
AddAcl(char *aclbuf, const char *keyword)
|
|
{
|
|
if (*aclbuf)
|
|
strcat(aclbuf, ",");
|
|
strcat(aclbuf, keyword);
|
|
}
|
|
|
|
/*
|
|
* This will take a string of 'arwR' and return a malloced,
|
|
* comma delimited string of SELECT,INSERT,UPDATE,DELETE,RULE
|
|
*/
|
|
static char *
|
|
GetPrivileges(const char *s)
|
|
{
|
|
char aclbuf[100];
|
|
|
|
aclbuf[0] = '\0';
|
|
|
|
if (strchr(s, 'a'))
|
|
AddAcl(aclbuf, "INSERT");
|
|
|
|
if (strchr(s, 'w'))
|
|
AddAcl(aclbuf, "UPDATE,DELETE");
|
|
|
|
if (strchr(s, 'r'))
|
|
AddAcl(aclbuf, "SELECT");
|
|
|
|
if (strchr(s, 'R'))
|
|
AddAcl(aclbuf, "RULE");
|
|
|
|
/* Special-case when they're all there */
|
|
if (strcmp(aclbuf, "INSERT,UPDATE,DELETE,SELECT,RULE") == 0)
|
|
return strdup("ALL");
|
|
|
|
return strdup(aclbuf);
|
|
}
|
|
|
|
/*
|
|
* The name says it all; a function to append a string is the dest
|
|
* is big enough. If not, it does a realloc.
|
|
*/
|
|
static void strcatalloc(char **dest, int *dSize, char *src)
|
|
{
|
|
int dLen = strlen(*dest);
|
|
int sLen = strlen(src);
|
|
if ( (dLen + sLen) >= *dSize) {
|
|
*dSize = (dLen + sLen) * 2;
|
|
*dest = realloc(*dest, *dSize);
|
|
}
|
|
strcpy(*dest + dLen, src);
|
|
}
|
|
|
|
|
|
/*
|
|
* dumpACL:
|
|
* Write out grant/revoke information
|
|
* Called for sequences and tables
|
|
*/
|
|
|
|
static void
|
|
dumpACL(Archive *fout, TableInfo tbinfo)
|
|
{
|
|
const char *acls = tbinfo.relacl;
|
|
char *aclbuf,
|
|
*tok,
|
|
*eqpos,
|
|
*priv;
|
|
char *sql;
|
|
char tmp[1024];
|
|
int sSize = 4096;
|
|
|
|
if (strlen(acls) == 0)
|
|
return; /* table has default permissions */
|
|
|
|
/*
|
|
* Allocate a larginsh buffer for the output SQL.
|
|
*/
|
|
sql = (char*)malloc(sSize);
|
|
|
|
/*
|
|
* Revoke Default permissions for PUBLIC. Is this actually necessary,
|
|
* or is it just a waste of time?
|
|
*/
|
|
sprintf(sql, "REVOKE ALL on %s from PUBLIC;\n",
|
|
fmtId(tbinfo.relname, force_quotes));
|
|
|
|
/* Make a working copy of acls so we can use strtok */
|
|
aclbuf = strdup(acls);
|
|
|
|
/* Scan comma-separated ACL items */
|
|
for (tok = strtok(aclbuf, ","); tok != NULL; tok = strtok(NULL, ","))
|
|
{
|
|
|
|
/*
|
|
* Token may start with '{' and/or '"'. Actually only the start
|
|
* of the string should have '{', but we don't verify that.
|
|
*/
|
|
if (*tok == '{')
|
|
tok++;
|
|
if (*tok == '"')
|
|
tok++;
|
|
|
|
/* User name is string up to = in tok */
|
|
eqpos = strchr(tok, '=');
|
|
if (!eqpos)
|
|
{
|
|
fprintf(stderr, "Could not parse ACL list ('%s') for '%s'...Exiting!\n",
|
|
acls, tbinfo.relname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/*
|
|
* Parse the privileges (right-hand side). Skip if there are
|
|
* none.
|
|
*/
|
|
priv = GetPrivileges(eqpos + 1);
|
|
if (*priv)
|
|
{
|
|
sprintf(tmp, "GRANT %s on %s to ",
|
|
priv, fmtId(tbinfo.relname, force_quotes));
|
|
strcatalloc(&sql, &sSize, tmp);
|
|
|
|
/*
|
|
* Note: fmtId() can only be called once per printf, so don't
|
|
* try to merge printing of username into the above printf.
|
|
*/
|
|
if (eqpos == tok)
|
|
{
|
|
/* Empty left-hand side means "PUBLIC" */
|
|
strcatalloc(&sql, &sSize, "PUBLIC;\n");
|
|
}
|
|
else
|
|
{
|
|
*eqpos = '\0'; /* it's ok to clobber aclbuf */
|
|
if (strncmp(tok, "group ", strlen("group ")) == 0)
|
|
sprintf(tmp, "GROUP %s;\n",
|
|
fmtId(tok + strlen("group "), force_quotes));
|
|
else
|
|
sprintf(tmp, "%s;\n", fmtId(tok, force_quotes));
|
|
strcatalloc(&sql, &sSize, tmp);
|
|
}
|
|
}
|
|
free(priv);
|
|
}
|
|
|
|
free(aclbuf);
|
|
|
|
ArchiveEntry(fout, tbinfo.oid, tbinfo.relname, "ACL", NULL, sql, "", "", "", NULL, NULL);
|
|
|
|
}
|
|
|
|
|
|
/*
|
|
* dumpTables:
|
|
* write out to fout all the user-define tables
|
|
*/
|
|
|
|
void
|
|
dumpTables(Archive *fout, TableInfo *tblinfo, int numTables,
|
|
IndInfo *indinfo, int numIndices,
|
|
InhInfo *inhinfo, int numInherits,
|
|
TypeInfo *tinfo, int numTypes, const char *tablename,
|
|
const bool aclsSkip, const bool oids,
|
|
const bool schemaOnly, const bool dataOnly)
|
|
{
|
|
int i,
|
|
j,
|
|
k;
|
|
PQExpBuffer q = createPQExpBuffer();
|
|
PQExpBuffer delq = createPQExpBuffer();
|
|
char *serialSeq = NULL; /* implicit sequence name created
|
|
* by SERIAL datatype */
|
|
const char *serialSeqSuffix = "_id_seq"; /* suffix for implicit
|
|
* SERIAL sequences */
|
|
char **parentRels; /* list of names of parent relations */
|
|
int numParents;
|
|
int actual_atts; /* number of attrs in this CREATE statment */
|
|
char *reltypename;
|
|
|
|
/* First - dump SEQUENCEs */
|
|
if (tablename && strlen(tablename) > 0)
|
|
{
|
|
serialSeq = malloc(strlen(tablename) + strlen(serialSeqSuffix) + 1);
|
|
strcpy(serialSeq, tablename);
|
|
strcat(serialSeq, serialSeqSuffix);
|
|
}
|
|
for (i = 0; i < numTables; i++)
|
|
{
|
|
if (!(tblinfo[i].sequence))
|
|
continue;
|
|
if (!tablename || (!strcmp(tblinfo[i].relname, tablename))
|
|
|| (serialSeq && !strcmp(tblinfo[i].relname, serialSeq)))
|
|
{
|
|
/* becomeUser(fout, tblinfo[i].usename); */
|
|
dumpSequence(fout, tblinfo[i]);
|
|
if (!aclsSkip)
|
|
dumpACL(fout, tblinfo[i]);
|
|
}
|
|
}
|
|
if (tablename)
|
|
free(serialSeq);
|
|
|
|
for (i = 0; i < numTables; i++)
|
|
{
|
|
if (tblinfo[i].sequence)/* already dumped */
|
|
continue;
|
|
|
|
if (!tablename || (!strcmp(tblinfo[i].relname, tablename)) || (strlen(tablename) == 0))
|
|
{
|
|
|
|
resetPQExpBuffer(delq);
|
|
resetPQExpBuffer(q);
|
|
|
|
/* Use the view definition if there is one */
|
|
if (tblinfo[i].viewdef != NULL)
|
|
{
|
|
reltypename = "VIEW";
|
|
|
|
appendPQExpBuffer(delq, "DROP VIEW %s;\n", fmtId(tblinfo[i].relname, force_quotes));
|
|
appendPQExpBuffer(q, "CREATE VIEW %s as %s", fmtId(tblinfo[i].relname, force_quotes), tblinfo[i].viewdef);
|
|
|
|
}
|
|
else
|
|
{
|
|
reltypename = "TABLE";
|
|
|
|
parentRels = tblinfo[i].parentRels;
|
|
numParents = tblinfo[i].numParents;
|
|
|
|
appendPQExpBuffer(delq, "DROP TABLE %s;\n", fmtId(tblinfo[i].relname, force_quotes));
|
|
|
|
appendPQExpBuffer(q, "CREATE TABLE %s (\n\t", fmtId(tblinfo[i].relname, force_quotes));
|
|
actual_atts = 0;
|
|
for (j = 0; j < tblinfo[i].numatts; j++)
|
|
{
|
|
/* Is this one of the table's own attrs ? */
|
|
if (tblinfo[i].inhAttrs[j] == 0)
|
|
{
|
|
/* Format properly if not first attr */
|
|
if (actual_atts > 0)
|
|
appendPQExpBuffer(q, ",\n\t");
|
|
|
|
/* Attr name & type */
|
|
appendPQExpBuffer(q, "%s %s",
|
|
fmtId(tblinfo[i].attnames[j], force_quotes),
|
|
tblinfo[i].atttypedefns[j]);
|
|
|
|
/* Default value */
|
|
if (tblinfo[i].adef_expr[j] != NULL)
|
|
appendPQExpBuffer(q, " DEFAULT %s",
|
|
tblinfo[i].adef_expr[j]);
|
|
|
|
/* Not Null constraint */
|
|
if (tblinfo[i].notnull[j])
|
|
appendPQExpBuffer(q, " NOT NULL");
|
|
|
|
actual_atts++;
|
|
}
|
|
}
|
|
|
|
|
|
|
|
/* put the CONSTRAINTS inside the table def */
|
|
for (k = 0; k < tblinfo[i].ncheck; k++)
|
|
{
|
|
if (actual_atts + k > 0)
|
|
appendPQExpBuffer(q, ",\n\t");
|
|
|
|
appendPQExpBuffer(q, "%s",
|
|
tblinfo[i].check_expr[k]);
|
|
}
|
|
|
|
/* Primary Key */
|
|
if (tblinfo[i].pkIndexOid != NULL)
|
|
{
|
|
PQExpBuffer consDef;
|
|
|
|
/* Find the corresponding index */
|
|
for (k = 0; k < numIndices; k++)
|
|
{
|
|
if (strcmp(indinfo[k].oid, tblinfo[i].pkIndexOid) == 0)
|
|
break;
|
|
}
|
|
|
|
if (k >= numIndices)
|
|
{
|
|
fprintf(stderr, "dumpTables(): failed sanity check, could not find index (%s) for PK constraint\n",
|
|
tblinfo[i].pkIndexOid);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
consDef = getPKconstraint(&tblinfo[i], &indinfo[k]);
|
|
|
|
if ( (actual_atts + tblinfo[i].ncheck) > 0)
|
|
appendPQExpBuffer(q, ",\n\t");
|
|
|
|
appendPQExpBuffer(q, "%s", consDef->data);
|
|
|
|
destroyPQExpBuffer(consDef);
|
|
}
|
|
|
|
|
|
appendPQExpBuffer(q, "\n)");
|
|
|
|
if (numParents > 0)
|
|
{
|
|
appendPQExpBuffer(q, "\ninherits (");
|
|
for (k = 0; k < numParents; k++)
|
|
{
|
|
appendPQExpBuffer(q, "%s%s",
|
|
(k > 0) ? ", " : "",
|
|
fmtId(parentRels[k], force_quotes));
|
|
}
|
|
appendPQExpBuffer(q, ")");
|
|
}
|
|
|
|
appendPQExpBuffer(q, ";\n");
|
|
}
|
|
|
|
if (!dataOnly) {
|
|
|
|
ArchiveEntry(fout, tblinfo[i].oid, fmtId(tblinfo[i].relname, false),
|
|
reltypename, NULL, q->data, delq->data, "", tblinfo[i].usename,
|
|
NULL, NULL);
|
|
|
|
if (!aclsSkip)
|
|
dumpACL(fout, tblinfo[i]);
|
|
|
|
}
|
|
|
|
/* Dump Field Comments */
|
|
|
|
for (j = 0; j < tblinfo[i].numatts; j++)
|
|
{
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "COLUMN %s", fmtId(tblinfo[i].relname, force_quotes));
|
|
appendPQExpBuffer(q, ".");
|
|
appendPQExpBuffer(q, "%s", fmtId(tblinfo[i].attnames[j], force_quotes));
|
|
dumpComment(fout, q->data, tblinfo[i].attoids[j]);
|
|
}
|
|
|
|
/* Dump Table Comments */
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "%s %s", reltypename, fmtId(tblinfo[i].relname, force_quotes));
|
|
dumpComment(fout, q->data, tblinfo[i].oid);
|
|
|
|
}
|
|
}
|
|
}
|
|
|
|
static PQExpBuffer getPKconstraint(TableInfo *tblInfo, IndInfo *indInfo)
|
|
{
|
|
PQExpBuffer pkBuf = createPQExpBuffer();
|
|
int k;
|
|
int indkey;
|
|
|
|
resetPQExpBuffer(pkBuf);
|
|
|
|
appendPQExpBuffer(pkBuf, "Constraint %s Primary Key (",
|
|
tblInfo->primary_key_name);
|
|
|
|
|
|
for (k = 0; k < INDEX_MAX_KEYS; k++)
|
|
{
|
|
char *attname;
|
|
|
|
indkey = atoi(indInfo->indkey[k]);
|
|
if (indkey == InvalidAttrNumber)
|
|
break;
|
|
indkey--;
|
|
if (indkey == ObjectIdAttributeNumber - 1)
|
|
attname = "oid";
|
|
else
|
|
attname = tblInfo->attnames[indkey];
|
|
|
|
appendPQExpBuffer(pkBuf, "%s%s",
|
|
(k == 0) ? "" : ", ",
|
|
fmtId(attname, force_quotes));
|
|
}
|
|
|
|
appendPQExpBuffer(pkBuf, ")");
|
|
|
|
return pkBuf;
|
|
}
|
|
|
|
/*
|
|
* dumpIndices:
|
|
* write out to fout all the user-define indices
|
|
*/
|
|
void
|
|
dumpIndices(Archive *fout, IndInfo *indinfo, int numIndices,
|
|
TableInfo *tblinfo, int numTables, const char *tablename)
|
|
{
|
|
int i,
|
|
k;
|
|
int tableInd;
|
|
PQExpBuffer attlist = createPQExpBuffer();
|
|
char *classname[INDEX_MAX_KEYS];
|
|
char *funcname; /* the name of the function to comput the
|
|
* index key from */
|
|
int indkey,
|
|
indclass;
|
|
int nclass;
|
|
|
|
PQExpBuffer q = createPQExpBuffer(),
|
|
delq = createPQExpBuffer(),
|
|
id1 = createPQExpBuffer(),
|
|
id2 = createPQExpBuffer();
|
|
PGresult *res;
|
|
|
|
for (i = 0; i < numIndices; i++)
|
|
{
|
|
tableInd = findTableByName(tblinfo, numTables,
|
|
indinfo[i].indrelname);
|
|
if (tableInd < 0)
|
|
{
|
|
fprintf(stderr, "failed sanity check, table %s was not found\n",
|
|
indinfo[i].indrelname);
|
|
exit(1);
|
|
}
|
|
|
|
/* Handle PK indexes */
|
|
if (strcmp(indinfo[i].indisprimary, "t") == 0)
|
|
{
|
|
/*
|
|
* ***PK: Enable this code when ALTER TABLE supports PK constraints. ***
|
|
*
|
|
* PQExpBuffer consDef = getPKconstraint(&tblinfo[tableInd], &indinfo[i]);
|
|
*
|
|
* resetPQExpBuffer(attlist);
|
|
*
|
|
* appendPQExpBuffer(attlist, "Alter Table %s Add %s;",
|
|
* fmtId(tblinfo[tableInd].relname, force_quotes),
|
|
* consDef->data);
|
|
*
|
|
* ArchiveEntry(fout, indinfo[i].oid, tblinfo[tableInd].primary_key_name, "CONSTRAINT", NULL,
|
|
* attlist->data, "",
|
|
* "", tblinfo[tableInd].usename, NULL, NULL);
|
|
*
|
|
* destroyPQExpBuffer(consDef);
|
|
*/
|
|
/* Don't need to do anything else for this system-generated index */
|
|
continue;
|
|
}
|
|
|
|
|
|
if (strcmp(indinfo[i].indproc, "0") == 0)
|
|
funcname = NULL;
|
|
else
|
|
{
|
|
int numFuncs;
|
|
|
|
/*
|
|
* the funcname is an oid which we use to find the name of the
|
|
* pg_proc. We need to do this because getFuncs() only reads
|
|
* in the user-defined funcs not all the funcs. We might not
|
|
* find what we want by looking in FuncInfo*
|
|
*/
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q,
|
|
"SELECT proname from pg_proc "
|
|
"where pg_proc.oid = '%s'::oid",
|
|
indinfo[i].indproc);
|
|
res = PQexec(g_conn, q->data);
|
|
if (!res || PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpIndices(): SELECT (funcname) failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Sanity: Check we got only one tuple */
|
|
numFuncs = PQntuples(res);
|
|
if (numFuncs != 1) {
|
|
fprintf(stderr, "dumpIndices(): SELECT (funcname) for index %s returned %d tuples. Expected 1.\n",
|
|
indinfo[i].indrelname, numFuncs);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
funcname = strdup(PQgetvalue(res, 0, PQfnumber(res, "proname")));
|
|
PQclear(res);
|
|
}
|
|
|
|
/* convert opclass oid(s) into names */
|
|
for (nclass = 0; nclass < INDEX_MAX_KEYS; nclass++)
|
|
{
|
|
int numRows;
|
|
|
|
indclass = atoi(indinfo[i].indclass[nclass]);
|
|
if (indclass == 0)
|
|
break;
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q,
|
|
"SELECT opcname from pg_opclass "
|
|
"where pg_opclass.oid = '%u'::oid",
|
|
indclass);
|
|
res = PQexec(g_conn, q->data);
|
|
if (!res || PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpIndices(): SELECT (classname) failed. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* Sanity: Check we got only one tuple */
|
|
numRows = PQntuples(res);
|
|
if (numRows != 1) {
|
|
fprintf(stderr, "dumpIndices(): SELECT (classname) for index %s returned %d tuples. Expected 1.\n",
|
|
indinfo[i].indrelname, numRows);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
classname[nclass] = strdup(PQgetvalue(res, 0, PQfnumber(res, "opcname")));
|
|
PQclear(res);
|
|
}
|
|
|
|
if (funcname && nclass != 1)
|
|
{
|
|
fprintf(stderr, "dumpIndices(): Must be exactly one OpClass "
|
|
"for functional index %s\n", indinfo[i].indexrelname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
/* convert attribute numbers into attribute list */
|
|
resetPQExpBuffer(attlist);
|
|
for (k = 0; k < INDEX_MAX_KEYS; k++)
|
|
{
|
|
char *attname;
|
|
|
|
indkey = atoi(indinfo[i].indkey[k]);
|
|
if (indkey == InvalidAttrNumber)
|
|
break;
|
|
indkey--;
|
|
if (indkey == ObjectIdAttributeNumber - 1)
|
|
attname = "oid";
|
|
else
|
|
attname = tblinfo[tableInd].attnames[indkey];
|
|
if (funcname)
|
|
appendPQExpBuffer(attlist, "%s%s",
|
|
(k == 0) ? "" : ", ", fmtId(attname, force_quotes));
|
|
else
|
|
{
|
|
if (k >= nclass)
|
|
{
|
|
fprintf(stderr, "dumpIndices(): OpClass not found for "
|
|
"attribute '%s' of index '%s'\n",
|
|
attname, indinfo[i].indexrelname);
|
|
exit_nicely(g_conn);
|
|
}
|
|
resetPQExpBuffer(id1);
|
|
resetPQExpBuffer(id2);
|
|
appendPQExpBuffer(id1, fmtId(attname, force_quotes));
|
|
appendPQExpBuffer(id2, fmtId(classname[k], force_quotes));
|
|
appendPQExpBuffer(attlist, "%s%s %s",
|
|
(k == 0) ? "" : ", ", id1->data, id2->data);
|
|
free(classname[k]);
|
|
}
|
|
}
|
|
|
|
if (!tablename || (strcmp(indinfo[i].indrelname, tablename) == 0) || (strlen(tablename) == 0) )
|
|
{
|
|
|
|
/*
|
|
* We make the index belong to the owner of its table, which
|
|
* is not necessarily right but should answer 99% of the time.
|
|
* Would have to add owner name to IndInfo to do it right.
|
|
*/
|
|
|
|
resetPQExpBuffer(id1);
|
|
resetPQExpBuffer(id2);
|
|
appendPQExpBuffer(id1, fmtId(indinfo[i].indexrelname, force_quotes));
|
|
appendPQExpBuffer(id2, fmtId(indinfo[i].indrelname, force_quotes));
|
|
|
|
resetPQExpBuffer(delq);
|
|
appendPQExpBuffer(delq, "DROP INDEX %s;\n", id1->data);
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "CREATE %s INDEX %s on %s using %s (",
|
|
(strcmp(indinfo[i].indisunique, "t") == 0) ? "UNIQUE" : "",
|
|
id1->data,
|
|
id2->data,
|
|
indinfo[i].indamname);
|
|
if (funcname)
|
|
{
|
|
/* need 2 printf's here cuz fmtId has static return area */
|
|
appendPQExpBuffer(q, " %s", fmtId(funcname, false));
|
|
appendPQExpBuffer(q, " (%s) %s );\n", attlist->data,
|
|
fmtId(classname[0], force_quotes));
|
|
free(funcname);
|
|
free(classname[0]);
|
|
}
|
|
else
|
|
appendPQExpBuffer(q, " %s );\n", attlist->data);
|
|
|
|
/* Dump Index Comments */
|
|
|
|
ArchiveEntry(fout, tblinfo[tableInd].oid, id1->data, "INDEX", NULL, q->data, delq->data,
|
|
"", tblinfo[tableInd].usename, NULL, NULL);
|
|
|
|
resetPQExpBuffer(q);
|
|
appendPQExpBuffer(q, "INDEX %s", id1->data);
|
|
dumpComment(fout, q->data, indinfo[i].indoid);
|
|
|
|
}
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
* dumpTuples
|
|
* prints out the tuples in ASCII representation. The output is a valid
|
|
* input to COPY FROM stdin.
|
|
*
|
|
* We only need to do this for POSTGRES 4.2 databases since the
|
|
* COPY TO statment doesn't escape newlines properly. It's been fixed
|
|
* in PostgreSQL.
|
|
*
|
|
* the attrmap passed in tells how to map the attributes copied in to the
|
|
* attributes copied out
|
|
*/
|
|
#ifdef NOT_USED
|
|
void
|
|
dumpTuples(PGresult *res, FILE *fout, int *attrmap)
|
|
{
|
|
int j,
|
|
k;
|
|
int m,
|
|
n;
|
|
char **outVals = NULL; /* values to copy out */
|
|
|
|
n = PQntuples(res);
|
|
m = PQnfields(res);
|
|
|
|
if (m > 0)
|
|
{
|
|
|
|
/*
|
|
* Print out the tuples but only print tuples with at least 1
|
|
* field.
|
|
*/
|
|
outVals = (char **) malloc(m * sizeof(char *));
|
|
|
|
for (j = 0; j < n; j++)
|
|
{
|
|
for (k = 0; k < m; k++)
|
|
outVals[attrmap[k]] = PQgetvalue(res, j, k);
|
|
for (k = 0; k < m; k++)
|
|
{
|
|
char *pval = outVals[k];
|
|
|
|
if (k != 0)
|
|
fputc('\t', fout); /* delimiter for attribute */
|
|
|
|
if (pval)
|
|
{
|
|
while (*pval != '\0')
|
|
{
|
|
/* escape tabs, newlines and backslashes */
|
|
if (*pval == '\t' || *pval == '\n' || *pval == '\\')
|
|
fputc('\\', fout);
|
|
fputc(*pval, fout);
|
|
pval++;
|
|
}
|
|
}
|
|
}
|
|
fputc('\n', fout); /* delimiter for a tuple */
|
|
}
|
|
free(outVals);
|
|
}
|
|
}
|
|
|
|
#endif
|
|
|
|
/*
|
|
* setMaxOid -
|
|
* find the maximum oid and generate a COPY statement to set it
|
|
*/
|
|
|
|
static void
|
|
setMaxOid(Archive *fout)
|
|
{
|
|
PGresult *res;
|
|
Oid max_oid;
|
|
char sql[1024];
|
|
int pos;
|
|
|
|
res = PQexec(g_conn, "CREATE TEMPORARY TABLE pgdump_oid (dummy int4)");
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_COMMAND_OK)
|
|
{
|
|
fprintf(stderr, "Can not create pgdump_oid table. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
PQclear(res);
|
|
res = PQexec(g_conn, "INSERT INTO pgdump_oid VALUES (0)");
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_COMMAND_OK)
|
|
{
|
|
fprintf(stderr, "Can not insert into pgdump_oid table. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
max_oid = atol(PQoidStatus(res));
|
|
if (max_oid == 0)
|
|
{
|
|
fprintf(stderr, "Invalid max id in setMaxOid\n");
|
|
exit_nicely(g_conn);
|
|
}
|
|
PQclear(res);
|
|
res = PQexec(g_conn, "DROP TABLE pgdump_oid;");
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_COMMAND_OK)
|
|
{
|
|
fprintf(stderr, "Can not drop pgdump_oid table. "
|
|
"Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
PQclear(res);
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s maximum system oid is %u %s\n",
|
|
g_comment_start, max_oid, g_comment_end);
|
|
pos = snprintf(sql, 1024, "CREATE TEMPORARY TABLE pgdump_oid (dummy int4);\n");
|
|
pos = pos + snprintf(sql+pos, 1024-pos, "COPY pgdump_oid WITH OIDS FROM stdin;\n");
|
|
pos = pos + snprintf(sql+pos, 1024-pos, "%-d\t0\n", max_oid);
|
|
pos = pos + snprintf(sql+pos, 1024-pos, "\\.\n");
|
|
pos = pos + snprintf(sql+pos, 1024-pos, "DROP TABLE pgdump_oid;\n");
|
|
|
|
ArchiveEntry(fout, "0", "Max OID", "<Init>", NULL, sql, "", "", "", NULL, NULL);
|
|
}
|
|
|
|
/*
|
|
* findLastBuiltInOid -
|
|
* find the last built in oid
|
|
* we do this by retrieving datlastsysoid from the pg_database entry for this database,
|
|
*/
|
|
|
|
static Oid
|
|
findLastBuiltinOid(const char* dbname)
|
|
{
|
|
PGresult *res;
|
|
int ntups;
|
|
Oid last_oid;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT datlastsysoid from pg_database where datname = ");
|
|
formatStringLiteral(query, dbname);
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (res == NULL ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "pg_dump: error in finding the last system OID");
|
|
fprintf(stderr, "Explanation from backend: '%s'.\n", PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
ntups = PQntuples(res);
|
|
if (ntups < 1)
|
|
{
|
|
fprintf(stderr, "pg_dump: couldn't find the pg_database entry.\n");
|
|
fprintf(stderr, "There is no entry in the 'pg_database' table for this database.\n");
|
|
exit_nicely(g_conn);
|
|
}
|
|
if (ntups > 1)
|
|
{
|
|
fprintf(stderr, "pg_dump: found more than one matching database.\n");
|
|
fprintf(stderr, "There is more than one entry for this database in the 'pg_database' table\n");
|
|
exit_nicely(g_conn);
|
|
}
|
|
last_oid = atoi(PQgetvalue(res, 0, PQfnumber(res, "datlastsysoid")));
|
|
PQclear(res);
|
|
return last_oid;
|
|
}
|
|
|
|
|
|
static void
|
|
dumpSequence(Archive *fout, TableInfo tbinfo)
|
|
{
|
|
PGresult *res;
|
|
int4 last,
|
|
incby,
|
|
maxv,
|
|
minv,
|
|
cache;
|
|
char cycled,
|
|
called;
|
|
const char *t;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
PQExpBuffer delqry = createPQExpBuffer();
|
|
|
|
appendPQExpBuffer(query,
|
|
"SELECT sequence_name, last_value, increment_by, max_value, "
|
|
"min_value, cache_value, is_cycled, is_called from %s",
|
|
fmtId(tbinfo.relname, force_quotes));
|
|
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res || PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpSequence(%s): SELECT failed. "
|
|
"Explanation from backend: '%s'.\n", tbinfo.relname, PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
if (PQntuples(res) != 1)
|
|
{
|
|
fprintf(stderr, "dumpSequence(%s): %d (!= 1) tuples returned by SELECT\n",
|
|
tbinfo.relname, PQntuples(res));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
if (strcmp(PQgetvalue(res, 0, 0), tbinfo.relname) != 0)
|
|
{
|
|
fprintf(stderr, "dumpSequence(%s): different sequence name "
|
|
"returned by SELECT: %s\n",
|
|
tbinfo.relname, PQgetvalue(res, 0, 0));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
|
|
last = atoi(PQgetvalue(res, 0, 1));
|
|
incby = atoi(PQgetvalue(res, 0, 2));
|
|
maxv = atoi(PQgetvalue(res, 0, 3));
|
|
minv = atoi(PQgetvalue(res, 0, 4));
|
|
cache = atoi(PQgetvalue(res, 0, 5));
|
|
t = PQgetvalue(res, 0, 6);
|
|
cycled = *t;
|
|
t = PQgetvalue(res, 0, 7);
|
|
called = *t;
|
|
|
|
PQclear(res);
|
|
|
|
resetPQExpBuffer(delqry);
|
|
appendPQExpBuffer(delqry, "DROP SEQUENCE %s;\n", fmtId(tbinfo.relname, force_quotes));
|
|
|
|
/*
|
|
* The logic we use for restoring sequences is as follows:
|
|
* - Add a basic CREATE SEQUENCE statement
|
|
* (use last_val for start if called == 'f', else use min_val for start_val).
|
|
* - Add a 'SETVAL(seq, last_val, iscalled)' at restore-time iff we load data
|
|
*/
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query,
|
|
"CREATE SEQUENCE %s start %d increment %d maxvalue %d "
|
|
"minvalue %d cache %d %s;\n",
|
|
fmtId(tbinfo.relname, force_quotes),
|
|
(called == 't') ? minv : last,
|
|
incby, maxv, minv, cache,
|
|
(cycled == 't') ? "cycle" : "");
|
|
|
|
ArchiveEntry(fout, tbinfo.oid, fmtId(tbinfo.relname, force_quotes), "SEQUENCE", NULL,
|
|
query->data, delqry->data, "", tbinfo.usename, NULL, NULL);
|
|
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT setval (");
|
|
formatStringLiteral(query, fmtId(tbinfo.relname, force_quotes));
|
|
appendPQExpBuffer(query, ", %d, '%c');\n", last, called);
|
|
|
|
ArchiveEntry(fout, tbinfo.oid, fmtId(tbinfo.relname, force_quotes), "SEQUENCE SET", NULL,
|
|
query->data, "" /* Del */, "", "", NULL, NULL);
|
|
|
|
/* Dump Sequence Comments */
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SEQUENCE %s", fmtId(tbinfo.relname, force_quotes));
|
|
dumpComment(fout, query->data, tbinfo.oid);
|
|
|
|
}
|
|
|
|
|
|
static void
|
|
dumpTriggers(Archive *fout, const char *tablename,
|
|
TableInfo *tblinfo, int numTables)
|
|
{
|
|
int i,
|
|
j;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s dumping out triggers %s\n",
|
|
g_comment_start, g_comment_end);
|
|
|
|
for (i = 0; i < numTables; i++)
|
|
{
|
|
if (tablename && (strcmp(tblinfo[i].relname, tablename) != 0) && (strlen(tablename) > 0) )
|
|
continue;
|
|
|
|
for (j = 0; j < tblinfo[i].ntrig; j++)
|
|
{
|
|
ArchiveEntry(fout, tblinfo[i].triggers[j].oid, tblinfo[i].triggers[j].tgname,
|
|
"TRIGGER", NULL, tblinfo[i].triggers[j].tgsrc, "", "",
|
|
tblinfo[i].usename, NULL, NULL);
|
|
dumpComment(fout, tblinfo[i].triggers[j].tgcomment, tblinfo[i].triggers[j].oid);
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
static void
|
|
dumpRules(Archive *fout, const char *tablename,
|
|
TableInfo *tblinfo, int numTables)
|
|
{
|
|
PGresult *res;
|
|
int nrules;
|
|
int i,
|
|
t;
|
|
PQExpBuffer query = createPQExpBuffer();
|
|
|
|
int i_definition;
|
|
int i_oid;
|
|
int i_owner;
|
|
int i_rulename;
|
|
|
|
if (g_verbose)
|
|
fprintf(stderr, "%s dumping out rules %s\n",
|
|
g_comment_start, g_comment_end);
|
|
|
|
/*
|
|
* For each table we dump
|
|
*/
|
|
for (t = 0; t < numTables; t++)
|
|
{
|
|
if (tablename && (strcmp(tblinfo[t].relname, tablename) != 0) && (strlen(tablename) > 0) )
|
|
continue;
|
|
|
|
/*
|
|
* Get all rules defined for this table
|
|
* We include pg_rules in the cross since it filters out
|
|
* all view rules (pjw 15-Sep-2000).
|
|
*
|
|
* XXXX: Use LOJ here
|
|
*/
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "SELECT definition,"
|
|
" (select usename from pg_user where pg_class.relowner = usesysid) AS viewowner, "
|
|
" pg_rewrite.oid, pg_rewrite.rulename "
|
|
"FROM pg_rewrite, pg_class, pg_rules "
|
|
"WHERE pg_class.relname = ");
|
|
formatStringLiteral(query, tblinfo[t].relname);
|
|
appendPQExpBuffer(query,
|
|
" AND pg_rewrite.ev_class = pg_class.oid "
|
|
" AND pg_rules.tablename = pg_class.relname "
|
|
" AND pg_rules.rulename = pg_rewrite.rulename "
|
|
"ORDER BY pg_rewrite.oid");
|
|
res = PQexec(g_conn, query->data);
|
|
if (!res ||
|
|
PQresultStatus(res) != PGRES_TUPLES_OK)
|
|
{
|
|
fprintf(stderr, "dumpRules(): SELECT failed for rules associated with table \"%s\".\n\tExplanation from backend: '%s'.\n",
|
|
tblinfo[t].relname, PQerrorMessage(g_conn));
|
|
exit_nicely(g_conn);
|
|
}
|
|
|
|
nrules = PQntuples(res);
|
|
i_definition = PQfnumber(res, "definition");
|
|
i_owner = PQfnumber(res, "viewowner");
|
|
i_oid = PQfnumber(res, "oid");
|
|
i_rulename = PQfnumber(res, "rulename");
|
|
|
|
/*
|
|
* Dump them out
|
|
*/
|
|
|
|
for (i = 0; i < nrules; i++)
|
|
{
|
|
ArchiveEntry(fout, PQgetvalue(res, i, i_oid), PQgetvalue(res, i, i_rulename),
|
|
"RULE", NULL, PQgetvalue(res, i, i_definition),
|
|
"", "", PQgetvalue(res, i, i_owner), NULL, NULL);
|
|
|
|
/* Dump rule comments */
|
|
|
|
resetPQExpBuffer(query);
|
|
appendPQExpBuffer(query, "RULE %s", fmtId(PQgetvalue(res, i, i_rulename), force_quotes));
|
|
dumpComment(fout, query->data, PQgetvalue(res, i, i_oid));
|
|
|
|
}
|
|
|
|
PQclear(res);
|
|
}
|
|
}
|
|
|