| /*------------------------------------------------------------------------- |
| * |
| * pg_dump.c |
| * pg_dump is a utility for dumping out a postgres database |
| * into a script file. |
| * |
| <<<<<<< HEAD |
| * Portions Copyright (c) 2005-2010, Greenplum inc |
| * Portions Copyright (c) 2012-Present VMware, Inc. or its affiliates. |
| * Portions Copyright (c) 1996-2021, PostgreSQL Global Development Group |
| ======= |
| * Portions Copyright (c) 1996-2023, PostgreSQL Global Development Group |
| >>>>>>> REL_16_9 |
| * Portions Copyright (c) 1994, Regents of the University of California |
| * |
| * pg_dump will read the system catalogs in a database and dump out a |
| * script that reproduces the schema in terms of SQL that is understood |
| * by PostgreSQL |
| * |
| * Note that pg_dump runs in a transaction-snapshot mode transaction, |
| * so it sees a consistent snapshot of the database including system |
| * catalogs. However, it relies in part on various specialized backend |
| * functions like pg_get_indexdef(), and those things tend to look at |
| * the currently committed state. So it is possible to get 'cache |
| * lookup failed' error if someone performs DDL changes while a dump is |
| * happening. The window for this sort of thing is from the acquisition |
| * of the transaction snapshot to getSchemaData() (when pg_dump acquires |
| * AccessShareLock on every table it intends to dump). It isn't very large, |
| * but it can happen. |
| * |
| * http://archives.postgresql.org/pgsql-bugs/2010-02/msg00187.php |
| * |
| * IDENTIFICATION |
| * src/bin/pg_dump/pg_dump.c |
| * |
| *------------------------------------------------------------------------- |
| */ |
| #include "postgres_fe.h" |
| |
| #include <unistd.h> |
| #include <ctype.h> |
| #include <limits.h> |
| #ifdef HAVE_TERMIOS_H |
| #include <termios.h> |
| #endif |
| |
| #include "access/attnum.h" |
| #include "access/sysattr.h" |
| #include "access/transam.h" |
| #include "catalog/pg_aggregate_d.h" |
| #include "catalog/pg_am_d.h" |
| #include "catalog/pg_attribute_d.h" |
| #include "catalog/pg_authid_d.h" |
| #include "catalog/pg_cast_d.h" |
| #include "catalog/pg_class_d.h" |
| #include "catalog/pg_default_acl_d.h" |
| #include "catalog/pg_foreign_server.h" |
| #include "catalog/pg_foreign_server_d.h" |
| #include "catalog/pg_largeobject_d.h" |
| #include "catalog/pg_largeobject_metadata_d.h" |
| #include "catalog/pg_namespace_d.h" |
| #include "catalog/pg_proc_d.h" |
| #include "catalog/pg_subscription.h" |
| #include "catalog/pg_trigger_d.h" |
| #include "catalog/pg_type_d.h" |
| #include "common/connect.h" |
| #include "common/relpath.h" |
| #include "compress_io.h" |
| #include "dumputils.h" |
| #include "fe_utils/option_utils.h" |
| #include "fe_utils/string_utils.h" |
| #include "getopt_long.h" |
| #include "libpq/libpq-fs.h" |
| #include "lib/stringinfo.h" |
| #include "parallel.h" |
| #include "pg_backup_db.h" |
| #include "pg_backup_utils.h" |
| #include "pg_dump.h" |
| #include "storage/block.h" |
| |
| typedef struct |
| { |
| Oid roleoid; /* role's OID */ |
| const char *rolename; /* role's name */ |
| } RoleNameItem; |
| |
| typedef struct |
| { |
| const char *descr; /* comment for an object */ |
| Oid classoid; /* object class (catalog OID) */ |
| Oid objoid; /* object OID */ |
| int objsubid; /* subobject (table column #) */ |
| } CommentItem; |
| |
| typedef struct |
| { |
| const char *provider; /* label provider of this security label */ |
| const char *label; /* security label for an object */ |
| Oid classoid; /* object class (catalog OID) */ |
| Oid objoid; /* object OID */ |
| int objsubid; /* subobject (table column #) */ |
| } SecLabelItem; |
| |
| typedef enum OidOptions |
| { |
| zeroIsError = 1, |
| zeroAsStar = 2, |
| zeroAsNone = 4 |
| } OidOptions; |
| |
| /* global decls */ |
| static bool dosync = true; /* Issue fsync() to make dump durable on disk. */ |
| |
| <<<<<<< HEAD |
| /* GPDB_95_MERGE_FIXME: put those flags to DumpOptions to avoid using global variables */ |
| /* START MPP ADDITION */ |
| bool dumpGpPolicy; |
| bool isGPbackend; |
| |
| /* END MPP ADDITION */ |
| |
| /* subquery used to convert user ID (eg, datdba) to user name */ |
| static const char *username_subquery; |
| |
| /* |
| * For 8.0 and earlier servers, pulled from pg_database, for 8.1+ we use |
| * FirstNormalObjectId - 1. |
| */ |
| ======= |
| >>>>>>> REL_16_9 |
| static Oid g_last_builtin_oid; /* value of the last builtin oid */ |
| |
| /* The specified names/patterns should to match at least one entity */ |
| static int strict_names = 0; |
| |
| static pg_compress_algorithm compression_algorithm = PG_COMPRESSION_NONE; |
| |
| /* |
| * Object inclusion/exclusion lists |
| * |
| * The string lists record the patterns given by command-line switches, |
| * which we then convert to lists of OIDs of matching objects. |
| */ |
| static SimpleStringList schema_include_patterns = {NULL, NULL}; |
| static SimpleOidList schema_include_oids = {NULL, NULL}; |
| static SimpleStringList schema_exclude_patterns = {NULL, NULL}; |
| static SimpleOidList schema_exclude_oids = {NULL, NULL}; |
| |
| static SimpleStringList table_include_patterns = {NULL, NULL}; |
| static SimpleStringList table_include_patterns_and_children = {NULL, NULL}; |
| static SimpleOidList table_include_oids = {NULL, NULL}; |
| static SimpleStringList table_exclude_patterns = {NULL, NULL}; |
| static SimpleStringList table_exclude_patterns_and_children = {NULL, NULL}; |
| static SimpleOidList table_exclude_oids = {NULL, NULL}; |
| static SimpleStringList tabledata_exclude_patterns = {NULL, NULL}; |
| static SimpleStringList tabledata_exclude_patterns_and_children = {NULL, NULL}; |
| static SimpleOidList tabledata_exclude_oids = {NULL, NULL}; |
| |
| <<<<<<< HEAD |
| static SimpleStringList relid_string_list = {NULL, NULL}; |
| static SimpleStringList funcid_string_list = {NULL, NULL}; |
| static SimpleOidList function_include_oids = {NULL, NULL}; |
| |
| static SimpleOidList preassigned_oids = {NULL, NULL}; |
| |
| ======= |
| >>>>>>> REL_16_9 |
| static SimpleStringList foreign_servers_include_patterns = {NULL, NULL}; |
| static SimpleOidList foreign_servers_include_oids = {NULL, NULL}; |
| |
| static SimpleStringList extension_include_patterns = {NULL, NULL}; |
| static SimpleOidList extension_include_oids = {NULL, NULL}; |
| |
| static const CatalogId nilCatalogId = {0, 0}; |
| |
| const char *EXT_PARTITION_NAME_POSTFIX = "_external_partition__"; |
| |
| /* pg_class.relstorage value used in GPDB 6.x and below to mark external tables. */ |
| #define RELSTORAGE_EXTERNAL 'x' |
| |
| /* flag indicating whether or not this GP database supports partitioning */ |
| static bool gp_partitioning_available = false; |
| |
| static DumpId binary_upgrade_dumpid; |
| |
| /* override for standard extra_float_digits setting */ |
| static bool have_extra_float_digits = false; |
| static int extra_float_digits; |
| |
| /* sorted table of role names */ |
| static RoleNameItem *rolenames = NULL; |
| static int nrolenames = 0; |
| |
| /* sorted table of comments */ |
| static CommentItem *comments = NULL; |
| static int ncomments = 0; |
| |
| /* sorted table of security labels */ |
| static SecLabelItem *seclabels = NULL; |
| static int nseclabels = 0; |
| |
| /* |
| * The default number of rows per INSERT when |
| * --inserts is specified without --rows-per-insert |
| */ |
| #define DUMP_DEFAULT_ROWS_PER_INSERT 1 |
| |
| /* |
| * FIXME: CBDB should not know the am oid of PAX. We put here because the kernel |
| * can't distinguish the PAX and renamed heap(heap_psql) in test `psql`. |
| * The definition of temporary is here and should be consistent with util/rel.h |
| */ |
| #define PAX_AM_OID 7047 |
| |
| /* |
| * Macro for producing quoted, schema-qualified name of a dumpable object. |
| */ |
| #define fmtQualifiedDumpable(obj) \ |
| fmtQualifiedId((obj)->dobj.namespace->dobj.name, \ |
| (obj)->dobj.name) |
| |
| static void help(const char *progname); |
| static void setup_connection(Archive *AH, |
| const char *dumpencoding, const char *dumpsnapshot, |
| char *use_role); |
| static ArchiveFormat parseArchiveFormat(const char *format, ArchiveMode *mode); |
| static void expand_schema_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids, |
| bool strict_names); |
| static void expand_extension_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids, |
| bool strict_names); |
| static void expand_foreign_server_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids); |
| static void expand_table_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids, |
| <<<<<<< HEAD |
| bool strict_names); |
| ======= |
| bool strict_names, |
| bool with_child_tables); |
| >>>>>>> REL_16_9 |
| static void prohibit_crossdb_refs(PGconn *conn, const char *dbname, |
| const char *pattern); |
| |
| static NamespaceInfo *findNamespace(Oid nsoid); |
| static void dumpTableData(Archive *fout, const TableDataInfo *tdinfo); |
| static void refreshMatViewData(Archive *fout, const TableDataInfo *tdinfo); |
| static const char *getRoleName(const char *roleoid_str); |
| static void collectRoleNames(Archive *fout); |
| static void getAdditionalACLs(Archive *fout); |
| static void dumpCommentExtended(Archive *fout, const char *type, |
| const char *name, const char *namespace, |
| const char *owner, CatalogId catalogId, |
| int subid, DumpId dumpId, |
| const char *initdb_comment); |
| static inline void dumpComment(Archive *fout, const char *type, |
| const char *name, const char *namespace, |
| const char *owner, CatalogId catalogId, |
| int subid, DumpId dumpId); |
| static int findComments(Oid classoid, Oid objoid, CommentItem **items); |
| static void collectComments(Archive *fout); |
| static void dumpSecLabel(Archive *fout, const char *type, const char *name, |
| const char *namespace, const char *owner, |
| CatalogId catalogId, int subid, DumpId dumpId); |
| static int findSecLabels(Oid classoid, Oid objoid, SecLabelItem **items); |
| static void collectSecLabels(Archive *fout); |
| static void dumpDumpableObject(Archive *fout, DumpableObject *dobj); |
| static void dumpNamespace(Archive *fout, const NamespaceInfo *nspinfo); |
| static void dumpExtension(Archive *fout, const ExtensionInfo *extinfo); |
| static void dumpType(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpBaseType(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpEnumType(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpRangeType(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpUndefinedType(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpDomain(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpCompositeType(Archive *fout, const TypeInfo *tyinfo); |
| static void dumpCompositeTypeColComments(Archive *fout, const TypeInfo *tyinfo, |
| PGresult *res); |
| static void dumpShellType(Archive *fout, const ShellTypeInfo *stinfo); |
| static void dumpProcLang(Archive *fout, const ProcLangInfo *plang); |
| static void dumpFunc(Archive *fout, const FuncInfo *finfo); |
| static void dumpCast(Archive *fout, const CastInfo *cast); |
| static void dumpTransform(Archive *fout, const TransformInfo *transform); |
| static void dumpOpr(Archive *fout, const OprInfo *oprinfo); |
| static void dumpAccessMethod(Archive *fout, const AccessMethodInfo *aminfo); |
| static void dumpOpclass(Archive *fout, const OpclassInfo *opcinfo); |
| static void dumpOpfamily(Archive *fout, const OpfamilyInfo *opfinfo); |
| static void dumpCollation(Archive *fout, const CollInfo *collinfo); |
| static void dumpConversion(Archive *fout, const ConvInfo *convinfo); |
| static void dumpRule(Archive *fout, const RuleInfo *rinfo); |
| static void dumpAgg(Archive *fout, const AggInfo *agginfo); |
| static void dumpTrigger(Archive *fout, const TriggerInfo *tginfo); |
| static void dumpEventTrigger(Archive *fout, const EventTriggerInfo *evtinfo); |
| static void dumpTable(Archive *fout, const TableInfo *tbinfo); |
| static void dumpTableSchema(Archive *fout, const TableInfo *tbinfo); |
| static void dumpTableAttach(Archive *fout, const TableAttachInfo *attachinfo); |
| static void dumpAttrDef(Archive *fout, const AttrDefInfo *adinfo); |
| static void dumpSequence(Archive *fout, const TableInfo *tbinfo); |
| static void dumpSequenceData(Archive *fout, const TableDataInfo *tdinfo); |
| static void dumpIndex(Archive *fout, const IndxInfo *indxinfo); |
| static void dumpIndexAttach(Archive *fout, const IndexAttachInfo *attachinfo); |
| static void dumpStatisticsExt(Archive *fout, const StatsExtInfo *statsextinfo); |
| static void dumpConstraint(Archive *fout, const ConstraintInfo *coninfo); |
| static void dumpTableConstraintComment(Archive *fout, const ConstraintInfo *coninfo); |
| static void dumpTSParser(Archive *fout, const TSParserInfo *prsinfo); |
| static void dumpTSDictionary(Archive *fout, const TSDictInfo *dictinfo); |
| static void dumpTSTemplate(Archive *fout, const TSTemplateInfo *tmplinfo); |
| static void dumpTSConfig(Archive *fout, const TSConfigInfo *cfginfo); |
| static void dumpForeignDataWrapper(Archive *fout, const FdwInfo *fdwinfo); |
| static void dumpForeignServer(Archive *fout, const ForeignServerInfo *srvinfo); |
| |
| /* GPDB follow upstream style */ |
| static void dumpExtProtocol(Archive *fout, const ExtProtInfo *ptcinfo); |
| static void dumpTypeStorageOptions(Archive *fout, const TypeStorageOptions *tstorageoptions); |
| |
| static void dumpUserMappings(Archive *fout, |
| const char *servername, const char *namespace, |
| const char *owner, CatalogId catalogId, DumpId dumpId); |
| static void dumpDefaultACL(Archive *fout, const DefaultACLInfo *daclinfo); |
| |
| static DumpId dumpACL(Archive *fout, DumpId objDumpId, DumpId altDumpId, |
| const char *type, const char *name, const char *subname, |
| const char *nspname, const char *owner, |
| const DumpableAcl *dacl); |
| |
| static void getDependencies(Archive *fout); |
| static void BuildArchiveDependencies(Archive *fout); |
| static void findDumpableDependencies(ArchiveHandle *AH, const DumpableObject *dobj, |
| DumpId **dependencies, int *nDeps, int *allocDeps); |
| |
| static DumpableObject *createBoundaryObjects(void); |
| static void addBoundaryDependencies(DumpableObject **dobjs, int numObjs, |
| DumpableObject *boundaryObjs); |
| |
| static void addConstrChildIdxDeps(DumpableObject *dobj, const IndxInfo *refidx); |
| static void getDomainConstraints(Archive *fout, TypeInfo *tyinfo); |
| static void getTableData(DumpOptions *dopt, TableInfo *tblinfo, int numTables, char relkind); |
| static void makeTableDataInfo(DumpOptions *dopt, TableInfo *tbinfo); |
| static void buildMatViewRefreshDependencies(Archive *fout); |
| static void getTableDataFKConstraints(void); |
| static char *format_function_arguments(const FuncInfo *finfo, const char *funcargs, |
| bool is_agg); |
| static char *format_function_signature(Archive *fout, |
| const FuncInfo *finfo, bool honor_quotes); |
| static char *convertRegProcReference(const char *proc); |
| static char *getFormattedOperatorName(const char *oproid); |
| static char *convertTSFunction(Archive *fout, Oid funcOid); |
| <<<<<<< HEAD |
| static Oid findLastBuiltinOid_V71(Archive *fout); |
| static const char *getFormattedTypeName(Archive *fout, Oid oid, OidOptions opts); |
| static void getBlobs(Archive *fout); |
| static void dumpBlob(Archive *fout, const BlobInfo *binfo); |
| static int dumpBlobs(Archive *fout, const void *arg); |
| static void dumpPreassignedOidArchiveEntry(Archive *fout, BinaryUpgradeInfo *binfo); |
| static void dumpPreassignedOidDefinition(Archive *fout, BinaryUpgradeInfo *binfo); |
| ======= |
| static const char *getFormattedTypeName(Archive *fout, Oid oid, OidOptions opts); |
| static void getLOs(Archive *fout); |
| static void dumpLO(Archive *fout, const LoInfo *loinfo); |
| static int dumpLOs(Archive *fout, const void *arg); |
| >>>>>>> REL_16_9 |
| static void dumpPolicy(Archive *fout, const PolicyInfo *polinfo); |
| static void dumpPublication(Archive *fout, const PublicationInfo *pubinfo); |
| static void dumpPublicationTable(Archive *fout, const PublicationRelInfo *pubrinfo); |
| static void dumpSubscription(Archive *fout, const SubscriptionInfo *subinfo); |
| static void dumpDatabase(Archive *fout); |
| static void dumpDatabaseConfig(Archive *AH, PQExpBuffer outbuf, |
| const char *dbname, Oid dboid); |
| static void dumpEncoding(Archive *AH); |
| static void dumpStdStrings(Archive *AH); |
| static void binary_upgrade_set_namespace_oid(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_namespace_oid); |
| static void binary_upgrade_set_type_oids_for_ao(Archive *fout, |
| PQExpBuffer upgrade_buffer, Oid pg_rel_oid, |
| char *ao_aux_typname); |
| static void binary_upgrade_set_type_oids_by_rel_oid_impl(Archive *fout, |
| PQExpBuffer upgrade_buffer, Oid pg_rel_oid, |
| char *typname_override); |
| static void binary_upgrade_set_pg_class_oids_for_ao(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_class_oid, bool is_index, |
| char *ao_aux_relname); |
| static void binary_upgrade_set_pg_class_oids_impl(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_class_oid, bool is_index, |
| char *relname_override); |
| static void dumpSearchPath(Archive *AH); |
| static void binary_upgrade_set_type_oids_by_type_oid(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_type_oid, Oid pg_type_ns_oid, |
| char *pg_type_name, |
| bool force_array_type, |
| bool include_multirange_type); |
| static void binary_upgrade_set_type_oids_by_rel(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| const TableInfo *tbinfo); |
| static void binary_upgrade_set_pg_class_oids(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_class_oid, bool is_index); |
| static void binary_upgrade_extension_member(PQExpBuffer upgrade_buffer, |
| const DumpableObject *dobj, |
| const char *objtype, |
| const char *objname, |
| const char *objnamespace); |
| static const char *getAttrName(int attrnum, const TableInfo *tblInfo); |
| static const char *fmtCopyColumnList(const TableInfo *ti, PQExpBuffer buffer); |
| static bool nonemptyReloptions(const char *reloptions); |
| static void appendReloptionsArrayAH(PQExpBuffer buffer, const char *reloptions, |
| const char *prefix, Archive *fout); |
| static char *get_synchronized_snapshot(Archive *fout); |
| static void set_restrict_relation_kind(Archive *AH, const char *value); |
| static void setupDumpWorker(Archive *AH); |
| static TableInfo *getRootTableInfo(const TableInfo *tbinfo); |
| static bool forcePartitionRootLoad(const TableInfo *tbinfo); |
| |
| |
| /* START MPP ADDITION */ |
| static void setExtPartDependency(TableInfo *tblinfo, int numTables); |
| static char *format_table_function_columns(Archive *fout, const FuncInfo *finfo, int nallargs, |
| char **allargtypes, |
| char **argmodes, |
| char **argnames); |
| static void expand_oid_patterns(SimpleStringList *patterns, |
| SimpleOidList *oids); |
| |
| static bool is_returns_table_function(int nallargs, char **argmodes); |
| static bool testGPbackend(Archive *fout); |
| static bool testPartitioningSupport(Archive *fout); |
| |
| static char *nextToken(register char **stringp, register const char *delim); |
| static void addDistributedBy(Archive *fout, PQExpBuffer q, const TableInfo *tbinfo, int actual_atts); |
| static void addDistributedByOld(Archive *fout, PQExpBuffer q, const TableInfo *tbinfo, int actual_atts); |
| static void addSchedule(Archive *fout, PQExpBuffer q, const TableInfo *tbinfo); |
| static bool isGPDB(Archive *fout); |
| static bool isGPDB5000OrLater(Archive *fout); |
| static bool isGPDB6000OrLater(Archive *fout); |
| |
| /* END MPP ADDITION */ |
| |
| /* |
| * Check if we are talking to GPDB |
| */ |
| static bool |
| isGPDB(Archive *fout) |
| { |
| static int value = -1; /* -1 = not known yet, 0 = no, 1 = yes */ |
| |
| /* Query the server on first call, and cache the result */ |
| if (value == -1) |
| { |
| const char *query = "select pg_catalog.version()"; |
| PGresult *res; |
| char *ver; |
| |
| res = ExecuteSqlQuery(fout, query, PGRES_TUPLES_OK); |
| |
| ver = (PQgetvalue(res, 0, 0)); |
| if (strstr(ver, "Cloudberry") != NULL) |
| value = 1; |
| else |
| value = 0; |
| |
| PQclear(res); |
| } |
| return (value == 1) ? true : false; |
| } |
| |
| |
| static bool |
| isGPDB5000OrLater(Archive *fout) |
| { |
| if (!isGPDB(fout)) |
| return false; /* Not Cloudberry at all. */ |
| |
| /* GPDB 5 is based on PostgreSQL 8.3 */ |
| return fout->remoteVersion >= 80300; |
| } |
| |
| |
| static bool |
| isGPDB6000OrLater(Archive *fout) |
| { |
| if (!isGPDB(fout)) |
| return false; /* Not Cloudberry at all. */ |
| |
| /* GPDB 6 is based on PostgreSQL 9.4 */ |
| return fout->remoteVersion >= 90400; |
| } |
| |
| int |
| main(int argc, char **argv) |
| { |
| int c; |
| const char *filename = NULL; |
| const char *format = "p"; |
| TableInfo *tblinfo; |
| int numTables; |
| DumpableObject **dobjs; |
| int numObjs; |
| DumpableObject *boundaryObjs; |
| int i; |
| int optindex; |
| RestoreOptions *ropt; |
| Archive *fout; /* the script file */ |
| bool g_verbose = false; |
| const char *dumpencoding = NULL; |
| const char *dumpsnapshot = NULL; |
| char *use_role = NULL; |
| int numWorkers = 1; |
| int plainText = 0; |
| ArchiveFormat archiveFormat = archUnknown; |
| ArchiveMode archiveMode; |
| pg_compress_specification compression_spec = {0}; |
| char *compression_detail = NULL; |
| char *compression_algorithm_str = "none"; |
| char *error_detail = NULL; |
| bool user_compression_defined = false; |
| |
| static DumpOptions dopt; |
| |
| /* |
| * The default value for gp_syntax_option depends upon whether or not the |
| * backend is a GP or non-GP backend -- a GP backend defaults to ENABLED. |
| */ |
| static enum |
| { |
| GPS_NOT_SPECIFIED, GPS_DISABLED, GPS_ENABLED |
| } gp_syntax_option = GPS_NOT_SPECIFIED; |
| |
| static struct option long_options[] = { |
| {"data-only", no_argument, NULL, 'a'}, |
| {"blobs", no_argument, NULL, 'b'}, |
| {"large-objects", no_argument, NULL, 'b'}, |
| {"no-blobs", no_argument, NULL, 'B'}, |
| {"no-large-objects", no_argument, NULL, 'B'}, |
| {"clean", no_argument, NULL, 'c'}, |
| {"create", no_argument, NULL, 'C'}, |
| {"dbname", required_argument, NULL, 'd'}, |
| {"extension", required_argument, NULL, 'e'}, |
| {"file", required_argument, NULL, 'f'}, |
| {"format", required_argument, NULL, 'F'}, |
| {"host", required_argument, NULL, 'h'}, |
| {"jobs", 1, NULL, 'j'}, |
| {"no-reconnect", no_argument, NULL, 'R'}, |
| {"no-owner", no_argument, NULL, 'O'}, |
| {"port", required_argument, NULL, 'p'}, |
| {"schema", required_argument, NULL, 'n'}, |
| {"exclude-schema", required_argument, NULL, 'N'}, |
| {"schema-only", no_argument, NULL, 's'}, |
| {"superuser", required_argument, NULL, 'S'}, |
| {"table", required_argument, NULL, 't'}, |
| {"exclude-table", required_argument, NULL, 'T'}, |
| {"no-password", no_argument, NULL, 'w'}, |
| {"password", no_argument, NULL, 'W'}, |
| {"username", required_argument, NULL, 'U'}, |
| {"verbose", no_argument, NULL, 'v'}, |
| {"no-privileges", no_argument, NULL, 'x'}, |
| {"no-acl", no_argument, NULL, 'x'}, |
| {"compress", required_argument, NULL, 'Z'}, |
| {"encoding", required_argument, NULL, 'E'}, |
| {"help", no_argument, NULL, '?'}, |
| {"version", no_argument, NULL, 'V'}, |
| |
| /* |
| * the following options don't have an equivalent short option letter |
| */ |
| {"attribute-inserts", no_argument, &dopt.column_inserts, 1}, |
| {"binary-upgrade", no_argument, &dopt.binary_upgrade, 1}, |
| {"column-inserts", no_argument, &dopt.column_inserts, 1}, |
| {"disable-dollar-quoting", no_argument, &dopt.disable_dollar_quoting, 1}, |
| {"disable-triggers", no_argument, &dopt.disable_triggers, 1}, |
| {"enable-row-security", no_argument, &dopt.enable_row_security, 1}, |
| {"exclude-table-data", required_argument, NULL, 4}, |
| {"extra-float-digits", required_argument, NULL, 8}, |
| {"if-exists", no_argument, &dopt.if_exists, 1}, |
| {"inserts", no_argument, NULL, 9}, |
| {"lock-wait-timeout", required_argument, NULL, 2}, |
| {"no-table-access-method", no_argument, &dopt.outputNoTableAm, 1}, |
| {"no-tablespaces", no_argument, &dopt.outputNoTablespaces, 1}, |
| {"quote-all-identifiers", no_argument, "e_all_identifiers, 1}, |
| {"load-via-partition-root", no_argument, &dopt.load_via_partition_root, 1}, |
| {"role", required_argument, NULL, 3}, |
| {"section", required_argument, NULL, 5}, |
| {"serializable-deferrable", no_argument, &dopt.serializable_deferrable, 1}, |
| {"snapshot", required_argument, NULL, 6}, |
| {"strict-names", no_argument, &strict_names, 1}, |
| {"use-set-session-authorization", no_argument, &dopt.use_setsessauth, 1}, |
| {"no-comments", no_argument, &dopt.no_comments, 1}, |
| {"no-publications", no_argument, &dopt.no_publications, 1}, |
| {"no-security-labels", no_argument, &dopt.no_security_labels, 1}, |
| {"no-subscriptions", no_argument, &dopt.no_subscriptions, 1}, |
| {"no-toast-compression", no_argument, &dopt.no_toast_compression, 1}, |
| {"no-unlogged-table-data", no_argument, &dopt.no_unlogged_table_data, 1}, |
| {"no-sync", no_argument, NULL, 7}, |
| {"on-conflict-do-nothing", no_argument, &dopt.do_nothing, 1}, |
| {"rows-per-insert", required_argument, NULL, 10}, |
| {"include-foreign-data", required_argument, NULL, 11}, |
| {"table-and-children", required_argument, NULL, 12}, |
| {"exclude-table-and-children", required_argument, NULL, 13}, |
| {"exclude-table-data-and-children", required_argument, NULL, 14}, |
| |
| /* START MPP ADDITION */ |
| |
| /* |
| * the following are mpp specific, and don't have an equivalent short |
| * option |
| */ |
| {"gp-syntax", no_argument, NULL, 1000}, |
| {"no-gp-syntax", no_argument, NULL, 1001}, |
| {"function-oids", required_argument, NULL, 1002}, |
| {"relation-oids", required_argument, NULL, 1003}, |
| /* END MPP ADDITION */ |
| {NULL, 0, NULL, 0} |
| }; |
| |
| pg_logging_init(argv[0]); |
| pg_logging_set_level(PG_LOG_WARNING); |
| set_pglocale_pgservice(argv[0], PG_TEXTDOMAIN("pg_dump")); |
| |
| /* |
| * Initialize what we need for parallel execution, especially for thread |
| * support on Windows. |
| */ |
| init_parallel_dump_utils(); |
| |
| progname = get_progname(argv[0]); |
| |
| if (argc > 1) |
| { |
| if (strcmp(argv[1], "--help") == 0 || strcmp(argv[1], "-?") == 0) |
| { |
| help(progname); |
| exit_nicely(0); |
| } |
| if (strcmp(argv[1], "--version") == 0 || strcmp(argv[1], "-V") == 0) |
| { |
| puts("pg_dump (Apache Cloudberry) " PG_VERSION); |
| exit_nicely(0); |
| } |
| } |
| |
| InitDumpOptions(&dopt); |
| |
| while ((c = getopt_long(argc, argv, "abBcCd:e:E:f:F:h:j:n:N:Op:RsS:t:T:U:vwWxZ:", |
| long_options, &optindex)) != -1) |
| { |
| switch (c) |
| { |
| case 'a': /* Dump data only */ |
| dopt.dataOnly = true; |
| break; |
| |
| case 'b': /* Dump LOs */ |
| dopt.outputLOs = true; |
| break; |
| |
| case 'B': /* Don't dump LOs */ |
| dopt.dontOutputLOs = true; |
| break; |
| |
| case 'c': /* clean (i.e., drop) schema prior to create */ |
| dopt.outputClean = 1; |
| break; |
| |
| case 'C': /* Create DB */ |
| dopt.outputCreateDB = 1; |
| break; |
| |
| case 'd': /* database name */ |
| dopt.cparams.dbname = pg_strdup(optarg); |
| break; |
| |
| case 'e': /* include extension(s) */ |
| simple_string_list_append(&extension_include_patterns, optarg); |
| dopt.include_everything = false; |
| break; |
| |
| case 'E': /* Dump encoding */ |
| dumpencoding = pg_strdup(optarg); |
| break; |
| |
| case 'f': |
| filename = pg_strdup(optarg); |
| break; |
| |
| case 'F': |
| format = pg_strdup(optarg); |
| break; |
| |
| case 'h': /* server host */ |
| dopt.cparams.pghost = pg_strdup(optarg); |
| break; |
| |
| case 'j': /* number of dump jobs */ |
| if (!option_parse_int(optarg, "-j/--jobs", 1, |
| PG_MAX_JOBS, |
| &numWorkers)) |
| exit_nicely(1); |
| break; |
| |
| case 'n': /* include schema(s) */ |
| simple_string_list_append(&schema_include_patterns, optarg); |
| dopt.include_everything = false; |
| break; |
| |
| case 'N': /* exclude schema(s) */ |
| simple_string_list_append(&schema_exclude_patterns, optarg); |
| break; |
| |
| case 'O': /* Don't reconnect to match owner */ |
| dopt.outputNoOwner = 1; |
| break; |
| |
| case 'p': /* server port */ |
| dopt.cparams.pgport = pg_strdup(optarg); |
| break; |
| |
| case 'R': |
| /* no-op, still accepted for backwards compatibility */ |
| break; |
| |
| case 's': /* dump schema only */ |
| dopt.schemaOnly = true; |
| break; |
| |
| case 'S': /* Username for superuser in plain text output */ |
| dopt.outputSuperuser = pg_strdup(optarg); |
| break; |
| |
| case 't': /* include table(s) */ |
| simple_string_list_append(&table_include_patterns, optarg); |
| dopt.include_everything = false; |
| break; |
| |
| case 'T': /* exclude table(s) */ |
| simple_string_list_append(&table_exclude_patterns, optarg); |
| break; |
| |
| case 'U': |
| dopt.cparams.username = pg_strdup(optarg); |
| break; |
| |
| case 'v': /* verbose */ |
| g_verbose = true; |
| pg_logging_increase_verbosity(); |
| break; |
| |
| case 'w': |
| dopt.cparams.promptPassword = TRI_NO; |
| break; |
| |
| case 'W': |
| dopt.cparams.promptPassword = TRI_YES; |
| break; |
| |
| case 'x': /* skip ACL dump */ |
| dopt.aclsSkip = true; |
| break; |
| |
| case 'Z': /* Compression */ |
| parse_compress_options(optarg, &compression_algorithm_str, |
| &compression_detail); |
| user_compression_defined = true; |
| break; |
| |
| case 0: |
| /* This covers the long options. */ |
| break; |
| |
| case 2: /* lock-wait-timeout */ |
| dopt.lockWaitTimeout = pg_strdup(optarg); |
| break; |
| |
| case 3: /* SET ROLE */ |
| use_role = pg_strdup(optarg); |
| break; |
| |
| case 4: /* exclude table(s) data */ |
| simple_string_list_append(&tabledata_exclude_patterns, optarg); |
| break; |
| |
| case 5: /* section */ |
| set_dump_section(optarg, &dopt.dumpSections); |
| break; |
| |
| case 6: /* snapshot */ |
| dumpsnapshot = pg_strdup(optarg); |
| break; |
| |
| case 7: /* no-sync */ |
| dosync = false; |
| break; |
| |
| <<<<<<< HEAD |
| case 8: |
| have_extra_float_digits = true; |
| extra_float_digits = atoi(optarg); |
| if (extra_float_digits < -15 || extra_float_digits > 3) |
| { |
| pg_log_error("extra_float_digits must be in range -15..3"); |
| exit_nicely(1); |
| } |
| break; |
| |
| case 9: /* inserts */ |
| |
| /* |
| * dump_inserts also stores --rows-per-insert, careful not to |
| * overwrite that. |
| */ |
| if (dopt.dump_inserts == 0) |
| dopt.dump_inserts = DUMP_DEFAULT_ROWS_PER_INSERT; |
| break; |
| |
| case 10: /* rows per insert */ |
| errno = 0; |
| rowsPerInsert = strtol(optarg, &endptr, 10); |
| |
| if (endptr == optarg || *endptr != '\0' || |
| rowsPerInsert <= 0 || rowsPerInsert > INT_MAX || |
| errno == ERANGE) |
| { |
| pg_log_error("rows-per-insert must be in range %d..%d", |
| 1, INT_MAX); |
| exit_nicely(1); |
| } |
| dopt.dump_inserts = (int) rowsPerInsert; |
| break; |
| |
| case 1000: /* gp-syntax */ |
| if (gp_syntax_option != GPS_NOT_SPECIFIED) |
| { |
| pg_log_warning("options \"--gp-syntax\" and \"--no-gp-syntax\" cannot be used together"); |
| exit(1); |
| } |
| gp_syntax_option = GPS_ENABLED; |
| break; |
| |
| case 1001: /* no-gp-syntax */ |
| if (gp_syntax_option != GPS_NOT_SPECIFIED) |
| { |
| pg_log_warning("options \"--gp-syntax\" and \"--no-gp-syntax\" cannot be used together"); |
| exit(1); |
| } |
| gp_syntax_option = GPS_DISABLED; |
| break; |
| |
| case 1002: |
| simple_string_list_append(&funcid_string_list, optarg); |
| dopt.include_everything = false; |
| break; |
| |
| case 1003: |
| simple_string_list_append(&relid_string_list, optarg); |
| dopt.include_everything = false; |
| ======= |
| case 8: |
| have_extra_float_digits = true; |
| if (!option_parse_int(optarg, "--extra-float-digits", -15, 3, |
| &extra_float_digits)) |
| exit_nicely(1); |
| break; |
| |
| case 9: /* inserts */ |
| |
| /* |
| * dump_inserts also stores --rows-per-insert, careful not to |
| * overwrite that. |
| */ |
| if (dopt.dump_inserts == 0) |
| dopt.dump_inserts = DUMP_DEFAULT_ROWS_PER_INSERT; |
| break; |
| |
| case 10: /* rows per insert */ |
| if (!option_parse_int(optarg, "--rows-per-insert", 1, INT_MAX, |
| &dopt.dump_inserts)) |
| exit_nicely(1); |
| >>>>>>> REL_16_9 |
| break; |
| |
| case 11: /* include foreign data */ |
| simple_string_list_append(&foreign_servers_include_patterns, |
| optarg); |
| break; |
| |
| case 12: /* include table(s) and their children */ |
| simple_string_list_append(&table_include_patterns_and_children, |
| optarg); |
| dopt.include_everything = false; |
| break; |
| |
| case 13: /* exclude table(s) and their children */ |
| simple_string_list_append(&table_exclude_patterns_and_children, |
| optarg); |
| break; |
| |
| case 14: /* exclude data of table(s) and children */ |
| simple_string_list_append(&tabledata_exclude_patterns_and_children, |
| optarg); |
| break; |
| |
| default: |
| /* getopt_long already emitted a complaint */ |
| pg_log_error_hint("Try \"%s --help\" for more information.", progname); |
| exit_nicely(1); |
| } |
| } |
| |
| /* |
| * Non-option argument specifies database name as long as it wasn't |
| * already specified with -d / --dbname |
| */ |
| if (optind < argc && dopt.cparams.dbname == NULL) |
| dopt.cparams.dbname = argv[optind++]; |
| |
| /* Complain if any arguments remain */ |
| if (optind < argc) |
| { |
| pg_log_error("too many command-line arguments (first is \"%s\")", |
| argv[optind]); |
| pg_log_error_hint("Try \"%s --help\" for more information.", progname); |
| exit_nicely(1); |
| } |
| |
| /* --column-inserts implies --inserts */ |
| if (dopt.column_inserts && dopt.dump_inserts == 0) |
| dopt.dump_inserts = DUMP_DEFAULT_ROWS_PER_INSERT; |
| |
| /* |
| * Binary upgrade mode implies dumping sequence data even in schema-only |
| * mode. This is not exposed as a separate option, but kept separate |
| * internally for clarity. |
| */ |
| if (dopt.binary_upgrade) |
| dopt.sequence_data = 1; |
| |
| if (dopt.dataOnly && dopt.schemaOnly) |
| pg_fatal("options -s/--schema-only and -a/--data-only cannot be used together"); |
| |
| if (dopt.schemaOnly && foreign_servers_include_patterns.head != NULL) |
| pg_fatal("options -s/--schema-only and --include-foreign-data cannot be used together"); |
| |
| if (numWorkers > 1 && foreign_servers_include_patterns.head != NULL) |
| pg_fatal("option --include-foreign-data is not supported with parallel backup"); |
| |
| if (dopt.dataOnly && dopt.outputClean) |
| pg_fatal("options -c/--clean and -a/--data-only cannot be used together"); |
| |
| if (dopt.if_exists && !dopt.outputClean) |
| pg_fatal("option --if-exists requires option -c/--clean"); |
| |
| /* |
| * --inserts are already implied above if --column-inserts or |
| * --rows-per-insert were specified. |
| */ |
| if (dopt.do_nothing && dopt.dump_inserts == 0) |
| pg_fatal("option --on-conflict-do-nothing requires option --inserts, --rows-per-insert, or --column-inserts"); |
| |
| /* Identify archive format to emit */ |
| archiveFormat = parseArchiveFormat(format, &archiveMode); |
| |
| /* archiveFormat specific setup */ |
| if (archiveFormat == archNull) |
| plainText = 1; |
| |
| /* |
| * Custom and directory formats are compressed by default with gzip when |
| * available, not the others. If gzip is not available, no compression is |
| * done by default. |
| */ |
| if ((archiveFormat == archCustom || archiveFormat == archDirectory) && |
| !user_compression_defined) |
| { |
| #ifdef HAVE_LIBZ |
| compression_algorithm_str = "gzip"; |
| #else |
| compression_algorithm_str = "none"; |
| #endif |
| } |
| |
| /* |
| * Compression options |
| */ |
| if (!parse_compress_algorithm(compression_algorithm_str, |
| &compression_algorithm)) |
| pg_fatal("unrecognized compression algorithm: \"%s\"", |
| compression_algorithm_str); |
| |
| parse_compress_specification(compression_algorithm, compression_detail, |
| &compression_spec); |
| error_detail = validate_compress_specification(&compression_spec); |
| if (error_detail != NULL) |
| pg_fatal("invalid compression specification: %s", |
| error_detail); |
| |
| error_detail = supports_compression(compression_spec); |
| if (error_detail != NULL) |
| pg_fatal("%s", error_detail); |
| |
| /* |
| * Disable support for zstd workers for now - these are based on |
| * threading, and it's unclear how it interacts with parallel dumps on |
| * platforms where that relies on threads too (e.g. Windows). |
| */ |
| if (compression_spec.options & PG_COMPRESSION_OPTION_WORKERS) |
| pg_log_warning("compression option \"%s\" is not currently supported by pg_dump", |
| "workers"); |
| |
| /* |
| * If emitting an archive format, we always want to emit a DATABASE item, |
| * in case --create is specified at pg_restore time. |
| */ |
| if (!plainText) |
| dopt.outputCreateDB = 1; |
| |
| /* Parallel backup only in the directory archive format so far */ |
| if (archiveFormat != archDirectory && numWorkers > 1) |
| pg_fatal("parallel backup only supported by the directory format"); |
| |
| /* Open the output file */ |
| fout = CreateArchive(filename, archiveFormat, compression_spec, |
| dosync, archiveMode, setupDumpWorker); |
| |
| /* Make dump options accessible right away */ |
| SetArchiveOptions(fout, &dopt, NULL); |
| |
| /* Register the cleanup hook */ |
| on_exit_close_archive(fout); |
| |
| /* Let the archiver know how noisy to be */ |
| fout->verbose = g_verbose; |
| |
| |
| /* |
| <<<<<<< HEAD |
| * We allow the server to be back to 8.3, and up to any minor release of |
| * our own major version. (See also version check in pg_dumpall.c.) |
| */ |
| fout->minRemoteVersion = 80300; /* we can handle back to 8.3 */ |
| ======= |
| * We allow the server to be back to 9.2, and up to any minor release of |
| * our own major version. (See also version check in pg_dumpall.c.) |
| */ |
| fout->minRemoteVersion = 90200; |
| >>>>>>> REL_16_9 |
| fout->maxRemoteVersion = (PG_VERSION_NUM / 100) * 100 + 99; |
| |
| fout->numWorkers = numWorkers; |
| |
| /* |
| * Open the database using the Archiver, so it knows about it. Errors mean |
| * death. |
| */ |
| ConnectDatabase(fout, &dopt.cparams, false, dopt.binary_upgrade); |
| setup_connection(fout, dumpencoding, dumpsnapshot, use_role); |
| |
| /* |
| <<<<<<< HEAD |
| * Determine whether or not we're interacting with a GP backend. |
| */ |
| isGPbackend = testGPbackend(fout); |
| |
| /* |
| * Now that the type of backend is known, determine the gp-syntax option |
| * value and set processing accordingly. |
| */ |
| switch (gp_syntax_option) |
| { |
| case GPS_NOT_SPECIFIED: |
| dumpGpPolicy = isGPbackend; |
| break; |
| case GPS_DISABLED: |
| dumpGpPolicy = false; |
| break; |
| case GPS_ENABLED: |
| dumpGpPolicy = isGPbackend; |
| if (!isGPbackend) |
| pg_log_warning("server is not a Apache Cloudberry instance; --gp-syntax option ignored"); |
| break; |
| } |
| |
| /* |
| * Disable security label support if server version < v9.1.x (prevents |
| * access to nonexistent pg_seclabel catalog) |
| */ |
| if (fout->remoteVersion < 90100) |
| dopt.no_security_labels = 1; |
| |
| /* |
| ======= |
| >>>>>>> REL_16_9 |
| * On hot standbys, never try to dump unlogged table data, since it will |
| * just throw an error. |
| */ |
| if (fout->isStandby) |
| dopt.no_unlogged_table_data = true; |
| |
| <<<<<<< HEAD |
| username_subquery = "SELECT rolname FROM pg_catalog.pg_roles WHERE oid ="; |
| |
| /* |
| * Remember whether or not this GP database supports partitioning. |
| */ |
| gp_partitioning_available = testPartitioningSupport(fout); |
| |
| /* check the version for the synchronized snapshots feature */ |
| if (numWorkers > 1 && fout->remoteVersion < 90200 |
| && !dopt.no_synchronized_snapshots) |
| fatal("Synchronized snapshots are not supported by this server version.\n" |
| "Run with --no-synchronized-snapshots instead if you do not need\n" |
| "synchronized snapshots."); |
| |
| /* check the version when a snapshot is explicitly specified by user */ |
| if (dumpsnapshot && fout->remoteVersion < 90200) |
| fatal("Exported snapshots are not supported by this server version."); |
| |
| ======= |
| >>>>>>> REL_16_9 |
| /* |
| * Find the last built-in OID, if needed (prior to 8.1) |
| * |
| * With 8.1 and above, we can just use FirstNormalObjectId - 1. |
| */ |
| g_last_builtin_oid = FirstNormalObjectId - 1; |
| |
| pg_log_info("last built-in OID is %u", g_last_builtin_oid); |
| |
| /* Expand schema selection patterns into OID lists */ |
| if (schema_include_patterns.head != NULL) |
| { |
| expand_schema_name_patterns(fout, &schema_include_patterns, |
| &schema_include_oids, |
| strict_names); |
| if (schema_include_oids.head == NULL) |
| pg_fatal("no matching schemas were found"); |
| } |
| expand_schema_name_patterns(fout, &schema_exclude_patterns, |
| &schema_exclude_oids, |
| false); |
| /* non-matching exclusion patterns aren't an error */ |
| |
| /* Expand table selection patterns into OID lists */ |
| expand_table_name_patterns(fout, &table_include_patterns, |
| &table_include_oids, |
| strict_names, false); |
| expand_table_name_patterns(fout, &table_include_patterns_and_children, |
| &table_include_oids, |
| strict_names, true); |
| if ((table_include_patterns.head != NULL || |
| table_include_patterns_and_children.head != NULL) && |
| table_include_oids.head == NULL) |
| pg_fatal("no matching tables were found"); |
| |
| expand_table_name_patterns(fout, &table_exclude_patterns, |
| &table_exclude_oids, |
| false, false); |
| expand_table_name_patterns(fout, &table_exclude_patterns_and_children, |
| &table_exclude_oids, |
| false, true); |
| |
| expand_table_name_patterns(fout, &tabledata_exclude_patterns, |
| &tabledata_exclude_oids, |
| false, false); |
| expand_table_name_patterns(fout, &tabledata_exclude_patterns_and_children, |
| &tabledata_exclude_oids, |
| false, true); |
| |
| expand_foreign_server_name_patterns(fout, &foreign_servers_include_patterns, |
| &foreign_servers_include_oids); |
| |
| /* non-matching exclusion patterns aren't an error */ |
| |
| |
| expand_oid_patterns(&relid_string_list, &table_include_oids); |
| expand_oid_patterns(&funcid_string_list, &function_include_oids); |
| |
| /* Expand extension selection patterns into OID lists */ |
| if (extension_include_patterns.head != NULL) |
| { |
| expand_extension_name_patterns(fout, &extension_include_patterns, |
| &extension_include_oids, |
| strict_names); |
| if (extension_include_oids.head == NULL) |
| pg_fatal("no matching extensions were found"); |
| } |
| |
| /* |
| * Dumping LOs is the default for dumps where an inclusion switch is not |
| * used (an "include everything" dump). -B can be used to exclude LOs |
| * from those dumps. -b can be used to include LOs even when an inclusion |
| * switch is used. |
| * |
| * -s means "schema only" and LOs are data, not schema, so we never |
| * include LOs when -s is used. |
| */ |
| if (dopt.include_everything && !dopt.schemaOnly && !dopt.dontOutputLOs) |
| dopt.outputLOs = true; |
| |
| /* |
| * Collect role names so we can map object owner OIDs to names. |
| */ |
| collectRoleNames(fout); |
| |
| /* |
| * Now scan the database and create DumpableObject structs for all the |
| * objects we intend to dump. |
| */ |
| tblinfo = getSchemaData(fout, &numTables); |
| |
| if (!dopt.schemaOnly) |
| { |
| getTableData(&dopt, tblinfo, numTables, 0); |
| buildMatViewRefreshDependencies(fout); |
| if (dopt.dataOnly) |
| getTableDataFKConstraints(); |
| } |
| |
| if (dopt.schemaOnly && dopt.sequence_data) |
| getTableData(&dopt, tblinfo, numTables, RELKIND_SEQUENCE); |
| |
| /* |
| * In binary-upgrade mode, we do not have to worry about the actual LO |
| * data or the associated metadata that resides in the pg_largeobject and |
| * pg_largeobject_metadata tables, respectively. |
| * |
| * However, we do need to collect LO information as there may be comments |
| * or other information on LOs that we do need to dump out. |
| */ |
| if (dopt.outputLOs || dopt.binary_upgrade) |
| getLOs(fout); |
| |
| /* |
| * Collect dependency data to assist in ordering the objects. |
| */ |
| getDependencies(fout); |
| |
| <<<<<<< HEAD |
| setExtPartDependency(tblinfo, numTables); |
| ======= |
| /* |
| * Collect ACLs, comments, and security labels, if wanted. |
| */ |
| if (!dopt.aclsSkip) |
| getAdditionalACLs(fout); |
| if (!dopt.no_comments) |
| collectComments(fout); |
| if (!dopt.no_security_labels) |
| collectSecLabels(fout); |
| >>>>>>> REL_16_9 |
| |
| /* Lastly, create dummy objects to represent the section boundaries */ |
| boundaryObjs = createBoundaryObjects(); |
| |
| /* Get pointers to all the known DumpableObjects */ |
| getDumpableObjects(&dobjs, &numObjs); |
| |
| /* |
| * Add dummy dependencies to enforce the dump section ordering. |
| */ |
| addBoundaryDependencies(dobjs, numObjs, boundaryObjs); |
| |
| /* |
| * Sort the objects into a safe dump order (no forward references). |
| * |
| * We rely on dependency information to help us determine a safe order, so |
| * the initial sort is mostly for cosmetic purposes: we sort by name to |
| * ensure that logically identical schemas will dump identically. |
| */ |
| sortDumpableObjectsByTypeName(dobjs, numObjs); |
| |
| sortDumpableObjects(dobjs, numObjs, |
| boundaryObjs[0].dumpId, boundaryObjs[1].dumpId); |
| |
| /* |
| * Create archive TOC entries for all the objects to be dumped, in a safe |
| * order. |
| */ |
| |
| /* |
| * First the special entries for ENCODING, STDSTRINGS, and SEARCHPATH. |
| */ |
| dumpEncoding(fout); |
| dumpStdStrings(fout); |
| dumpSearchPath(fout); |
| |
| /* The database items are always next, unless we don't want them at all */ |
| if (dopt.outputCreateDB) |
| dumpDatabase(fout); |
| |
| int binfo_index = -1; |
| /* Now the rearrangeable objects. */ |
| for (i = 0; i < numObjs; i++) |
| { |
| dumpDumpableObject(fout, dobjs[i]); |
| if (dobjs[i]->objType == DO_BINARY_UPGRADE) |
| binfo_index = i; |
| } |
| |
| /* Amend the Oid preassignment TOC with the actual Oids gathered */ |
| if (dopt.binary_upgrade && binfo_index >= 0) |
| dumpPreassignedOidDefinition(fout, (BinaryUpgradeInfo *) dobjs[binfo_index]); |
| |
| /* |
| * Set up options info to ensure we dump what we want. |
| */ |
| ropt = NewRestoreOptions(); |
| ropt->filename = filename; |
| |
| /* if you change this list, see dumpOptionsFromRestoreOptions */ |
| ropt->cparams.dbname = dopt.cparams.dbname ? pg_strdup(dopt.cparams.dbname) : NULL; |
| ropt->cparams.pgport = dopt.cparams.pgport ? pg_strdup(dopt.cparams.pgport) : NULL; |
| ropt->cparams.pghost = dopt.cparams.pghost ? pg_strdup(dopt.cparams.pghost) : NULL; |
| ropt->cparams.username = dopt.cparams.username ? pg_strdup(dopt.cparams.username) : NULL; |
| ropt->cparams.promptPassword = dopt.cparams.promptPassword; |
| ropt->dropSchema = dopt.outputClean; |
| ropt->dataOnly = dopt.dataOnly; |
| ropt->schemaOnly = dopt.schemaOnly; |
| ropt->if_exists = dopt.if_exists; |
| ropt->column_inserts = dopt.column_inserts; |
| ropt->dumpSections = dopt.dumpSections; |
| ropt->aclsSkip = dopt.aclsSkip; |
| ropt->superuser = dopt.outputSuperuser; |
| ropt->createDB = dopt.outputCreateDB; |
| ropt->noOwner = dopt.outputNoOwner; |
| ropt->noTableAm = dopt.outputNoTableAm; |
| ropt->noTablespace = dopt.outputNoTablespaces; |
| ropt->disable_triggers = dopt.disable_triggers; |
| ropt->use_setsessauth = dopt.use_setsessauth; |
| ropt->disable_dollar_quoting = dopt.disable_dollar_quoting; |
| ropt->dump_inserts = dopt.dump_inserts; |
| ropt->no_comments = dopt.no_comments; |
| ropt->no_publications = dopt.no_publications; |
| ropt->no_security_labels = dopt.no_security_labels; |
| ropt->no_subscriptions = dopt.no_subscriptions; |
| ropt->lockWaitTimeout = dopt.lockWaitTimeout; |
| ropt->include_everything = dopt.include_everything; |
| ropt->enable_row_security = dopt.enable_row_security; |
| ropt->sequence_data = dopt.sequence_data; |
| ropt->binary_upgrade = dopt.binary_upgrade; |
| |
| ropt->compression_spec = compression_spec; |
| |
| ropt->suppressDumpWarnings = true; /* We've already shown them */ |
| |
| SetArchiveOptions(fout, &dopt, ropt); |
| |
| /* Mark which entries should be output */ |
| ProcessArchiveRestoreOptions(fout); |
| |
| /* |
| * The archive's TOC entries are now marked as to which ones will actually |
| * be output, so we can set up their dependency lists properly. This isn't |
| * necessary for plain-text output, though. |
| */ |
| if (!plainText) |
| BuildArchiveDependencies(fout); |
| |
| /* |
| * And finally we can do the actual output. |
| * |
| * Note: for non-plain-text output formats, the output file is written |
| * inside CloseArchive(). This is, um, bizarre; but not worth changing |
| * right now. |
| */ |
| if (plainText) |
| RestoreArchive(fout); |
| |
| CloseArchive(fout); |
| |
| exit_nicely(0); |
| } |
| |
| |
| static void |
| help(const char *progname) |
| { |
| printf(_("%s dumps a database as a text file or to other formats.\n\n"), progname); |
| printf(_("Usage:\n")); |
| printf(_(" %s [OPTION]... [DBNAME]\n"), progname); |
| |
| printf(_("\nGeneral options:\n")); |
| printf(_(" -f, --file=FILENAME output file or directory name\n")); |
| printf(_(" -F, --format=c|d|t|p output file format (custom, directory, tar,\n" |
| " plain text (default))\n")); |
| printf(_(" -j, --jobs=NUM use this many parallel jobs to dump\n")); |
| printf(_(" -v, --verbose verbose mode\n")); |
| printf(_(" -V, --version output version information, then exit\n")); |
| printf(_(" -Z, --compress=METHOD[:DETAIL]\n" |
| " compress as specified\n")); |
| printf(_(" --lock-wait-timeout=TIMEOUT fail after waiting TIMEOUT for a table lock\n")); |
| printf(_(" --no-sync do not wait for changes to be written safely to disk\n")); |
| printf(_(" -?, --help show this help, then exit\n")); |
| |
| printf(_("\nOptions controlling the output content:\n")); |
| printf(_(" -a, --data-only dump only the data, not the schema\n")); |
| printf(_(" -b, --large-objects include large objects in dump\n")); |
| printf(_(" --blobs (same as --large-objects, deprecated)\n")); |
| printf(_(" -B, --no-large-objects exclude large objects in dump\n")); |
| printf(_(" --no-blobs (same as --no-large-objects, deprecated)\n")); |
| printf(_(" -c, --clean clean (drop) database objects before recreating\n")); |
| printf(_(" -C, --create include commands to create database in dump\n")); |
| printf(_(" -e, --extension=PATTERN dump the specified extension(s) only\n")); |
| printf(_(" -E, --encoding=ENCODING dump the data in encoding ENCODING\n")); |
| printf(_(" -n, --schema=PATTERN dump the specified schema(s) only\n")); |
| printf(_(" -N, --exclude-schema=PATTERN do NOT dump the specified schema(s)\n")); |
| printf(_(" -O, --no-owner skip restoration of object ownership in\n" |
| " plain-text format\n")); |
| printf(_(" -s, --schema-only dump only the schema, no data\n")); |
| printf(_(" -S, --superuser=NAME superuser user name to use in plain-text format\n")); |
| printf(_(" -t, --table=PATTERN dump only the specified table(s)\n")); |
| printf(_(" -T, --exclude-table=PATTERN do NOT dump the specified table(s)\n")); |
| printf(_(" -x, --no-privileges do not dump privileges (grant/revoke)\n")); |
| printf(_(" --binary-upgrade for use by upgrade utilities only\n")); |
| printf(_(" --column-inserts dump data as INSERT commands with column names\n")); |
| printf(_(" --disable-dollar-quoting disable dollar quoting, use SQL standard quoting\n")); |
| printf(_(" --disable-triggers disable triggers during data-only restore\n")); |
| printf(_(" --enable-row-security enable row security (dump only content user has\n" |
| " access to)\n")); |
| printf(_(" --exclude-table-and-children=PATTERN\n" |
| " do NOT dump the specified table(s), including\n" |
| " child and partition tables\n")); |
| printf(_(" --exclude-table-data=PATTERN do NOT dump data for the specified table(s)\n")); |
| printf(_(" --exclude-table-data-and-children=PATTERN\n" |
| " do NOT dump data for the specified table(s),\n" |
| " including child and partition tables\n")); |
| printf(_(" --extra-float-digits=NUM override default setting for extra_float_digits\n")); |
| printf(_(" --if-exists use IF EXISTS when dropping objects\n")); |
| printf(_(" --include-foreign-data=PATTERN\n" |
| " include data of foreign tables on foreign\n" |
| " servers matching PATTERN\n")); |
| printf(_(" --inserts dump data as INSERT commands, rather than COPY\n")); |
| printf(_(" --load-via-partition-root load partitions via the root table\n")); |
| printf(_(" --no-comments do not dump comments\n")); |
| printf(_(" --no-publications do not dump publications\n")); |
| printf(_(" --no-security-labels do not dump security label assignments\n")); |
| printf(_(" --no-subscriptions do not dump subscriptions\n")); |
| printf(_(" --no-table-access-method do not dump table access methods\n")); |
| printf(_(" --no-tablespaces do not dump tablespace assignments\n")); |
| printf(_(" --no-toast-compression do not dump TOAST compression methods\n")); |
| printf(_(" --no-unlogged-table-data do not dump unlogged table data\n")); |
| printf(_(" --on-conflict-do-nothing add ON CONFLICT DO NOTHING to INSERT commands\n")); |
| printf(_(" --quote-all-identifiers quote all identifiers, even if not key words\n")); |
| printf(_(" --rows-per-insert=NROWS number of rows per INSERT; implies --inserts\n")); |
| printf(_(" --section=SECTION dump named section (pre-data, data, or post-data)\n")); |
| printf(_(" --serializable-deferrable wait until the dump can run without anomalies\n")); |
| printf(_(" --snapshot=SNAPSHOT use given snapshot for the dump\n")); |
| printf(_(" --strict-names require table and/or schema include patterns to\n" |
| " match at least one entity each\n")); |
| printf(_(" --table-and-children=PATTERN dump only the specified table(s), including\n" |
| " child and partition tables\n")); |
| printf(_(" --use-set-session-authorization\n" |
| " use SET SESSION AUTHORIZATION commands instead of\n" |
| " ALTER OWNER commands to set ownership\n")); |
| |
| /* START MPP ADDITION */ |
| printf(_(" --gp-syntax dump with Apache Cloudberry syntax (default if gpdb)\n")); |
| printf(_(" --no-gp-syntax dump without Apache Cloudberry syntax (default if postgresql)\n")); |
| printf(_(" --function-oids dump only function(s) of given list of oids\n")); |
| printf(_(" --relation-oids dump only relation(s) of given list of oids\n")); |
| /* END MPP ADDITION */ |
| |
| printf(_("\nConnection options:\n")); |
| printf(_(" -d, --dbname=DBNAME database to dump\n")); |
| printf(_(" -h, --host=HOSTNAME database server host or socket directory\n")); |
| printf(_(" -p, --port=PORT database server port number\n")); |
| printf(_(" -U, --username=NAME connect as specified database user\n")); |
| printf(_(" -w, --no-password never prompt for password\n")); |
| printf(_(" -W, --password force password prompt (should happen automatically)\n")); |
| printf(_(" --role=ROLENAME do SET ROLE before dump\n")); |
| |
| printf(_("\nIf no database name is supplied, then the PGDATABASE environment\n" |
| "variable value is used.\n\n")); |
| printf(_("Report bugs to <%s>.\n"), PACKAGE_BUGREPORT); |
| printf(_("%s home page: <%s>\n"), PACKAGE_NAME, PACKAGE_URL); |
| } |
| |
| static void |
| setup_connection(Archive *AH, const char *dumpencoding, |
| const char *dumpsnapshot, char *use_role) |
| { |
| DumpOptions *dopt = AH->dopt; |
| PGconn *conn = GetConnection(AH); |
| const char *std_strings; |
| |
| PQclear(ExecuteSqlQueryForSingleRow(AH, ALWAYS_SECURE_SEARCH_PATH_SQL)); |
| |
| /* |
| * Set the client encoding if requested. |
| */ |
| if (dumpencoding) |
| { |
| if (PQsetClientEncoding(conn, dumpencoding) < 0) |
| pg_fatal("invalid client encoding \"%s\" specified", |
| dumpencoding); |
| } |
| |
| /* |
| * Get the active encoding and the standard_conforming_strings setting, so |
| * we know how to escape strings. |
| */ |
| AH->encoding = PQclientEncoding(conn); |
| setFmtEncoding(AH->encoding); |
| |
| std_strings = PQparameterStatus(conn, "standard_conforming_strings"); |
| AH->std_strings = (std_strings && strcmp(std_strings, "on") == 0); |
| |
| /* |
| * Set the role if requested. In a parallel dump worker, we'll be passed |
| * use_role == NULL, but AH->use_role is already set (if user specified it |
| * originally) and we should use that. |
| */ |
| if (!use_role && AH->use_role) |
| use_role = AH->use_role; |
| |
| /* Set the role if requested */ |
| if (use_role) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "SET ROLE %s", fmtId(use_role)); |
| ExecuteSqlStatement(AH, query->data); |
| destroyPQExpBuffer(query); |
| |
| /* save it for possible later use by parallel workers */ |
| if (!AH->use_role) |
| AH->use_role = pg_strdup(use_role); |
| } |
| |
| /* Set the datestyle to ISO to ensure the dump's portability */ |
| ExecuteSqlStatement(AH, "SET DATESTYLE = ISO"); |
| |
| /* Likewise, avoid using sql_standard intervalstyle */ |
| ExecuteSqlStatement(AH, "SET INTERVALSTYLE = POSTGRES"); |
| |
| /* |
| * Use an explicitly specified extra_float_digits if it has been provided. |
| * Otherwise, set extra_float_digits so that we can dump float data |
| * exactly (given correctly implemented float I/O code, anyway). |
| */ |
| if (have_extra_float_digits) |
| { |
| PQExpBuffer q = createPQExpBuffer(); |
| |
| appendPQExpBuffer(q, "SET extra_float_digits TO %d", |
| extra_float_digits); |
| ExecuteSqlStatement(AH, q->data); |
| destroyPQExpBuffer(q); |
| } |
| else |
| ExecuteSqlStatement(AH, "SET extra_float_digits TO 3"); |
| |
| /* |
| * Disable synchronized scanning, to prevent unpredictable changes in row |
| * ordering across a dump and reload. |
| */ |
| ExecuteSqlStatement(AH, "SET synchronize_seqscans TO off"); |
| |
| /* |
| * The default for enable_nestloop is off in GPDB. However, many of the queries |
| * that we issue best run with nested loop joins, so enable it. |
| */ |
| ExecuteSqlStatement(AH, "SET enable_nestloop TO on"); |
| |
| /* |
| * Disable timeouts if supported. |
| */ |
| ExecuteSqlStatement(AH, "SET statement_timeout = 0"); |
| if (AH->remoteVersion >= 90300) |
| ExecuteSqlStatement(AH, "SET lock_timeout = 0"); |
| if (AH->remoteVersion >= 90600) |
| ExecuteSqlStatement(AH, "SET idle_in_transaction_session_timeout = 0"); |
| |
| /* |
| * Quote all identifiers, if requested. |
| */ |
| <<<<<<< HEAD |
| if (quote_all_identifiers && AH->remoteVersion >= 80300) |
| ======= |
| if (quote_all_identifiers) |
| >>>>>>> REL_16_9 |
| ExecuteSqlStatement(AH, "SET quote_all_identifiers = true"); |
| |
| /* |
| * Adjust row-security mode, if supported. |
| */ |
| if (AH->remoteVersion >= 90500) |
| { |
| if (dopt->enable_row_security) |
| ExecuteSqlStatement(AH, "SET row_security = on"); |
| else |
| ExecuteSqlStatement(AH, "SET row_security = off"); |
| } |
| |
| /* |
| * For security reasons, we restrict the expansion of non-system views and |
| * access to foreign tables during the pg_dump process. This restriction |
| * is adjusted when dumping foreign table data. |
| */ |
| set_restrict_relation_kind(AH, "view, foreign-table"); |
| |
| /* |
| * Initialize prepared-query state to "nothing prepared". We do this here |
| * so that a parallel dump worker will have its own state. |
| */ |
| AH->is_prepared = (bool *) pg_malloc0(NUM_PREP_QUERIES * sizeof(bool)); |
| |
| /* |
| * Start transaction-snapshot mode transaction to dump consistent data. |
| */ |
| ExecuteSqlStatement(AH, "BEGIN"); |
| |
| /* |
| * To support the combination of serializable_deferrable with the jobs |
| * option we use REPEATABLE READ for the worker connections that are |
| * passed a snapshot. As long as the snapshot is acquired in a |
| * SERIALIZABLE, READ ONLY, DEFERRABLE transaction, its use within a |
| * REPEATABLE READ transaction provides the appropriate integrity |
| * guarantees. This is a kluge, but safe for back-patching. |
| */ |
| if (dopt->serializable_deferrable && AH->sync_snapshot_id == NULL) |
| ExecuteSqlStatement(AH, |
| "SET TRANSACTION ISOLATION LEVEL " |
| "SERIALIZABLE, READ ONLY, DEFERRABLE"); |
| else |
| ExecuteSqlStatement(AH, |
| "SET TRANSACTION ISOLATION LEVEL " |
| "REPEATABLE READ, READ ONLY"); |
| |
| /* |
| * If user specified a snapshot to use, select that. In a parallel dump |
| * worker, we'll be passed dumpsnapshot == NULL, but AH->sync_snapshot_id |
| * is already set (if the server can handle it) and we should use that. |
| */ |
| if (dumpsnapshot) |
| AH->sync_snapshot_id = pg_strdup(dumpsnapshot); |
| |
| if (AH->sync_snapshot_id) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SET TRANSACTION SNAPSHOT "); |
| appendStringLiteralConn(query, AH->sync_snapshot_id, conn); |
| ExecuteSqlStatement(AH, query->data); |
| destroyPQExpBuffer(query); |
| } |
| else if (AH->numWorkers > 1) |
| { |
| if (AH->isStandby && AH->remoteVersion < 100000) |
| pg_fatal("parallel dumps from standby servers are not supported by this server version"); |
| AH->sync_snapshot_id = get_synchronized_snapshot(AH); |
| } |
| } |
| |
| /* Set up connection for a parallel worker process */ |
| static void |
| setupDumpWorker(Archive *AH) |
| { |
| /* |
| * We want to re-select all the same values the leader connection is |
| * using. We'll have inherited directly-usable values in |
| * AH->sync_snapshot_id and AH->use_role, but we need to translate the |
| * inherited encoding value back to a string to pass to setup_connection. |
| */ |
| setup_connection(AH, |
| pg_encoding_to_char(AH->encoding), |
| NULL, |
| NULL); |
| } |
| |
| static char * |
| get_synchronized_snapshot(Archive *fout) |
| { |
| char *query = "SELECT pg_catalog.pg_export_snapshot()"; |
| char *result; |
| PGresult *res; |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query); |
| result = pg_strdup(PQgetvalue(res, 0, 0)); |
| PQclear(res); |
| |
| return result; |
| } |
| |
| static ArchiveFormat |
| parseArchiveFormat(const char *format, ArchiveMode *mode) |
| { |
| ArchiveFormat archiveFormat; |
| |
| *mode = archModeWrite; |
| |
| if (pg_strcasecmp(format, "a") == 0 || pg_strcasecmp(format, "append") == 0) |
| { |
| /* This is used by pg_dumpall, and is not documented */ |
| archiveFormat = archNull; |
| *mode = archModeAppend; |
| } |
| else if (pg_strcasecmp(format, "c") == 0) |
| archiveFormat = archCustom; |
| else if (pg_strcasecmp(format, "custom") == 0) |
| archiveFormat = archCustom; |
| else if (pg_strcasecmp(format, "d") == 0) |
| archiveFormat = archDirectory; |
| else if (pg_strcasecmp(format, "directory") == 0) |
| archiveFormat = archDirectory; |
| else if (pg_strcasecmp(format, "p") == 0) |
| archiveFormat = archNull; |
| else if (pg_strcasecmp(format, "plain") == 0) |
| archiveFormat = archNull; |
| else if (pg_strcasecmp(format, "t") == 0) |
| archiveFormat = archTar; |
| else if (pg_strcasecmp(format, "tar") == 0) |
| archiveFormat = archTar; |
| else |
| pg_fatal("invalid output format \"%s\" specified", format); |
| return archiveFormat; |
| } |
| |
| /* |
| * Find the OIDs of all schemas matching the given list of patterns, |
| * and append them to the given OID list. |
| */ |
| static void |
| expand_schema_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids, |
| bool strict_names) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| SimpleStringListCell *cell; |
| int i; |
| |
| if (patterns->head == NULL) |
| return; /* nothing to do */ |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * The loop below runs multiple SELECTs might sometimes result in |
| * duplicate entries in the OID list, but we don't care. |
| */ |
| |
| for (cell = patterns->head; cell; cell = cell->next) |
| { |
| <<<<<<< HEAD |
| PQExpBufferData dbbuf; |
| int dotcnt; |
| ======= |
| PQExpBufferData dbbuf; |
| int dotcnt; |
| >>>>>>> REL_16_9 |
| |
| appendPQExpBufferStr(query, |
| "SELECT oid FROM pg_catalog.pg_namespace n\n"); |
| initPQExpBuffer(&dbbuf); |
| processSQLNamePattern(GetConnection(fout), query, cell->val, false, |
| false, NULL, "n.nspname", NULL, NULL, &dbbuf, |
| &dotcnt); |
| if (dotcnt > 1) |
| <<<<<<< HEAD |
| fatal("improper qualified name (too many dotted names): %s", |
| cell->val); |
| ======= |
| pg_fatal("improper qualified name (too many dotted names): %s", |
| cell->val); |
| >>>>>>> REL_16_9 |
| else if (dotcnt == 1) |
| prohibit_crossdb_refs(GetConnection(fout), dbbuf.data, cell->val); |
| termPQExpBuffer(&dbbuf); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| if (strict_names && PQntuples(res) == 0) |
| pg_fatal("no matching schemas were found for pattern \"%s\"", cell->val); |
| |
| for (i = 0; i < PQntuples(res); i++) |
| { |
| simple_oid_list_append(oids, atooid(PQgetvalue(res, i, 0))); |
| } |
| |
| PQclear(res); |
| resetPQExpBuffer(query); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * Find the OIDs of all extensions matching the given list of patterns, |
| * and append them to the given OID list. |
| */ |
| static void |
| expand_extension_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids, |
| bool strict_names) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| SimpleStringListCell *cell; |
| int i; |
| |
| if (patterns->head == NULL) |
| return; /* nothing to do */ |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * The loop below runs multiple SELECTs might sometimes result in |
| * duplicate entries in the OID list, but we don't care. |
| */ |
| for (cell = patterns->head; cell; cell = cell->next) |
| { |
| <<<<<<< HEAD |
| int dotcnt; |
| ======= |
| int dotcnt; |
| >>>>>>> REL_16_9 |
| |
| appendPQExpBufferStr(query, |
| "SELECT oid FROM pg_catalog.pg_extension e\n"); |
| processSQLNamePattern(GetConnection(fout), query, cell->val, false, |
| false, NULL, "e.extname", NULL, NULL, NULL, |
| &dotcnt); |
| if (dotcnt > 0) |
| <<<<<<< HEAD |
| fatal("improper qualified name (too many dotted names): %s", |
| cell->val); |
| ======= |
| pg_fatal("improper qualified name (too many dotted names): %s", |
| cell->val); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| if (strict_names && PQntuples(res) == 0) |
| pg_fatal("no matching extensions were found for pattern \"%s\"", cell->val); |
| |
| for (i = 0; i < PQntuples(res); i++) |
| { |
| simple_oid_list_append(oids, atooid(PQgetvalue(res, i, 0))); |
| } |
| |
| PQclear(res); |
| resetPQExpBuffer(query); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * Find the OIDs of all foreign servers matching the given list of patterns, |
| * and append them to the given OID list. |
| */ |
| static void |
| expand_foreign_server_name_patterns(Archive *fout, |
| SimpleStringList *patterns, |
| SimpleOidList *oids) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| SimpleStringListCell *cell; |
| int i; |
| |
| if (patterns->head == NULL) |
| return; /* nothing to do */ |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * The loop below runs multiple SELECTs might sometimes result in |
| * duplicate entries in the OID list, but we don't care. |
| */ |
| |
| for (cell = patterns->head; cell; cell = cell->next) |
| { |
| <<<<<<< HEAD |
| int dotcnt; |
| ======= |
| int dotcnt; |
| >>>>>>> REL_16_9 |
| |
| appendPQExpBufferStr(query, |
| "SELECT oid FROM pg_catalog.pg_foreign_server s\n"); |
| processSQLNamePattern(GetConnection(fout), query, cell->val, false, |
| false, NULL, "s.srvname", NULL, NULL, NULL, |
| &dotcnt); |
| if (dotcnt > 0) |
| <<<<<<< HEAD |
| fatal("improper qualified name (too many dotted names): %s", |
| cell->val); |
| ======= |
| pg_fatal("improper qualified name (too many dotted names): %s", |
| cell->val); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| if (PQntuples(res) == 0) |
| pg_fatal("no matching foreign servers were found for pattern \"%s\"", cell->val); |
| |
| for (i = 0; i < PQntuples(res); i++) |
| simple_oid_list_append(oids, atooid(PQgetvalue(res, i, 0))); |
| |
| PQclear(res); |
| resetPQExpBuffer(query); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * Find the OIDs of all tables matching the given list of patterns, |
| * and append them to the given OID list. See also expand_dbname_patterns() |
| * in pg_dumpall.c |
| */ |
| static void |
| expand_table_name_patterns(Archive *fout, |
| SimpleStringList *patterns, SimpleOidList *oids, |
| bool strict_names, bool with_child_tables) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| SimpleStringListCell *cell; |
| int i; |
| |
| if (patterns->head == NULL) |
| return; /* nothing to do */ |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * this might sometimes result in duplicate entries in the OID list, but |
| * we don't care. |
| */ |
| |
| for (cell = patterns->head; cell; cell = cell->next) |
| { |
| <<<<<<< HEAD |
| PQExpBufferData dbbuf; |
| int dotcnt; |
| ======= |
| PQExpBufferData dbbuf; |
| int dotcnt; |
| >>>>>>> REL_16_9 |
| |
| /* |
| * Query must remain ABSOLUTELY devoid of unqualified names. This |
| * would be unnecessary given a pg_table_is_visible() variant taking a |
| * search_path argument. |
| * |
| * For with_child_tables, we start with the basic query's results and |
| * recursively search the inheritance tree to add child tables. |
| */ |
| if (with_child_tables) |
| { |
| appendPQExpBuffer(query, "WITH RECURSIVE partition_tree (relid) AS (\n"); |
| } |
| |
| appendPQExpBuffer(query, |
| "SELECT c.oid" |
| "\nFROM pg_catalog.pg_class c" |
| "\n LEFT JOIN pg_catalog.pg_namespace n" |
| "\n ON n.oid OPERATOR(pg_catalog.=) c.relnamespace" |
| "\nWHERE c.relkind OPERATOR(pg_catalog.=) ANY" |
| "\n (array['%c', '%c', '%c', '%c', '%c', '%c'])\n", |
| RELKIND_RELATION, RELKIND_SEQUENCE, RELKIND_VIEW, |
| RELKIND_MATVIEW, RELKIND_FOREIGN_TABLE, |
| RELKIND_PARTITIONED_TABLE); |
| initPQExpBuffer(&dbbuf); |
| processSQLNamePattern(GetConnection(fout), query, cell->val, true, |
| false, "n.nspname", "c.relname", NULL, |
| "pg_catalog.pg_table_is_visible(c.oid)", &dbbuf, |
| &dotcnt); |
| if (dotcnt > 2) |
| <<<<<<< HEAD |
| fatal("improper relation name (too many dotted names): %s", |
| cell->val); |
| else if (dotcnt == 2) |
| prohibit_crossdb_refs(GetConnection(fout), dbbuf.data, cell->val); |
| termPQExpBuffer(&dbbuf); |
| ======= |
| pg_fatal("improper relation name (too many dotted names): %s", |
| cell->val); |
| else if (dotcnt == 2) |
| prohibit_crossdb_refs(GetConnection(fout), dbbuf.data, cell->val); |
| termPQExpBuffer(&dbbuf); |
| |
| if (with_child_tables) |
| { |
| appendPQExpBuffer(query, "UNION" |
| "\nSELECT i.inhrelid" |
| "\nFROM partition_tree p" |
| "\n JOIN pg_catalog.pg_inherits i" |
| "\n ON p.relid OPERATOR(pg_catalog.=) i.inhparent" |
| "\n)" |
| "\nSELECT relid FROM partition_tree"); |
| } |
| >>>>>>> REL_16_9 |
| |
| ExecuteSqlStatement(fout, "RESET search_path"); |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| PQclear(ExecuteSqlQueryForSingleRow(fout, |
| ALWAYS_SECURE_SEARCH_PATH_SQL)); |
| if (strict_names && PQntuples(res) == 0) |
| pg_fatal("no matching tables were found for pattern \"%s\"", cell->val); |
| |
| for (i = 0; i < PQntuples(res); i++) |
| { |
| simple_oid_list_append(oids, atooid(PQgetvalue(res, i, 0))); |
| } |
| |
| PQclear(res); |
| resetPQExpBuffer(query); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| <<<<<<< HEAD |
| * Parse the OIDs matching the given list of patterns separated by non-digit |
| * characters, and append them to the given OID list. |
| */ |
| static void |
| expand_oid_patterns(SimpleStringList *patterns, SimpleOidList *oids) |
| { |
| SimpleStringListCell *cell; |
| |
| if (patterns->head == NULL) |
| return; /* nothing to do */ |
| |
| for (cell = patterns->head; cell; cell = cell->next) |
| { |
| const char *seperator = ","; |
| char *oidstr = pg_strdup(cell->val); |
| char *token = strtok(oidstr, seperator); |
| while (token) |
| { |
| if (strstr(seperator, token) == NULL) |
| simple_oid_list_append(oids, atooid(token)); |
| |
| token = strtok(NULL, seperator); |
| } |
| |
| free(oidstr); |
| } |
| } |
| |
| /* |
| ======= |
| >>>>>>> REL_16_9 |
| * Verifies that the connected database name matches the given database name, |
| * and if not, dies with an error about the given pattern. |
| * |
| * The 'dbname' argument should be a literal name parsed from 'pattern'. |
| */ |
| static void |
| prohibit_crossdb_refs(PGconn *conn, const char *dbname, const char *pattern) |
| { |
| const char *db; |
| |
| db = PQdb(conn); |
| if (db == NULL) |
| <<<<<<< HEAD |
| fatal("You are currently not connected to a database."); |
| |
| if (strcmp(db, dbname) != 0) |
| fatal("cross-database references are not implemented: %s", pattern); |
| ======= |
| pg_fatal("You are currently not connected to a database."); |
| |
| if (strcmp(db, dbname) != 0) |
| pg_fatal("cross-database references are not implemented: %s", |
| pattern); |
| >>>>>>> REL_16_9 |
| } |
| |
| /* |
| * checkExtensionMembership |
| * Determine whether object is an extension member, and if so, |
| * record an appropriate dependency and set the object's dump flag. |
| * |
| * It's important to call this for each object that could be an extension |
| * member. Generally, we integrate this with determining the object's |
| * to-be-dumped-ness, since extension membership overrides other rules for that. |
| * |
| * Returns true if object is an extension member, else false. |
| */ |
| static bool |
| checkExtensionMembership(DumpableObject *dobj, Archive *fout) |
| { |
| ExtensionInfo *ext = findOwningExtension(dobj->catId); |
| |
| if (ext == NULL) |
| return false; |
| |
| dobj->ext_member = true; |
| |
| /* Record dependency so that getDependencies needn't deal with that */ |
| addObjectDependency(dobj, ext->dobj.dumpId); |
| |
| /* |
| * In 9.6 and above, mark the member object to have any non-initial ACLs |
| * dumped. (Any initial ACLs will be removed later, using data from |
| * pg_init_privs, so that we'll dump only the delta from the extension's |
| * initial setup.) |
| * |
| * Prior to 9.6, we do not include any extension member components. |
| * |
| * In binary upgrades, we still dump all components of the members |
| * individually, since the idea is to exactly reproduce the database |
| * contents rather than replace the extension contents with something |
| * different. |
| * |
| * Note: it might be interesting someday to implement storage and delta |
| * dumping of extension members' RLS policies and/or security labels. |
| * However there is a pitfall for RLS policies: trying to dump them |
| * requires getting a lock on their tables, and the calling user might not |
| * have privileges for that. We need no lock to examine a table's ACLs, |
| * so the current feature doesn't have a problem of that sort. |
| */ |
| if (fout->dopt->binary_upgrade) |
| dobj->dump = ext->dobj.dump; |
| else |
| { |
| if (fout->remoteVersion < 90600) |
| dobj->dump = DUMP_COMPONENT_NONE; |
| else |
| dobj->dump = ext->dobj.dump_contains & (DUMP_COMPONENT_ACL); |
| } |
| |
| return true; |
| } |
| |
| /* |
| * selectDumpableNamespace: policy-setting subroutine |
| * Mark a namespace as to be dumped or not |
| * |
| * Normally, we dump all extensions, or none of them if include_everything |
| * is false (i.e., a --schema or --table switch was given). However, in |
| * binary-upgrade mode it's necessary to skip built-in extensions, since we |
| * assume those will already be installed in the target database. We identify |
| * such extensions by their having OIDs in the range reserved for initdb. |
| */ |
| static void |
| selectDumpableNamespace(NamespaceInfo *nsinfo, Archive *fout) |
| { |
| DumpOptions *dopt = fout->dopt; |
| |
| /* |
| * DUMP_COMPONENT_DEFINITION typically implies a CREATE SCHEMA statement |
| * and (for --clean) a DROP SCHEMA statement. (In the absence of |
| * DUMP_COMPONENT_DEFINITION, this value is irrelevant.) |
| */ |
| nsinfo->create = true; |
| |
| /* |
| * If specific tables are being dumped, do not dump any complete |
| * namespaces. If specific namespaces are being dumped, dump just those |
| * namespaces. Otherwise, dump all non-system namespaces. |
| * |
| * GPDB: dump the gp_toolkit schema only in binary upgrade, where the |
| * new cluster expects the schema of an extension to exist before it |
| * creates the *empty* extension and copy the objects from old cluster. |
| * In non-binary-upgrade cases, gp_toolkit schema should be created when |
| * creating the extension itself, so no need to dump it; in fact, |
| * dumping it could result in error during replaying the script on a new |
| * database, because new database would inherit the gp_toolkit schema |
| * and extension from 'template1' already. |
| */ |
| if (table_include_oids.head != NULL) |
| nsinfo->dobj.dump_contains = nsinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| else if (schema_include_oids.head != NULL) |
| nsinfo->dobj.dump_contains = nsinfo->dobj.dump = |
| simple_oid_list_member(&schema_include_oids, |
| nsinfo->dobj.catId.oid) ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| else if (fout->remoteVersion >= 90600 && |
| strcmp(nsinfo->dobj.name, "pg_catalog") == 0) |
| { |
| /* |
| * In 9.6 and above, we dump out any ACLs defined in pg_catalog, if |
| * they are interesting (and not the original ACLs which were set at |
| * initdb time, see pg_init_privs). |
| */ |
| nsinfo->dobj.dump_contains = nsinfo->dobj.dump = DUMP_COMPONENT_ACL; |
| } |
| else if (strncmp(nsinfo->dobj.name, "pg_", 3) == 0 || |
| strcmp(nsinfo->dobj.name, "information_schema") == 0 || |
| (!dopt->binary_upgrade && strcmp(nsinfo->dobj.name, "gp_toolkit") == 0)) |
| { |
| /* Other system schemas don't get dumped */ |
| nsinfo->dobj.dump_contains = nsinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| } |
| else if (strcmp(nsinfo->dobj.name, "public") == 0) |
| { |
| /* |
| * The public schema is a strange beast that sits in a sort of |
| <<<<<<< HEAD |
| * no-mans-land between being a system object and a user object. We |
| * don't want to dump creation or comment commands for it, because |
| * that complicates matters for non-superuser use of pg_dump. But we |
| * should dump any ACL changes that have occurred for it, and of |
| * course we should dump contained objects. |
| * |
| * In Cloudberry binary upgrades, we need to dump the public schema in |
| * order to maintain its Oid. |
| */ |
| if (dopt->binary_upgrade) |
| nsinfo->dobj.dump = DUMP_COMPONENT_ALL; |
| else |
| nsinfo->dobj.dump = DUMP_COMPONENT_ACL; |
| ======= |
| * no-mans-land between being a system object and a user object. |
| * CREATE SCHEMA would fail, so its DUMP_COMPONENT_DEFINITION is just |
| * a comment and an indication of ownership. If the owner is the |
| * default, omit that superfluous DUMP_COMPONENT_DEFINITION. Before |
| * v15, the default owner was BOOTSTRAP_SUPERUSERID. |
| */ |
| nsinfo->create = false; |
| nsinfo->dobj.dump = DUMP_COMPONENT_ALL; |
| if (nsinfo->nspowner == ROLE_PG_DATABASE_OWNER) |
| nsinfo->dobj.dump &= ~DUMP_COMPONENT_DEFINITION; |
| >>>>>>> REL_16_9 |
| nsinfo->dobj.dump_contains = DUMP_COMPONENT_ALL; |
| |
| /* |
| * Also, make like it has a comment even if it doesn't; this is so |
| * that we'll emit a command to drop the comment, if appropriate. |
| * (Without this, we'd not call dumpCommentExtended for it.) |
| */ |
| nsinfo->dobj.components |= DUMP_COMPONENT_COMMENT; |
| } |
| else |
| nsinfo->dobj.dump_contains = nsinfo->dobj.dump = DUMP_COMPONENT_ALL; |
| |
| /* |
| * In any case, a namespace can be excluded by an exclusion switch |
| */ |
| if (nsinfo->dobj.dump_contains && |
| simple_oid_list_member(&schema_exclude_oids, |
| nsinfo->dobj.catId.oid)) |
| nsinfo->dobj.dump_contains = nsinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| |
| /* |
| * If the schema belongs to an extension, allow extension membership to |
| * override the dump decision for the schema itself. However, this does |
| * not change dump_contains, so this won't change what we do with objects |
| * within the schema. (If they belong to the extension, they'll get |
| * suppressed by it, otherwise not.) |
| */ |
| (void) checkExtensionMembership(&nsinfo->dobj, fout); |
| } |
| |
| /* |
| * selectDumpableTable: policy-setting subroutine |
| * Mark a table as to be dumped or not |
| */ |
| static void |
| selectDumpableTable(TableInfo *tbinfo, Archive *fout) |
| { |
| if (checkExtensionMembership(&tbinfo->dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| /* |
| * If specific tables are being dumped, dump just those tables; else, dump |
| * according to the parent namespace's dump flag. |
| */ |
| if (table_include_oids.head != NULL) |
| tbinfo->dobj.dump = simple_oid_list_member(&table_include_oids, |
| tbinfo->dobj.catId.oid) ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| else |
| tbinfo->dobj.dump = tbinfo->dobj.namespace->dobj.dump_contains; |
| |
| /* |
| * In any case, a table can be excluded by an exclusion switch |
| */ |
| if (tbinfo->dobj.dump && |
| simple_oid_list_member(&table_exclude_oids, |
| tbinfo->dobj.catId.oid)) |
| tbinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| |
| /* |
| * Pax not support pg_dump yet |
| */ |
| if (tbinfo->amoid == PAX_AM_OID) { |
| tbinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| |
| pg_log_warning("unsupport am pax yet, current relation \"%s\" will be ignore", |
| tbinfo->dobj.name); |
| } |
| } |
| |
| /* |
| * selectDumpableType: policy-setting subroutine |
| * Mark a type as to be dumped or not |
| * |
| * If it's a table's rowtype or an autogenerated array type, we also apply a |
| * special type code to facilitate sorting into the desired order. (We don't |
| * want to consider those to be ordinary types because that would bring tables |
| * up into the datatype part of the dump order.) We still set the object's |
| * dump flag; that's not going to cause the dummy type to be dumped, but we |
| * need it so that casts involving such types will be dumped correctly -- see |
| * dumpCast. This means the flag should be set the same as for the underlying |
| * object (the table or base type). |
| */ |
| static void |
| selectDumpableType(TypeInfo *tyinfo, Archive *fout) |
| { |
| /* skip complex types, except for standalone composite types */ |
| if (OidIsValid(tyinfo->typrelid) && |
| tyinfo->typrelkind != RELKIND_COMPOSITE_TYPE) |
| { |
| TableInfo *tytable = findTableByOid(tyinfo->typrelid); |
| |
| tyinfo->dobj.objType = DO_DUMMY_TYPE; |
| if (tytable != NULL) |
| tyinfo->dobj.dump = tytable->dobj.dump; |
| else |
| tyinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| return; |
| } |
| |
| /* skip auto-generated array types */ |
| if (tyinfo->isArray || tyinfo->isMultirange) |
| { |
| tyinfo->dobj.objType = DO_DUMMY_TYPE; |
| |
| /* |
| * Fall through to set the dump flag; we assume that the subsequent |
| * rules will do the same thing as they would for the array's base |
| * type. (We cannot reliably look up the base type here, since |
| * getTypes may not have processed it yet.) |
| */ |
| } |
| |
| if (checkExtensionMembership(&tyinfo->dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| /* Dump based on if the contents of the namespace are being dumped */ |
| tyinfo->dobj.dump = tyinfo->dobj.namespace->dobj.dump_contains; |
| } |
| |
| /* |
| * selectDumpableFunction: policy-setting subroutine |
| * Mark a function as to be dumped or not |
| */ |
| static void |
| selectDumpableFunction(FuncInfo *finfo) |
| { |
| /* |
| * If specific functions are being dumped, dump just those functions; else, dump |
| * according to the parent namespace's dump flag if parent namespace is not null; |
| * else, always dump the function. |
| */ |
| if (function_include_oids.head != NULL) |
| finfo->dobj.dump = simple_oid_list_member(&function_include_oids, |
| finfo->dobj.catId.oid); |
| else if (finfo->dobj.namespace) |
| finfo->dobj.dump = finfo->dobj.namespace->dobj.dump; |
| else |
| finfo->dobj.dump = true; |
| } |
| |
| /* |
| * selectDumpableDefaultACL: policy-setting subroutine |
| * Mark a default ACL as to be dumped or not |
| * |
| * For per-schema default ACLs, dump if the schema is to be dumped. |
| * Otherwise dump if we are dumping "everything". Note that dataOnly |
| * and aclsSkip are checked separately. |
| */ |
| static void |
| selectDumpableDefaultACL(DefaultACLInfo *dinfo, DumpOptions *dopt) |
| { |
| /* Default ACLs can't be extension members */ |
| |
| if (dinfo->dobj.namespace) |
| /* default ACLs are considered part of the namespace */ |
| dinfo->dobj.dump = dinfo->dobj.namespace->dobj.dump_contains; |
| else |
| dinfo->dobj.dump = dopt->include_everything ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| } |
| |
| /* |
| * selectDumpableCast: policy-setting subroutine |
| * Mark a cast as to be dumped or not |
| * |
| * Casts do not belong to any particular namespace (since they haven't got |
| * names), nor do they have identifiable owners. To distinguish user-defined |
| * casts from built-in ones, we must resort to checking whether the cast's |
| * OID is in the range reserved for initdb. |
| */ |
| static void |
| selectDumpableCast(CastInfo *cast, Archive *fout) |
| { |
| if (checkExtensionMembership(&cast->dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| /* |
| * This would be DUMP_COMPONENT_ACL for from-initdb casts, but they do not |
| * support ACLs currently. |
| */ |
| if (cast->dobj.catId.oid <= (Oid) g_last_builtin_oid) |
| cast->dobj.dump = DUMP_COMPONENT_NONE; |
| else |
| cast->dobj.dump = fout->dopt->include_everything ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| } |
| |
| /* |
| * selectDumpableProcLang: policy-setting subroutine |
| * Mark a procedural language as to be dumped or not |
| * |
| * Procedural languages do not belong to any particular namespace. To |
| * identify built-in languages, we must resort to checking whether the |
| * language's OID is in the range reserved for initdb. |
| */ |
| static void |
| selectDumpableProcLang(ProcLangInfo *plang, Archive *fout) |
| { |
| if (checkExtensionMembership(&plang->dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| /* |
| * Only include procedural languages when we are dumping everything. |
| * |
| * For from-initdb procedural languages, only include ACLs, as we do for |
| * the pg_catalog namespace. We need this because procedural languages do |
| * not live in any namespace. |
| */ |
| if (!fout->dopt->include_everything) |
| plang->dobj.dump = DUMP_COMPONENT_NONE; |
| else |
| { |
| if (plang->dobj.catId.oid <= (Oid) g_last_builtin_oid) |
| plang->dobj.dump = fout->remoteVersion < 90600 ? |
| DUMP_COMPONENT_NONE : DUMP_COMPONENT_ACL; |
| else |
| plang->dobj.dump = DUMP_COMPONENT_ALL; |
| } |
| } |
| |
| /* |
| * selectDumpableAccessMethod: policy-setting subroutine |
| * Mark an access method as to be dumped or not |
| * |
| * Access methods do not belong to any particular namespace. To identify |
| * built-in access methods, we must resort to checking whether the |
| * method's OID is in the range reserved for initdb. |
| */ |
| static void |
| selectDumpableAccessMethod(AccessMethodInfo *method, Archive *fout) |
| { |
| if (checkExtensionMembership(&method->dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| /* |
| * This would be DUMP_COMPONENT_ACL for from-initdb access methods, but |
| * they do not support ACLs currently. |
| */ |
| if (method->dobj.catId.oid <= (Oid) g_last_builtin_oid) |
| method->dobj.dump = DUMP_COMPONENT_NONE; |
| else |
| method->dobj.dump = fout->dopt->include_everything ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| } |
| |
| /* |
| * selectDumpableExtension: policy-setting subroutine |
| * Mark an extension as to be dumped or not |
| * |
| * Built-in extensions should be skipped except for checking ACLs, since we |
| * assume those will already be installed in the target database. We identify |
| * such extensions by their having OIDs in the range reserved for initdb. |
| * We dump all user-added extensions by default. No extensions are dumped |
| * if include_everything is false (i.e., a --schema or --table switch was |
| * given), except if --extension specifies a list of extensions to dump. |
| */ |
| static void |
| selectDumpableExtension(ExtensionInfo *extinfo, DumpOptions *dopt) |
| { |
| /* |
| * Use DUMP_COMPONENT_ACL for built-in extensions, to allow users to |
| * change permissions on their member objects, if they wish to, and have |
| * those changes preserved. |
| */ |
| if (extinfo->dobj.catId.oid <= (Oid) g_last_builtin_oid) |
| extinfo->dobj.dump = extinfo->dobj.dump_contains = DUMP_COMPONENT_ACL; |
| else |
| { |
| /* check if there is a list of extensions to dump */ |
| if (extension_include_oids.head != NULL) |
| extinfo->dobj.dump = extinfo->dobj.dump_contains = |
| simple_oid_list_member(&extension_include_oids, |
| extinfo->dobj.catId.oid) ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| else |
| extinfo->dobj.dump = extinfo->dobj.dump_contains = |
| dopt->include_everything ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| } |
| } |
| |
| /* |
| * selectDumpablePublicationObject: policy-setting subroutine |
| * Mark a publication object as to be dumped or not |
| * |
| * A publication can have schemas and tables which have schemas, but those are |
| * ignored in decision making, because publications are only dumped when we are |
| * dumping everything. |
| */ |
| static void |
| selectDumpablePublicationObject(DumpableObject *dobj, Archive *fout) |
| { |
| if (checkExtensionMembership(dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| dobj->dump = fout->dopt->include_everything ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| } |
| |
| /* |
| * selectDumpableStatisticsObject: policy-setting subroutine |
| * Mark an extended statistics object as to be dumped or not |
| * |
| * We dump an extended statistics object if the schema it's in and the table |
| * it's for are being dumped. (This'll need more thought if statistics |
| * objects ever support cross-table stats.) |
| */ |
| static void |
| selectDumpableStatisticsObject(StatsExtInfo *sobj, Archive *fout) |
| { |
| if (checkExtensionMembership(&sobj->dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| sobj->dobj.dump = sobj->dobj.namespace->dobj.dump_contains; |
| if (sobj->stattable == NULL || |
| !(sobj->stattable->dobj.dump & DUMP_COMPONENT_DEFINITION)) |
| sobj->dobj.dump = DUMP_COMPONENT_NONE; |
| } |
| |
| /* |
| * selectDumpableObject: policy-setting subroutine |
| * Mark a generic dumpable object as to be dumped or not |
| * |
| * Use this only for object types without a special-case routine above. |
| */ |
| static void |
| selectDumpableObject(DumpableObject *dobj, Archive *fout) |
| { |
| if (checkExtensionMembership(dobj, fout)) |
| return; /* extension membership overrides all else */ |
| |
| /* |
| * Default policy is to dump if parent namespace is dumpable, or for |
| * non-namespace-associated items, dump if we're dumping "everything". |
| */ |
| if (dobj->namespace) |
| dobj->dump = dobj->namespace->dobj.dump_contains; |
| else |
| dobj->dump = fout->dopt->include_everything ? |
| DUMP_COMPONENT_ALL : DUMP_COMPONENT_NONE; |
| } |
| |
| /* |
| * Dump a table's contents for loading using the COPY command |
| * - this routine is called by the Archiver when it wants the table |
| * to be dumped. |
| */ |
| static int |
| dumpTableData_copy(Archive *fout, const void *dcontext) |
| { |
| TableDataInfo *tdinfo = (TableDataInfo *) dcontext; |
| TableInfo *tbinfo = tdinfo->tdtable; |
| const char *classname = tbinfo->dobj.name; |
| PQExpBuffer q = createPQExpBuffer(); |
| |
| /* |
| * Note: can't use getThreadLocalPQExpBuffer() here, we're calling fmtId |
| * which uses it already. |
| */ |
| PQExpBuffer clistBuf = createPQExpBuffer(); |
| PGconn *conn = GetConnection(fout); |
| PGresult *res; |
| int ret; |
| char *copybuf; |
| const char *column_list; |
| |
| pg_log_info("dumping contents of table \"%s.%s\"", |
| tbinfo->dobj.namespace->dobj.name, classname); |
| |
| /* |
| * Specify the column list explicitly so that we have no possibility of |
| * retrieving data in the wrong column order. (The default column |
| * ordering of COPY will not be what we want in certain corner cases |
| * involving ADD COLUMN and inheritance.) |
| */ |
| column_list = fmtCopyColumnList(tbinfo, clistBuf); |
| |
| /* |
| * Use COPY (SELECT ...) TO when dumping a foreign table's data, and when |
| * a filter condition was specified. For other cases a simple COPY |
| * suffices. |
| */ |
| if (tdinfo->filtercond || tbinfo->relkind == RELKIND_FOREIGN_TABLE) |
| { |
| <<<<<<< HEAD |
| ======= |
| /* Temporary allows to access to foreign tables to dump data */ |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE) |
| set_restrict_relation_kind(fout, "view"); |
| |
| >>>>>>> REL_16_9 |
| appendPQExpBufferStr(q, "COPY (SELECT "); |
| /* klugery to get rid of parens in column list */ |
| if (strlen(column_list) > 2) |
| { |
| appendPQExpBufferStr(q, column_list + 1); |
| q->data[q->len - 1] = ' '; |
| } |
| else |
| appendPQExpBufferStr(q, "* "); |
| |
| appendPQExpBuffer(q, "FROM %s %s) TO stdout;", |
| fmtQualifiedDumpable(tbinfo), |
| tdinfo->filtercond ? tdinfo->filtercond : ""); |
| } |
| else |
| { |
| appendPQExpBuffer(q, "COPY %s %s TO stdout;", |
| fmtQualifiedDumpable(tbinfo), |
| column_list); |
| } |
| res = ExecuteSqlQuery(fout, q->data, PGRES_COPY_OUT); |
| PQclear(res); |
| destroyPQExpBuffer(clistBuf); |
| |
| for (;;) |
| { |
| ret = PQgetCopyData(conn, ©buf, 0); |
| |
| if (ret < 0) |
| break; /* done or error */ |
| |
| if (copybuf) |
| { |
| WriteData(fout, copybuf, ret); |
| PQfreemem(copybuf); |
| } |
| |
| /* ---------- |
| * THROTTLE: |
| * |
| * There was considerable discussion in late July, 2000 regarding |
| * slowing down pg_dump when backing up large tables. Users with both |
| * slow & fast (multi-processor) machines experienced performance |
| * degradation when doing a backup. |
| * |
| * Initial attempts based on sleeping for a number of ms for each ms |
| * of work were deemed too complex, then a simple 'sleep in each loop' |
| * implementation was suggested. The latter failed because the loop |
| * was too tight. Finally, the following was implemented: |
| * |
| * If throttle is non-zero, then |
| * See how long since the last sleep. |
| * Work out how long to sleep (based on ratio). |
| * If sleep is more than 100ms, then |
| * sleep |
| * reset timer |
| * EndIf |
| * EndIf |
| * |
| * where the throttle value was the number of ms to sleep per ms of |
| * work. The calculation was done in each loop. |
| * |
| * Most of the hard work is done in the backend, and this solution |
| * still did not work particularly well: on slow machines, the ratio |
| * was 50:1, and on medium paced machines, 1:1, and on fast |
| * multi-processor machines, it had little or no effect, for reasons |
| * that were unclear. |
| * |
| * Further discussion ensued, and the proposal was dropped. |
| * |
| * For those people who want this feature, it can be implemented using |
| * gettimeofday in each loop, calculating the time since last sleep, |
| * multiplying that by the sleep ratio, then if the result is more |
| * than a preset 'minimum sleep time' (say 100ms), call the 'select' |
| * function to sleep for a subsecond period ie. |
| * |
| * select(0, NULL, NULL, NULL, &tvi); |
| * |
| * This will return after the interval specified in the structure tvi. |
| * Finally, call gettimeofday again to save the 'last sleep time'. |
| * ---------- |
| */ |
| } |
| archprintf(fout, "\\.\n\n\n"); |
| |
| if (ret == -2) |
| { |
| /* copy data transfer failed */ |
| pg_log_error("Dumping the contents of table \"%s\" failed: PQgetCopyData() failed.", classname); |
| pg_log_error_detail("Error message from server: %s", PQerrorMessage(conn)); |
| pg_log_error_detail("Command was: %s", q->data); |
| exit_nicely(1); |
| } |
| |
| /* Check command status and return to normal libpq state */ |
| res = PQgetResult(conn); |
| if (PQresultStatus(res) != PGRES_COMMAND_OK) |
| { |
| pg_log_error("Dumping the contents of table \"%s\" failed: PQgetResult() failed.", classname); |
| pg_log_error_detail("Error message from server: %s", PQerrorMessage(conn)); |
| pg_log_error_detail("Command was: %s", q->data); |
| exit_nicely(1); |
| } |
| PQclear(res); |
| |
| /* Do this to ensure we've pumped libpq back to idle state */ |
| if (PQgetResult(conn) != NULL) |
| pg_log_warning("unexpected extra results during COPY of table \"%s\"", |
| classname); |
| |
| destroyPQExpBuffer(q); |
| |
| /* Revert back the setting */ |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE) |
| set_restrict_relation_kind(fout, "view, foreign-table"); |
| |
| return 1; |
| } |
| |
| /* |
| * Dump table data using INSERT commands. |
| * |
| * Caution: when we restore from an archive file direct to database, the |
| * INSERT commands emitted by this function have to be parsed by |
| * pg_backup_db.c's ExecuteSimpleCommands(), which will not handle comments, |
| * E'' strings, or dollar-quoted strings. So don't emit anything like that. |
| */ |
| static int |
| dumpTableData_insert(Archive *fout, const void *dcontext) |
| { |
| TableDataInfo *tdinfo = (TableDataInfo *) dcontext; |
| TableInfo *tbinfo = tdinfo->tdtable; |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer insertStmt = NULL; |
| char *attgenerated; |
| PGresult *res; |
| int nfields, |
| i; |
| int rows_per_statement = dopt->dump_inserts; |
| int rows_this_statement = 0; |
| |
| <<<<<<< HEAD |
| ======= |
| /* Temporary allows to access to foreign tables to dump data */ |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE) |
| set_restrict_relation_kind(fout, "view"); |
| |
| >>>>>>> REL_16_9 |
| /* |
| * If we're going to emit INSERTs with column names, the most efficient |
| * way to deal with generated columns is to exclude them entirely. For |
| * INSERTs without column names, we have to emit DEFAULT rather than the |
| * actual column value --- but we can save a few cycles by fetching nulls |
| * rather than the uninteresting-to-us value. |
| */ |
| attgenerated = (char *) pg_malloc(tbinfo->numatts * sizeof(char)); |
| appendPQExpBufferStr(q, "DECLARE _pg_dump_cursor CURSOR FOR SELECT "); |
| nfields = 0; |
| for (i = 0; i < tbinfo->numatts; i++) |
| { |
| if (tbinfo->attisdropped[i]) |
| continue; |
| if (tbinfo->attgenerated[i] && dopt->column_inserts) |
| continue; |
| if (nfields > 0) |
| appendPQExpBufferStr(q, ", "); |
| if (tbinfo->attgenerated[i]) |
| appendPQExpBufferStr(q, "NULL"); |
| else |
| appendPQExpBufferStr(q, fmtId(tbinfo->attnames[i])); |
| attgenerated[nfields] = tbinfo->attgenerated[i]; |
| nfields++; |
| } |
| /* Servers before 9.4 will complain about zero-column SELECT */ |
| if (nfields == 0) |
| appendPQExpBufferStr(q, "NULL"); |
| appendPQExpBuffer(q, " FROM ONLY %s", |
| fmtQualifiedDumpable(tbinfo)); |
| if (tdinfo->filtercond) |
| appendPQExpBuffer(q, " %s", tdinfo->filtercond); |
| |
| ExecuteSqlStatement(fout, q->data); |
| |
| while (1) |
| { |
| res = ExecuteSqlQuery(fout, "FETCH 100 FROM _pg_dump_cursor", |
| PGRES_TUPLES_OK); |
| |
| /* cross-check field count, allowing for dummy NULL if any */ |
| if (nfields != PQnfields(res) && |
| !(nfields == 0 && PQnfields(res) == 1)) |
| <<<<<<< HEAD |
| fatal("wrong number of fields retrieved from table \"%s\"", |
| tbinfo->dobj.name); |
| ======= |
| pg_fatal("wrong number of fields retrieved from table \"%s\"", |
| tbinfo->dobj.name); |
| >>>>>>> REL_16_9 |
| |
| /* |
| * First time through, we build as much of the INSERT statement as |
| * possible in "insertStmt", which we can then just print for each |
| * statement. If the table happens to have zero dumpable columns then |
| * this will be a complete statement, otherwise it will end in |
| * "VALUES" and be ready to have the row's column values printed. |
| */ |
| if (insertStmt == NULL) |
| { |
| TableInfo *targettab; |
| |
| insertStmt = createPQExpBuffer(); |
| |
| /* |
| * When load-via-partition-root is set or forced, get the root |
| * table name for the partition table, so that we can reload data |
| * through the root table. |
| */ |
| if (tbinfo->ispartition && |
| (dopt->load_via_partition_root || |
| forcePartitionRootLoad(tbinfo))) |
| targettab = getRootTableInfo(tbinfo); |
| else |
| targettab = tbinfo; |
| |
| appendPQExpBuffer(insertStmt, "INSERT INTO %s ", |
| fmtQualifiedDumpable(targettab)); |
| |
| /* corner case for zero-column table */ |
| if (nfields == 0) |
| { |
| appendPQExpBufferStr(insertStmt, "DEFAULT VALUES;\n"); |
| } |
| else |
| { |
| /* append the list of column names if required */ |
| if (dopt->column_inserts) |
| { |
| appendPQExpBufferChar(insertStmt, '('); |
| for (int field = 0; field < nfields; field++) |
| { |
| if (field > 0) |
| appendPQExpBufferStr(insertStmt, ", "); |
| appendPQExpBufferStr(insertStmt, |
| fmtId(PQfname(res, field))); |
| } |
| appendPQExpBufferStr(insertStmt, ") "); |
| } |
| |
| if (tbinfo->needs_override) |
| appendPQExpBufferStr(insertStmt, "OVERRIDING SYSTEM VALUE "); |
| |
| appendPQExpBufferStr(insertStmt, "VALUES"); |
| } |
| } |
| |
| for (int tuple = 0; tuple < PQntuples(res); tuple++) |
| { |
| /* Write the INSERT if not in the middle of a multi-row INSERT. */ |
| if (rows_this_statement == 0) |
| archputs(insertStmt->data, fout); |
| |
| /* |
| * If it is zero-column table then we've already written the |
| * complete statement, which will mean we've disobeyed |
| * --rows-per-insert when it's set greater than 1. We do support |
| * a way to make this multi-row with: SELECT UNION ALL SELECT |
| * UNION ALL ... but that's non-standard so we should avoid it |
| * given that using INSERTs is mostly only ever needed for |
| * cross-database exports. |
| */ |
| if (nfields == 0) |
| continue; |
| |
| /* Emit a row heading */ |
| if (rows_per_statement == 1) |
| archputs(" (", fout); |
| else if (rows_this_statement > 0) |
| archputs(",\n\t(", fout); |
| else |
| archputs("\n\t(", fout); |
| |
| for (int field = 0; field < nfields; field++) |
| { |
| if (field > 0) |
| archputs(", ", fout); |
| if (attgenerated[field]) |
| { |
| archputs("DEFAULT", fout); |
| continue; |
| } |
| if (PQgetisnull(res, tuple, field)) |
| { |
| archputs("NULL", fout); |
| continue; |
| } |
| |
| /* XXX This code is partially duplicated in ruleutils.c */ |
| switch (PQftype(res, field)) |
| { |
| case INT2OID: |
| case INT4OID: |
| case INT8OID: |
| case OIDOID: |
| case FLOAT4OID: |
| case FLOAT8OID: |
| case NUMERICOID: |
| { |
| /* |
| * These types are printed without quotes unless |
| * they contain values that aren't accepted by the |
| * scanner unquoted (e.g., 'NaN'). Note that |
| * strtod() and friends might accept NaN, so we |
| * can't use that to test. |
| * |
| * In reality we only need to defend against |
| * infinity and NaN, so we need not get too crazy |
| * about pattern matching here. |
| */ |
| const char *s = PQgetvalue(res, tuple, field); |
| |
| if (strspn(s, "0123456789 +-eE.") == strlen(s)) |
| archputs(s, fout); |
| else |
| archprintf(fout, "'%s'", s); |
| } |
| break; |
| |
| case BITOID: |
| case VARBITOID: |
| archprintf(fout, "B'%s'", |
| PQgetvalue(res, tuple, field)); |
| break; |
| |
| case BOOLOID: |
| if (strcmp(PQgetvalue(res, tuple, field), "t") == 0) |
| archputs("true", fout); |
| else |
| archputs("false", fout); |
| break; |
| |
| default: |
| /* All other types are printed as string literals. */ |
| resetPQExpBuffer(q); |
| appendStringLiteralAH(q, |
| PQgetvalue(res, tuple, field), |
| fout); |
| archputs(q->data, fout); |
| break; |
| } |
| } |
| |
| /* Terminate the row ... */ |
| archputs(")", fout); |
| |
| /* ... and the statement, if the target no. of rows is reached */ |
| if (++rows_this_statement >= rows_per_statement) |
| { |
| if (dopt->do_nothing) |
| archputs(" ON CONFLICT DO NOTHING;\n", fout); |
| else |
| archputs(";\n", fout); |
| /* Reset the row counter */ |
| rows_this_statement = 0; |
| } |
| } |
| |
| if (PQntuples(res) <= 0) |
| { |
| PQclear(res); |
| break; |
| } |
| PQclear(res); |
| } |
| |
| /* Terminate any statements that didn't make the row count. */ |
| if (rows_this_statement > 0) |
| { |
| if (dopt->do_nothing) |
| archputs(" ON CONFLICT DO NOTHING;\n", fout); |
| else |
| archputs(";\n", fout); |
| } |
| |
| archputs("\n\n", fout); |
| |
| ExecuteSqlStatement(fout, "CLOSE _pg_dump_cursor"); |
| |
| destroyPQExpBuffer(q); |
| if (insertStmt != NULL) |
| destroyPQExpBuffer(insertStmt); |
| free(attgenerated); |
| <<<<<<< HEAD |
| ======= |
| |
| /* Revert back the setting */ |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE) |
| set_restrict_relation_kind(fout, "view, foreign-table"); |
| >>>>>>> REL_16_9 |
| |
| return 1; |
| } |
| |
| /* |
| * getRootTableInfo: |
| * get the root TableInfo for the given partition table. |
| */ |
| static TableInfo * |
| getRootTableInfo(const TableInfo *tbinfo) |
| { |
| TableInfo *parentTbinfo; |
| |
| Assert(tbinfo->ispartition); |
| Assert(tbinfo->numParents == 1); |
| |
| parentTbinfo = tbinfo->parents[0]; |
| while (parentTbinfo->ispartition) |
| { |
| Assert(parentTbinfo->numParents == 1); |
| parentTbinfo = parentTbinfo->parents[0]; |
| } |
| |
| return parentTbinfo; |
| } |
| |
| /* |
| * forcePartitionRootLoad |
| * Check if we must force load_via_partition_root for this partition. |
| * |
| * This is required if any level of ancestral partitioned table has an |
| * unsafe partitioning scheme. |
| */ |
| static bool |
| forcePartitionRootLoad(const TableInfo *tbinfo) |
| { |
| TableInfo *parentTbinfo; |
| |
| Assert(tbinfo->ispartition); |
| Assert(tbinfo->numParents == 1); |
| |
| parentTbinfo = tbinfo->parents[0]; |
| if (parentTbinfo->unsafe_partitions) |
| return true; |
| while (parentTbinfo->ispartition) |
| { |
| Assert(parentTbinfo->numParents == 1); |
| parentTbinfo = parentTbinfo->parents[0]; |
| if (parentTbinfo->unsafe_partitions) |
| return true; |
| } |
| |
| return false; |
| } |
| |
| /* |
| * dumpTableData - |
| * dump the contents of a single table |
| * |
| * Actually, this just makes an ArchiveEntry for the table contents. |
| */ |
| static void |
| dumpTableData(Archive *fout, const TableDataInfo *tdinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = tdinfo->tdtable; |
| PQExpBuffer copyBuf = createPQExpBuffer(); |
| PQExpBuffer clistBuf = createPQExpBuffer(); |
| DataDumperPtr dumpFn; |
| char *tdDefn = NULL; |
| char *copyStmt; |
| const char *copyFrom; |
| |
| /* We had better have loaded per-column details about this table */ |
| Assert(tbinfo->interesting); |
| |
| /* |
| * When load-via-partition-root is set or forced, get the root table name |
| * for the partition table, so that we can reload data through the root |
| * table. Then construct a comment to be inserted into the TOC entry's |
| * defn field, so that such cases can be identified reliably. |
| */ |
| if (tbinfo->ispartition && |
| (dopt->load_via_partition_root || |
| forcePartitionRootLoad(tbinfo))) |
| { |
| TableInfo *parentTbinfo; |
| |
| parentTbinfo = getRootTableInfo(tbinfo); |
| copyFrom = fmtQualifiedDumpable(parentTbinfo); |
| printfPQExpBuffer(copyBuf, "-- load via partition root %s", |
| copyFrom); |
| tdDefn = pg_strdup(copyBuf->data); |
| } |
| else |
| copyFrom = fmtQualifiedDumpable(tbinfo); |
| |
| if (dopt->dump_inserts == 0) |
| { |
| /* Dump/restore using COPY */ |
| dumpFn = dumpTableData_copy; |
| /* must use 2 steps here 'cause fmtId is nonreentrant */ |
| printfPQExpBuffer(copyBuf, "COPY %s ", |
| copyFrom); |
| appendPQExpBuffer(copyBuf, "%s FROM stdin;\n", |
| fmtCopyColumnList(tbinfo, clistBuf)); |
| copyStmt = copyBuf->data; |
| } |
| else |
| { |
| /* Restore using INSERT */ |
| dumpFn = dumpTableData_insert; |
| copyStmt = NULL; |
| } |
| |
| /* |
| * Note: although the TableDataInfo is a full DumpableObject, we treat its |
| * dependency on its table as "special" and pass it to ArchiveEntry now. |
| * See comments for BuildArchiveDependencies. |
| */ |
| if (tdinfo->dobj.dump & DUMP_COMPONENT_DATA) |
| { |
| TocEntry *te; |
| |
| te = ArchiveEntry(fout, tdinfo->dobj.catId, tdinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tbinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "TABLE DATA", |
| .section = SECTION_DATA, |
| .createStmt = tdDefn, |
| .copyStmt = copyStmt, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1, |
| .dumpFn = dumpFn, |
| .dumpArg = tdinfo)); |
| |
| /* |
| * Set the TocEntry's dataLength in case we are doing a parallel dump |
| * and want to order dump jobs by table size. We choose to measure |
| * dataLength in table pages (including TOAST pages) during dump, so |
| * no scaling is needed. |
| * |
| * However, relpages is declared as "integer" in pg_class, and hence |
| * also in TableInfo, but it's really BlockNumber a/k/a unsigned int. |
| * Cast so that we get the right interpretation of table sizes |
| * exceeding INT_MAX pages. |
| */ |
| te->dataLength = (BlockNumber) tbinfo->relpages; |
| te->dataLength += (BlockNumber) tbinfo->toastpages; |
| |
| /* |
| * If pgoff_t is only 32 bits wide, the above refinement is useless, |
| * and instead we'd better worry about integer overflow. Clamp to |
| * INT_MAX if the correct result exceeds that. |
| */ |
| if (sizeof(te->dataLength) == 4 && |
| (tbinfo->relpages < 0 || tbinfo->toastpages < 0 || |
| te->dataLength < 0)) |
| te->dataLength = INT_MAX; |
| } |
| |
| destroyPQExpBuffer(copyBuf); |
| destroyPQExpBuffer(clistBuf); |
| } |
| |
| /* |
| * refreshMatViewData - |
| * load or refresh the contents of a single materialized view |
| * |
| * Actually, this just makes an ArchiveEntry for the REFRESH MATERIALIZED VIEW |
| * statement. |
| */ |
| static void |
| refreshMatViewData(Archive *fout, const TableDataInfo *tdinfo) |
| { |
| TableInfo *tbinfo = tdinfo->tdtable; |
| PQExpBuffer q; |
| |
| /* If the materialized view is not flagged as populated, skip this. */ |
| if (!tbinfo->relispopulated) |
| return; |
| |
| q = createPQExpBuffer(); |
| |
| if (tbinfo->isdynamic) |
| { |
| appendPQExpBuffer(q, "REFRESH DYNAMIC TABLE %s;\n", |
| fmtQualifiedDumpable(tbinfo)); |
| } |
| else |
| { |
| appendPQExpBuffer(q, "REFRESH MATERIALIZED VIEW %s;\n", |
| fmtQualifiedDumpable(tbinfo)); |
| } |
| |
| if (tdinfo->dobj.dump & DUMP_COMPONENT_DATA) |
| ArchiveEntry(fout, |
| tdinfo->dobj.catId, /* catalog ID */ |
| tdinfo->dobj.dumpId, /* dump ID */ |
| ARCHIVE_OPTS(.tag = tbinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = tbinfo->isdynamic ? "DYNAMIC TABLE DATA": "MATERIALIZED VIEW DATA", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .deps = tdinfo->dobj.dependencies, |
| .nDeps = tdinfo->dobj.nDeps)); |
| |
| destroyPQExpBuffer(q); |
| } |
| |
| /* |
| * getTableData - |
| * set up dumpable objects representing the contents of tables |
| */ |
| static void |
| getTableData(DumpOptions *dopt, TableInfo *tblinfo, int numTables, char relkind) |
| { |
| int i; |
| |
| for (i = 0; i < numTables; i++) |
| { |
| if (tblinfo[i].dobj.dump & DUMP_COMPONENT_DATA && |
| (!relkind || tblinfo[i].relkind == relkind)) |
| makeTableDataInfo(dopt, &(tblinfo[i])); |
| } |
| } |
| |
| /* |
| * Make a dumpable object for the data of this specific table |
| * |
| * Note: we make a TableDataInfo if and only if we are going to dump the |
| * table data; the "dump" field in such objects isn't very interesting. |
| */ |
| static void |
| makeTableDataInfo(DumpOptions *dopt, TableInfo *tbinfo) |
| { |
| TableDataInfo *tdinfo; |
| |
| /* |
| * Nothing to do if we already decided to dump the table. This will |
| * happen for "config" tables. |
| */ |
| if (tbinfo->dataObj != NULL) |
| return; |
| |
| /* Skip VIEWs (no data to dump) */ |
| if (tbinfo->relkind == RELKIND_VIEW) |
| return; |
| /* Skip FOREIGN TABLEs (no data to dump) unless requested explicitly */ |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE && |
| (foreign_servers_include_oids.head == NULL || |
| !simple_oid_list_member(&foreign_servers_include_oids, |
| tbinfo->foreign_server))) |
| return; |
| /* Skip EXTERNAL TABLEs (like foreign tables in GPDB 6.x and below) */ |
| if (tbinfo->relstorage == RELSTORAGE_EXTERNAL) |
| return; |
| /* Skip partitioned tables (data in partitions) */ |
| if (tbinfo->relkind == RELKIND_PARTITIONED_TABLE) |
| return; |
| |
| /* Don't dump data in unlogged tables, if so requested */ |
| if (tbinfo->relpersistence == RELPERSISTENCE_UNLOGGED && |
| dopt->no_unlogged_table_data) |
| return; |
| |
| /* Check that the data is not explicitly excluded */ |
| if (simple_oid_list_member(&tabledata_exclude_oids, |
| tbinfo->dobj.catId.oid)) |
| return; |
| |
| /* OK, let's dump it */ |
| tdinfo = (TableDataInfo *) pg_malloc(sizeof(TableDataInfo)); |
| |
| if (tbinfo->relkind == RELKIND_MATVIEW) |
| tdinfo->dobj.objType = DO_REFRESH_MATVIEW; |
| else if (tbinfo->relkind == RELKIND_SEQUENCE) |
| tdinfo->dobj.objType = DO_SEQUENCE_SET; |
| else |
| tdinfo->dobj.objType = DO_TABLE_DATA; |
| |
| /* |
| * Note: use tableoid 0 so that this object won't be mistaken for |
| * something that pg_depend entries apply to. |
| */ |
| tdinfo->dobj.catId.tableoid = 0; |
| tdinfo->dobj.catId.oid = tbinfo->dobj.catId.oid; |
| AssignDumpId(&tdinfo->dobj); |
| tdinfo->dobj.name = tbinfo->dobj.name; |
| tdinfo->dobj.namespace = tbinfo->dobj.namespace; |
| tdinfo->tdtable = tbinfo; |
| tdinfo->filtercond = NULL; /* might get set later */ |
| addObjectDependency(&tdinfo->dobj, tbinfo->dobj.dumpId); |
| |
| /* A TableDataInfo contains data, of course */ |
| tdinfo->dobj.components |= DUMP_COMPONENT_DATA; |
| |
| tbinfo->dataObj = tdinfo; |
| |
| /* Make sure that we'll collect per-column info for this table. */ |
| tbinfo->interesting = true; |
| } |
| |
| /* |
| * The refresh for a materialized view must be dependent on the refresh for |
| * any materialized view that this one is dependent on. |
| * |
| * This must be called after all the objects are created, but before they are |
| * sorted. |
| */ |
| static void |
| buildMatViewRefreshDependencies(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| int ntups, |
| i; |
| int i_classid, |
| i_objid, |
| i_refobjid; |
| |
| /* No Mat Views before 9.3. */ |
| if (fout->remoteVersion < 90300) |
| return; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "WITH RECURSIVE w AS " |
| "( " |
| "SELECT d1.objid, d2.refobjid, c2.relkind AS refrelkind " |
| "FROM pg_depend d1 " |
| "JOIN pg_class c1 ON c1.oid = d1.objid " |
| "AND c1.relkind = " CppAsString2(RELKIND_MATVIEW) |
| " JOIN pg_rewrite r1 ON r1.ev_class = d1.objid " |
| "JOIN pg_depend d2 ON d2.classid = 'pg_rewrite'::regclass " |
| "AND d2.objid = r1.oid " |
| "AND d2.refobjid <> d1.objid " |
| "JOIN pg_class c2 ON c2.oid = d2.refobjid " |
| "AND c2.relkind IN (" CppAsString2(RELKIND_MATVIEW) "," |
| CppAsString2(RELKIND_VIEW) ") " |
| "WHERE d1.classid = 'pg_class'::regclass " |
| "UNION " |
| "SELECT w.objid, d3.refobjid, c3.relkind " |
| "FROM w " |
| "JOIN pg_rewrite r3 ON r3.ev_class = w.refobjid " |
| "JOIN pg_depend d3 ON d3.classid = 'pg_rewrite'::regclass " |
| "AND d3.objid = r3.oid " |
| "AND d3.refobjid <> w.refobjid " |
| "JOIN pg_class c3 ON c3.oid = d3.refobjid " |
| "AND c3.relkind IN (" CppAsString2(RELKIND_MATVIEW) "," |
| CppAsString2(RELKIND_VIEW) ") " |
| ") " |
| "SELECT 'pg_class'::regclass::oid AS classid, objid, refobjid " |
| "FROM w " |
| "WHERE refrelkind = " CppAsString2(RELKIND_MATVIEW)); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_classid = PQfnumber(res, "classid"); |
| i_objid = PQfnumber(res, "objid"); |
| i_refobjid = PQfnumber(res, "refobjid"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| CatalogId objId; |
| CatalogId refobjId; |
| DumpableObject *dobj; |
| DumpableObject *refdobj; |
| TableInfo *tbinfo; |
| TableInfo *reftbinfo; |
| |
| objId.tableoid = atooid(PQgetvalue(res, i, i_classid)); |
| objId.oid = atooid(PQgetvalue(res, i, i_objid)); |
| refobjId.tableoid = objId.tableoid; |
| refobjId.oid = atooid(PQgetvalue(res, i, i_refobjid)); |
| |
| dobj = findObjectByCatalogId(objId); |
| if (dobj == NULL) |
| continue; |
| |
| Assert(dobj->objType == DO_TABLE); |
| tbinfo = (TableInfo *) dobj; |
| Assert(tbinfo->relkind == RELKIND_MATVIEW); |
| dobj = (DumpableObject *) tbinfo->dataObj; |
| if (dobj == NULL) |
| continue; |
| Assert(dobj->objType == DO_REFRESH_MATVIEW); |
| |
| refdobj = findObjectByCatalogId(refobjId); |
| if (refdobj == NULL) |
| continue; |
| |
| Assert(refdobj->objType == DO_TABLE); |
| reftbinfo = (TableInfo *) refdobj; |
| Assert(reftbinfo->relkind == RELKIND_MATVIEW); |
| refdobj = (DumpableObject *) reftbinfo->dataObj; |
| if (refdobj == NULL) |
| continue; |
| Assert(refdobj->objType == DO_REFRESH_MATVIEW); |
| |
| addObjectDependency(dobj, refdobj->dumpId); |
| |
| if (!reftbinfo->relispopulated) |
| tbinfo->relispopulated = false; |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getTableDataFKConstraints - |
| * add dump-order dependencies reflecting foreign key constraints |
| * |
| * This code is executed only in a data-only dump --- in schema+data dumps |
| * we handle foreign key issues by not creating the FK constraints until |
| * after the data is loaded. In a data-only dump, however, we want to |
| * order the table data objects in such a way that a table's referenced |
| * tables are restored first. (In the presence of circular references or |
| * self-references this may be impossible; we'll detect and complain about |
| * that during the dependency sorting step.) |
| */ |
| static void |
| getTableDataFKConstraints(void) |
| { |
| DumpableObject **dobjs; |
| int numObjs; |
| int i; |
| |
| /* Search through all the dumpable objects for FK constraints */ |
| getDumpableObjects(&dobjs, &numObjs); |
| for (i = 0; i < numObjs; i++) |
| { |
| if (dobjs[i]->objType == DO_FK_CONSTRAINT) |
| { |
| ConstraintInfo *cinfo = (ConstraintInfo *) dobjs[i]; |
| TableInfo *ftable; |
| |
| /* Not interesting unless both tables are to be dumped */ |
| if (cinfo->contable == NULL || |
| cinfo->contable->dataObj == NULL) |
| continue; |
| ftable = findTableByOid(cinfo->confrelid); |
| if (ftable == NULL || |
| ftable->dataObj == NULL) |
| continue; |
| |
| /* |
| * Okay, make referencing table's TABLE_DATA object depend on the |
| * referenced table's TABLE_DATA object. |
| */ |
| addObjectDependency(&cinfo->contable->dataObj->dobj, |
| ftable->dataObj->dobj.dumpId); |
| } |
| } |
| free(dobjs); |
| } |
| |
| |
| /* |
| <<<<<<< HEAD |
| * guessConstraintInheritance: |
| * In pre-8.4 databases, we can't tell for certain which constraints |
| * are inherited. We assume a CHECK constraint is inherited if its name |
| * matches the name of any constraint in the parent. Originally this code |
| * tried to compare the expression texts, but that can fail for various |
| * reasons --- for example, if the parent and child tables are in different |
| * schemas, reverse-listing of function calls may produce different text |
| * (schema-qualified or not) depending on search path. |
| * |
| * In 8.4 and up we can rely on the conislocal field to decide which |
| * constraints must be dumped; much safer. |
| * |
| * This function assumes all conislocal flags were initialized to true. |
| * It clears the flag on anything that seems to be inherited. |
| */ |
| static void |
| guessConstraintInheritance(TableInfo *tblinfo, int numTables) |
| { |
| int i, |
| j, |
| k; |
| |
| for (i = 0; i < numTables; i++) |
| { |
| TableInfo *tbinfo = &(tblinfo[i]); |
| int numParents; |
| TableInfo **parents; |
| TableInfo *parent; |
| |
| /* Some kinds never have parents */ |
| if (tbinfo->relkind == RELKIND_SEQUENCE || |
| tbinfo->relkind == RELKIND_VIEW || |
| tbinfo->relkind == RELKIND_MATVIEW) |
| continue; |
| |
| /* Don't bother computing anything for non-target tables, either */ |
| if (!(tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION)) |
| continue; |
| |
| numParents = tbinfo->numParents; |
| parents = tbinfo->parents; |
| |
| if (numParents == 0) |
| continue; /* nothing to see here, move along */ |
| |
| /* scan for inherited CHECK constraints */ |
| for (j = 0; j < tbinfo->ncheck; j++) |
| { |
| ConstraintInfo *constr; |
| |
| constr = &(tbinfo->checkexprs[j]); |
| |
| for (k = 0; k < numParents; k++) |
| { |
| int l; |
| |
| parent = parents[k]; |
| for (l = 0; l < parent->ncheck; l++) |
| { |
| ConstraintInfo *pconstr = &(parent->checkexprs[l]); |
| |
| if (strcmp(pconstr->dobj.name, constr->dobj.name) == 0) |
| { |
| constr->conislocal = false; |
| break; |
| } |
| } |
| if (!constr->conislocal) |
| break; |
| } |
| } |
| } |
| } |
| |
| static void |
| dumpPreassignedOidArchiveEntry(Archive *fout, BinaryUpgradeInfo *binfo) |
| { |
| PQExpBuffer setoidquery; |
| CatalogId maxoidid; |
| |
| setoidquery = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(setoidquery, |
| "-- Placeholder for binary_upgrade_set_preassigned_oids()\n\n"); |
| |
| maxoidid.oid = 0; |
| maxoidid.tableoid = 0; |
| |
| char *tag = pg_strdup("binary_upgrade"); |
| |
| ArchiveEntry(fout, maxoidid, binfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .description = tag, |
| .section = SECTION_PRE_DATA, |
| .createStmt = setoidquery->data)); |
| |
| destroyPQExpBuffer(setoidquery); |
| free(tag); |
| } |
| |
| static void |
| dumpPreassignedOidDefinition(Archive *fout, BinaryUpgradeInfo *binfo) |
| { |
| PQExpBuffer setoidquery; |
| SimpleOidListCell *cell; |
| |
| setoidquery = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(setoidquery, |
| "SELECT binary_upgrade_set_preassigned_oids(ARRAY["); |
| for (cell = preassigned_oids.head; cell; cell = cell->next) |
| { |
| appendPQExpBuffer(setoidquery, "%u%s", |
| cell->val, (cell->next ? "," : "")); |
| } |
| appendPQExpBufferStr(setoidquery, "]::pg_catalog.oid[]);\n\n"); |
| |
| AmendArchiveEntry(fout, binfo->dobj.dumpId, setoidquery->data); |
| |
| destroyPQExpBuffer(setoidquery); |
| } |
| |
| /* |
| ======= |
| >>>>>>> REL_16_9 |
| * dumpDatabase: |
| * dump the database definition |
| */ |
| static void |
| dumpDatabase(Archive *fout) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer dbQry = createPQExpBuffer(); |
| PQExpBuffer delQry = createPQExpBuffer(); |
| PQExpBuffer creaQry = createPQExpBuffer(); |
| PQExpBuffer labelq = createPQExpBuffer(); |
| PGconn *conn = GetConnection(fout); |
| PGresult *res; |
| int i_tableoid, |
| i_oid, |
| i_datname, |
| i_datdba, |
| i_encoding, |
| i_datlocprovider, |
| i_collate, |
| i_ctype, |
| i_daticulocale, |
| i_daticurules, |
| i_frozenxid, |
| i_minmxid, |
| i_datacl, |
| i_acldefault, |
| i_datistemplate, |
| i_datconnlimit, |
| i_datcollversion, |
| i_tablespace; |
| CatalogId dbCatId; |
| DumpId dbDumpId; |
| DumpableAcl dbdacl; |
| const char *datname, |
| *dba, |
| *encoding, |
| *datlocprovider, |
| *collate, |
| *ctype, |
| *iculocale, |
| *icurules, |
| *datistemplate, |
| *datconnlimit, |
| *tablespace; |
| uint32 frozenxid, |
| minmxid; |
| char *qdatname; |
| |
| pg_log_info("saving database definition"); |
| |
| /* |
| * Fetch the database-level properties for this database. |
| */ |
| <<<<<<< HEAD |
| if (fout->remoteVersion >= 90600) |
| { |
| appendPQExpBuffer(dbQry, "SELECT tableoid, oid, datname, " |
| "(%s datdba) AS dba, " |
| "pg_encoding_to_char(encoding) AS encoding, " |
| "datcollate, datctype, datfrozenxid, datminmxid, " |
| "(SELECT array_agg(acl ORDER BY row_n) FROM " |
| " (SELECT acl, row_n FROM " |
| " unnest(coalesce(datacl,acldefault('d',datdba))) " |
| " WITH ORDINALITY AS perm(acl,row_n) " |
| " WHERE NOT EXISTS ( " |
| " SELECT 1 " |
| " FROM unnest(acldefault('d',datdba)) " |
| " AS init(init_acl) " |
| " WHERE acl = init_acl)) AS datacls) " |
| " AS datacl, " |
| "(SELECT array_agg(acl ORDER BY row_n) FROM " |
| " (SELECT acl, row_n FROM " |
| " unnest(acldefault('d',datdba)) " |
| " WITH ORDINALITY AS initp(acl,row_n) " |
| " WHERE NOT EXISTS ( " |
| " SELECT 1 " |
| " FROM unnest(coalesce(datacl,acldefault('d',datdba))) " |
| " AS permp(orig_acl) " |
| " WHERE acl = orig_acl)) AS rdatacls) " |
| " AS rdatacl, " |
| "datistemplate, datconnlimit, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = dattablespace) AS tablespace, " |
| "shobj_description(oid, 'pg_database') AS description " |
| |
| "FROM pg_database " |
| "WHERE datname = current_database()", |
| username_subquery); |
| } |
| else if (fout->remoteVersion >= 90300) |
| { |
| appendPQExpBuffer(dbQry, "SELECT tableoid, oid, datname, " |
| "(%s datdba) AS dba, " |
| "pg_encoding_to_char(encoding) AS encoding, " |
| "datcollate, datctype, datfrozenxid, datminmxid, " |
| "datacl, '' as rdatacl, datistemplate, datconnlimit, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = dattablespace) AS tablespace, " |
| "shobj_description(oid, 'pg_database') AS description " |
| |
| "FROM pg_database " |
| "WHERE datname = current_database()", |
| username_subquery); |
| } |
| else if (fout->remoteVersion >= 80400) |
| { |
| appendPQExpBuffer(dbQry, "SELECT tableoid, oid, datname, " |
| "(%s datdba) AS dba, " |
| "pg_encoding_to_char(encoding) AS encoding, " |
| "datcollate, datctype, datfrozenxid, 0 AS datminmxid, " |
| "datacl, '' as rdatacl, datistemplate, datconnlimit, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = dattablespace) AS tablespace, " |
| "shobj_description(oid, 'pg_database') AS description " |
| |
| "FROM pg_database " |
| "WHERE datname = current_database()", |
| username_subquery); |
| } |
| else |
| { |
| appendPQExpBuffer(dbQry, "SELECT tableoid, oid, datname, " |
| "(%s datdba) AS dba, " |
| "pg_encoding_to_char(encoding) AS encoding, " |
| "NULL AS datcollate, NULL AS datctype, datfrozenxid, 0 AS datminmxid, " |
| "datacl, '' as rdatacl, datistemplate, datconnlimit, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = dattablespace) AS tablespace, " |
| "shobj_description(oid, 'pg_database') AS description " |
| |
| "FROM pg_database " |
| "WHERE datname = current_database()", |
| username_subquery); |
| } |
| ======= |
| appendPQExpBufferStr(dbQry, "SELECT tableoid, oid, datname, " |
| "datdba, " |
| "pg_encoding_to_char(encoding) AS encoding, " |
| "datcollate, datctype, datfrozenxid, " |
| "datacl, acldefault('d', datdba) AS acldefault, " |
| "datistemplate, datconnlimit, "); |
| if (fout->remoteVersion >= 90300) |
| appendPQExpBufferStr(dbQry, "datminmxid, "); |
| else |
| appendPQExpBufferStr(dbQry, "0 AS datminmxid, "); |
| if (fout->remoteVersion >= 150000) |
| appendPQExpBufferStr(dbQry, "datlocprovider, daticulocale, datcollversion, "); |
| else |
| appendPQExpBufferStr(dbQry, "'c' AS datlocprovider, NULL AS daticulocale, NULL AS datcollversion, "); |
| if (fout->remoteVersion >= 160000) |
| appendPQExpBufferStr(dbQry, "daticurules, "); |
| else |
| appendPQExpBufferStr(dbQry, "NULL AS daticurules, "); |
| appendPQExpBufferStr(dbQry, |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = dattablespace) AS tablespace, " |
| "shobj_description(oid, 'pg_database') AS description " |
| "FROM pg_database " |
| "WHERE datname = current_database()"); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQueryForSingleRow(fout, dbQry->data); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_datname = PQfnumber(res, "datname"); |
| i_datdba = PQfnumber(res, "datdba"); |
| i_encoding = PQfnumber(res, "encoding"); |
| i_datlocprovider = PQfnumber(res, "datlocprovider"); |
| i_collate = PQfnumber(res, "datcollate"); |
| i_ctype = PQfnumber(res, "datctype"); |
| i_daticulocale = PQfnumber(res, "daticulocale"); |
| i_daticurules = PQfnumber(res, "daticurules"); |
| i_frozenxid = PQfnumber(res, "datfrozenxid"); |
| i_minmxid = PQfnumber(res, "datminmxid"); |
| i_datacl = PQfnumber(res, "datacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| i_datistemplate = PQfnumber(res, "datistemplate"); |
| i_datconnlimit = PQfnumber(res, "datconnlimit"); |
| i_datcollversion = PQfnumber(res, "datcollversion"); |
| i_tablespace = PQfnumber(res, "tablespace"); |
| |
| dbCatId.tableoid = atooid(PQgetvalue(res, 0, i_tableoid)); |
| dbCatId.oid = atooid(PQgetvalue(res, 0, i_oid)); |
| datname = PQgetvalue(res, 0, i_datname); |
| dba = getRoleName(PQgetvalue(res, 0, i_datdba)); |
| encoding = PQgetvalue(res, 0, i_encoding); |
| datlocprovider = PQgetvalue(res, 0, i_datlocprovider); |
| collate = PQgetvalue(res, 0, i_collate); |
| ctype = PQgetvalue(res, 0, i_ctype); |
| if (!PQgetisnull(res, 0, i_daticulocale)) |
| iculocale = PQgetvalue(res, 0, i_daticulocale); |
| else |
| iculocale = NULL; |
| if (!PQgetisnull(res, 0, i_daticurules)) |
| icurules = PQgetvalue(res, 0, i_daticurules); |
| else |
| icurules = NULL; |
| frozenxid = atooid(PQgetvalue(res, 0, i_frozenxid)); |
| minmxid = atooid(PQgetvalue(res, 0, i_minmxid)); |
| dbdacl.acl = PQgetvalue(res, 0, i_datacl); |
| dbdacl.acldefault = PQgetvalue(res, 0, i_acldefault); |
| datistemplate = PQgetvalue(res, 0, i_datistemplate); |
| datconnlimit = PQgetvalue(res, 0, i_datconnlimit); |
| tablespace = PQgetvalue(res, 0, i_tablespace); |
| |
| qdatname = pg_strdup(fmtId(datname)); |
| |
| /* |
| * Prepare the CREATE DATABASE command. We must specify OID (if we want |
| * to preserve that), as well as the encoding, locale, and tablespace |
| * since those can't be altered later. Other DB properties are left to |
| * the DATABASE PROPERTIES entry, so that they can be applied after |
| * reconnecting to the target DB. |
| */ |
| if (dopt->binary_upgrade) |
| { |
| appendPQExpBuffer(creaQry, "CREATE DATABASE %s WITH TEMPLATE = template0 OID = %u", |
| qdatname, dbCatId.oid); |
| } |
| else |
| { |
| appendPQExpBuffer(creaQry, "CREATE DATABASE %s WITH TEMPLATE = template0", |
| qdatname); |
| } |
| if (strlen(encoding) > 0) |
| { |
| appendPQExpBufferStr(creaQry, " ENCODING = "); |
| appendStringLiteralAH(creaQry, encoding, fout); |
| } |
| |
| appendPQExpBufferStr(creaQry, " LOCALE_PROVIDER = "); |
| if (datlocprovider[0] == 'c') |
| appendPQExpBufferStr(creaQry, "libc"); |
| else if (datlocprovider[0] == 'i') |
| appendPQExpBufferStr(creaQry, "icu"); |
| else |
| pg_fatal("unrecognized locale provider: %s", |
| datlocprovider); |
| |
| if (strlen(collate) > 0 && strcmp(collate, ctype) == 0) |
| { |
| appendPQExpBufferStr(creaQry, " LOCALE = "); |
| appendStringLiteralAH(creaQry, collate, fout); |
| } |
| else |
| { |
| if (strlen(collate) > 0) |
| { |
| appendPQExpBufferStr(creaQry, " LC_COLLATE = "); |
| appendStringLiteralAH(creaQry, collate, fout); |
| } |
| if (strlen(ctype) > 0) |
| { |
| appendPQExpBufferStr(creaQry, " LC_CTYPE = "); |
| appendStringLiteralAH(creaQry, ctype, fout); |
| } |
| } |
| if (iculocale) |
| { |
| appendPQExpBufferStr(creaQry, " ICU_LOCALE = "); |
| appendStringLiteralAH(creaQry, iculocale, fout); |
| } |
| if (icurules) |
| { |
| appendPQExpBufferStr(creaQry, " ICU_RULES = "); |
| appendStringLiteralAH(creaQry, icurules, fout); |
| } |
| |
| /* |
| * For binary upgrade, carry over the collation version. For normal |
| * dump/restore, omit the version, so that it is computed upon restore. |
| */ |
| if (dopt->binary_upgrade) |
| { |
| if (!PQgetisnull(res, 0, i_datcollversion)) |
| { |
| appendPQExpBufferStr(creaQry, " COLLATION_VERSION = "); |
| appendStringLiteralAH(creaQry, |
| PQgetvalue(res, 0, i_datcollversion), |
| fout); |
| } |
| } |
| |
| /* |
| * Note: looking at dopt->outputNoTablespaces here is completely the wrong |
| * thing; the decision whether to specify a tablespace should be left till |
| * pg_restore, so that pg_restore --no-tablespaces applies. Ideally we'd |
| * label the DATABASE entry with the tablespace and let the normal |
| * tablespace selection logic work ... but CREATE DATABASE doesn't pay |
| * attention to default_tablespace, so that won't work. |
| */ |
| if (strlen(tablespace) > 0 && strcmp(tablespace, "pg_default") != 0 && |
| !dopt->outputNoTablespaces) |
| appendPQExpBuffer(creaQry, " TABLESPACE = %s", |
| fmtId(tablespace)); |
| appendPQExpBufferStr(creaQry, ";\n"); |
| |
| appendPQExpBuffer(delQry, "DROP DATABASE %s;\n", |
| qdatname); |
| |
| dbDumpId = createDumpId(); |
| |
| ArchiveEntry(fout, |
| dbCatId, /* catalog ID */ |
| dbDumpId, /* dump ID */ |
| ARCHIVE_OPTS(.tag = datname, |
| .owner = dba, |
| .description = "DATABASE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = creaQry->data, |
| .dropStmt = delQry->data)); |
| |
| /* Compute correct tag for archive entry */ |
| appendPQExpBuffer(labelq, "DATABASE %s", qdatname); |
| |
| <<<<<<< HEAD |
| char *comment = PQgetvalue(res, 0, PQfnumber(res, "description")); |
| |
| if (comment && *comment && !dopt->no_comments) |
| ======= |
| /* Dump DB comment if any */ |
| >>>>>>> REL_16_9 |
| { |
| resetPQExpBuffer(dbQry); |
| |
| /* |
| * Generates warning when loaded into a differently-named |
| * database. |
| */ |
| appendPQExpBuffer(dbQry, "COMMENT ON DATABASE %s IS ", qdatname); |
| appendStringLiteralAH(dbQry, comment, fout); |
| appendPQExpBufferStr(dbQry, ";\n"); |
| |
| <<<<<<< HEAD |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = labelq->data, |
| .owner = dba, |
| .description = "COMMENT", |
| .section = SECTION_NONE, |
| .createStmt = dbQry->data, |
| .deps = &dbDumpId, |
| .nDeps = 1)); |
| ======= |
| if (comment && *comment && !dopt->no_comments) |
| { |
| resetPQExpBuffer(dbQry); |
| |
| /* |
| * Generates warning when loaded into a differently-named |
| * database. |
| */ |
| appendPQExpBuffer(dbQry, "COMMENT ON DATABASE %s IS ", qdatname); |
| appendStringLiteralAH(dbQry, comment, fout); |
| appendPQExpBufferStr(dbQry, ";\n"); |
| |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = labelq->data, |
| .owner = dba, |
| .description = "COMMENT", |
| .section = SECTION_NONE, |
| .createStmt = dbQry->data, |
| .deps = &dbDumpId, |
| .nDeps = 1)); |
| } |
| >>>>>>> REL_16_9 |
| } |
| |
| /* Dump DB security label, if enabled */ |
| if (!dopt->no_security_labels) |
| { |
| PGresult *shres; |
| PQExpBuffer seclabelQry; |
| |
| seclabelQry = createPQExpBuffer(); |
| |
| buildShSecLabelQuery("pg_database", dbCatId.oid, seclabelQry); |
| shres = ExecuteSqlQuery(fout, seclabelQry->data, PGRES_TUPLES_OK); |
| resetPQExpBuffer(seclabelQry); |
| emitShSecLabels(conn, shres, seclabelQry, "DATABASE", datname); |
| if (seclabelQry->len > 0) |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = labelq->data, |
| .owner = dba, |
| .description = "SECURITY LABEL", |
| .section = SECTION_NONE, |
| .createStmt = seclabelQry->data, |
| .deps = &dbDumpId, |
| .nDeps = 1)); |
| destroyPQExpBuffer(seclabelQry); |
| PQclear(shres); |
| } |
| |
| /* |
| * Dump ACL if any. Note that we do not support initial privileges |
| * (pg_init_privs) on databases. |
| */ |
| dbdacl.privtype = 0; |
| dbdacl.initprivs = NULL; |
| |
| dumpACL(fout, dbDumpId, InvalidDumpId, "DATABASE", |
| qdatname, NULL, NULL, |
| dba, &dbdacl); |
| |
| /* |
| * Now construct a DATABASE PROPERTIES archive entry to restore any |
| * non-default database-level properties. (The reason this must be |
| * separate is that we cannot put any additional commands into the TOC |
| * entry that has CREATE DATABASE. pg_restore would execute such a group |
| * in an implicit transaction block, and the backend won't allow CREATE |
| * DATABASE in that context.) |
| */ |
| resetPQExpBuffer(creaQry); |
| resetPQExpBuffer(delQry); |
| |
| if (strlen(datconnlimit) > 0 && strcmp(datconnlimit, "-1") != 0) |
| appendPQExpBuffer(creaQry, "ALTER DATABASE %s CONNECTION LIMIT = %s;\n", |
| qdatname, datconnlimit); |
| |
| if (strcmp(datistemplate, "t") == 0) |
| { |
| appendPQExpBuffer(creaQry, "ALTER DATABASE %s IS_TEMPLATE = true;\n", |
| qdatname); |
| |
| appendPQExpBuffer(delQry, "SET allow_system_table_mods = true;\n"); |
| |
| /* |
| * The backend won't accept DROP DATABASE on a template database. We |
| * can deal with that by removing the template marking before the DROP |
| * gets issued. We'd prefer to use ALTER DATABASE IF EXISTS here, but |
| * since no such command is currently supported, fake it with a direct |
| * UPDATE on pg_database. |
| */ |
| appendPQExpBufferStr(delQry, "UPDATE pg_catalog.pg_database " |
| "SET datistemplate = false WHERE datname = "); |
| appendStringLiteralAH(delQry, datname, fout); |
| appendPQExpBufferStr(delQry, ";\n"); |
| appendPQExpBuffer(delQry, "RESET allow_system_table_mods;\n"); |
| } |
| |
| /* Add database-specific SET options */ |
| dumpDatabaseConfig(fout, creaQry, datname, dbCatId.oid); |
| |
| /* |
| * We stick this binary-upgrade query into the DATABASE PROPERTIES archive |
| * entry, too, for lack of a better place. |
| */ |
| if (dopt->binary_upgrade) |
| { |
| appendPQExpBufferStr(creaQry, "\n-- For binary upgrade, set datfrozenxid and datminmxid.\n"); |
| appendPQExpBufferStr(creaQry, "SET allow_system_table_mods = true;\n"); |
| appendPQExpBuffer(creaQry, "UPDATE pg_catalog.pg_database\n" |
| "SET datfrozenxid = '%u', datminmxid = '%u'\n" |
| "WHERE datname = ", |
| frozenxid, minmxid); |
| appendStringLiteralAH(creaQry, datname, fout); |
| appendPQExpBufferStr(creaQry, ";\n"); |
| |
| appendPQExpBuffer(creaQry, "RESET allow_system_table_mods;\n"); |
| } |
| |
| if (creaQry->len > 0) |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = datname, |
| .owner = dba, |
| .description = "DATABASE PROPERTIES", |
| .section = SECTION_PRE_DATA, |
| .createStmt = creaQry->data, |
| .dropStmt = delQry->data, |
| .deps = &dbDumpId)); |
| |
| /* |
| * pg_largeobject comes from the old system intact, so set its |
| * relfrozenxids, relminmxids and relfilenode. |
| */ |
| if (dopt->binary_upgrade) |
| { |
| PGresult *lo_res; |
| PQExpBuffer loFrozenQry = createPQExpBuffer(); |
| PQExpBuffer loOutQry = createPQExpBuffer(); |
| PQExpBuffer loHorizonQry = createPQExpBuffer(); |
| int ii_relfrozenxid, |
| ii_relfilenode, |
| ii_oid, |
| ii_relminmxid; |
| |
| /* |
| * pg_largeobject |
| */ |
| if (fout->remoteVersion >= 90300) |
| appendPQExpBuffer(loFrozenQry, "SELECT relfrozenxid, relminmxid, relfilenode, oid\n" |
| "FROM pg_catalog.pg_class\n" |
| "WHERE oid IN (%u, %u);\n", |
| LargeObjectRelationId, LargeObjectLOidPNIndexId); |
| else |
| appendPQExpBuffer(loFrozenQry, "SELECT relfrozenxid, 0 AS relminmxid, relfilenode, oid\n" |
| "FROM pg_catalog.pg_class\n" |
| "WHERE oid IN (%u, %u);\n", |
| LargeObjectRelationId, LargeObjectLOidPNIndexId); |
| |
| lo_res = ExecuteSqlQuery(fout, loFrozenQry->data, PGRES_TUPLES_OK); |
| |
| ii_relfrozenxid = PQfnumber(lo_res, "relfrozenxid"); |
| ii_relminmxid = PQfnumber(lo_res, "relminmxid"); |
| ii_relfilenode = PQfnumber(lo_res, "relfilenode"); |
| ii_oid = PQfnumber(lo_res, "oid"); |
| |
| appendPQExpBufferStr(loHorizonQry, "\n-- For binary upgrade, set pg_largeobject relfrozenxid and relminmxid\n"); |
| appendPQExpBufferStr(loOutQry, "\n-- For binary upgrade, preserve pg_largeobject and index relfilenodes\n"); |
| for (int i = 0; i < PQntuples(lo_res); ++i) |
| { |
| Oid oid; |
| RelFileNumber relfilenumber; |
| |
| appendPQExpBuffer(loHorizonQry, "UPDATE pg_catalog.pg_class\n" |
| "SET relfrozenxid = '%u', relminmxid = '%u'\n" |
| "WHERE oid = %u;\n", |
| atooid(PQgetvalue(lo_res, i, ii_relfrozenxid)), |
| atooid(PQgetvalue(lo_res, i, ii_relminmxid)), |
| atooid(PQgetvalue(lo_res, i, ii_oid))); |
| |
| oid = atooid(PQgetvalue(lo_res, i, ii_oid)); |
| relfilenumber = atooid(PQgetvalue(lo_res, i, ii_relfilenode)); |
| |
| if (oid == LargeObjectRelationId) |
| appendPQExpBuffer(loOutQry, |
| "SELECT pg_catalog.binary_upgrade_set_next_heap_relfilenode('%u'::pg_catalog.oid);\n", |
| relfilenumber); |
| else if (oid == LargeObjectLOidPNIndexId) |
| appendPQExpBuffer(loOutQry, |
| "SELECT pg_catalog.binary_upgrade_set_next_index_relfilenode('%u'::pg_catalog.oid);\n", |
| relfilenumber); |
| } |
| |
| appendPQExpBufferStr(loOutQry, |
| "TRUNCATE pg_catalog.pg_largeobject;\n"); |
| appendPQExpBufferStr(loOutQry, loHorizonQry->data); |
| |
| <<<<<<< HEAD |
| appendPQExpBufferStr(loOutQry, "\n-- For binary upgrade, set pg_largeobject relfrozenxid and relminmxid\n"); |
| appendPQExpBufferStr(loOutQry, "SET allow_system_table_mods = true;\n"); |
| appendPQExpBuffer(loOutQry, "UPDATE pg_catalog.pg_class\n" |
| "SET relfrozenxid = '%u', relminmxid = '%u'\n" |
| "WHERE oid = %u;\n", |
| atooid(PQgetvalue(lo_res, 0, i_relfrozenxid)), |
| atooid(PQgetvalue(lo_res, 0, i_relminmxid)), |
| LargeObjectRelationId); |
| appendPQExpBuffer(loOutQry, "RESET allow_system_table_mods;\n"); |
| ======= |
| >>>>>>> REL_16_9 |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = "pg_largeobject", |
| .description = "pg_largeobject", |
| .section = SECTION_PRE_DATA, |
| .createStmt = loOutQry->data)); |
| |
| PQclear(lo_res); |
| |
| destroyPQExpBuffer(loFrozenQry); |
| destroyPQExpBuffer(loHorizonQry); |
| destroyPQExpBuffer(loOutQry); |
| } |
| |
| PQclear(res); |
| |
| free(qdatname); |
| destroyPQExpBuffer(dbQry); |
| destroyPQExpBuffer(delQry); |
| destroyPQExpBuffer(creaQry); |
| destroyPQExpBuffer(labelq); |
| } |
| |
| /* |
| * Collect any database-specific or role-and-database-specific SET options |
| * for this database, and append them to outbuf. |
| */ |
| static void |
| dumpDatabaseConfig(Archive *AH, PQExpBuffer outbuf, |
| const char *dbname, Oid dboid) |
| { |
| PGconn *conn = GetConnection(AH); |
| PQExpBuffer buf = createPQExpBuffer(); |
| PGresult *res; |
| |
| /* First collect database-specific options */ |
| printfPQExpBuffer(buf, "SELECT unnest(setconfig) FROM pg_db_role_setting " |
| "WHERE setrole = 0 AND setdatabase = '%u'::oid", |
| dboid); |
| |
| res = ExecuteSqlQuery(AH, buf->data, PGRES_TUPLES_OK); |
| |
| for (int i = 0; i < PQntuples(res); i++) |
| makeAlterConfigCommand(conn, PQgetvalue(res, i, 0), |
| "DATABASE", dbname, NULL, NULL, |
| outbuf); |
| |
| PQclear(res); |
| |
| /* Now look for role-and-database-specific options */ |
| printfPQExpBuffer(buf, "SELECT rolname, unnest(setconfig) " |
| "FROM pg_db_role_setting s, pg_roles r " |
| "WHERE setrole = r.oid AND setdatabase = '%u'::oid", |
| dboid); |
| |
| res = ExecuteSqlQuery(AH, buf->data, PGRES_TUPLES_OK); |
| |
| for (int i = 0; i < PQntuples(res); i++) |
| makeAlterConfigCommand(conn, PQgetvalue(res, i, 1), |
| "ROLE", PQgetvalue(res, i, 0), |
| "DATABASE", dbname, |
| outbuf); |
| |
| PQclear(res); |
| |
| /* |
| * If we're upgrading from GPDB 5 or below, use the legacy hash ops. |
| */ |
| if (AH->dopt->binary_upgrade && AH->remoteVersion < 90400) |
| { |
| makeAlterConfigCommand(conn, "gp_use_legacy_hashops=on", |
| "DATABASE", dbname, NULL, NULL, outbuf); |
| } |
| |
| destroyPQExpBuffer(buf); |
| } |
| |
| /* |
| * dumpEncoding: put the correct encoding into the archive |
| */ |
| static void |
| dumpEncoding(Archive *AH) |
| { |
| const char *encname = pg_encoding_to_char(AH->encoding); |
| PQExpBuffer qry = createPQExpBuffer(); |
| |
| pg_log_info("saving encoding = %s", encname); |
| |
| appendPQExpBufferStr(qry, "SET client_encoding = "); |
| appendStringLiteralAH(qry, encname, AH); |
| appendPQExpBufferStr(qry, ";\n"); |
| |
| ArchiveEntry(AH, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = "ENCODING", |
| .description = "ENCODING", |
| .section = SECTION_PRE_DATA, |
| .createStmt = qry->data)); |
| |
| destroyPQExpBuffer(qry); |
| } |
| |
| |
| /* |
| * dumpStdStrings: put the correct escape string behavior into the archive |
| */ |
| static void |
| dumpStdStrings(Archive *AH) |
| { |
| const char *stdstrings = AH->std_strings ? "on" : "off"; |
| PQExpBuffer qry = createPQExpBuffer(); |
| |
| pg_log_info("saving standard_conforming_strings = %s", |
| stdstrings); |
| |
| appendPQExpBuffer(qry, "SET standard_conforming_strings = '%s';\n", |
| stdstrings); |
| |
| ArchiveEntry(AH, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = "STDSTRINGS", |
| .description = "STDSTRINGS", |
| .section = SECTION_PRE_DATA, |
| .createStmt = qry->data)); |
| |
| destroyPQExpBuffer(qry); |
| } |
| |
| /* |
| * dumpSearchPath: record the active search_path in the archive |
| */ |
| static void |
| dumpSearchPath(Archive *AH) |
| { |
| PQExpBuffer qry = createPQExpBuffer(); |
| PQExpBuffer path = createPQExpBuffer(); |
| PGresult *res; |
| char **schemanames = NULL; |
| int nschemanames = 0; |
| int i; |
| |
| /* |
| * We use the result of current_schemas(), not the search_path GUC, |
| * because that might contain wildcards such as "$user", which won't |
| * necessarily have the same value during restore. Also, this way avoids |
| * listing schemas that may appear in search_path but not actually exist, |
| * which seems like a prudent exclusion. |
| */ |
| res = ExecuteSqlQueryForSingleRow(AH, |
| "SELECT pg_catalog.current_schemas(false)"); |
| |
| if (!parsePGArray(PQgetvalue(res, 0, 0), &schemanames, &nschemanames)) |
| pg_fatal("could not parse result of current_schemas()"); |
| |
| /* |
| * We use set_config(), not a simple "SET search_path" command, because |
| * the latter has less-clean behavior if the search path is empty. While |
| * that's likely to get fixed at some point, it seems like a good idea to |
| * be as backwards-compatible as possible in what we put into archives. |
| */ |
| for (i = 0; i < nschemanames; i++) |
| { |
| if (i > 0) |
| appendPQExpBufferStr(path, ", "); |
| appendPQExpBufferStr(path, fmtId(schemanames[i])); |
| } |
| |
| appendPQExpBufferStr(qry, "SELECT pg_catalog.set_config('search_path', "); |
| appendStringLiteralAH(qry, path->data, AH); |
| appendPQExpBufferStr(qry, ", false);\n"); |
| |
| pg_log_info("saving search_path = %s", path->data); |
| |
| ArchiveEntry(AH, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = "SEARCHPATH", |
| .description = "SEARCHPATH", |
| .section = SECTION_PRE_DATA, |
| .createStmt = qry->data)); |
| |
| /* Also save it in AH->searchpath, in case we're doing plain text dump */ |
| AH->searchpath = pg_strdup(qry->data); |
| |
| free(schemanames); |
| PQclear(res); |
| destroyPQExpBuffer(qry); |
| destroyPQExpBuffer(path); |
| } |
| |
| |
| /* |
| * getLOs: |
| * Collect schema-level data about large objects |
| */ |
| static void |
| getLOs(Archive *fout) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer loQry = createPQExpBuffer(); |
| LoInfo *loinfo; |
| DumpableObject *lodata; |
| PGresult *res; |
| int ntups; |
| int i; |
| int i_oid; |
| int i_lomowner; |
| int i_lomacl; |
| int i_acldefault; |
| |
| pg_log_info("reading large objects"); |
| |
| /* Fetch LO OIDs, and owner/ACL data */ |
| appendPQExpBufferStr(loQry, |
| "SELECT oid, lomowner, lomacl, " |
| "acldefault('L', lomowner) AS acldefault " |
| "FROM pg_largeobject_metadata"); |
| |
| res = ExecuteSqlQuery(fout, loQry->data, PGRES_TUPLES_OK); |
| |
| i_oid = PQfnumber(res, "oid"); |
| i_lomowner = PQfnumber(res, "lomowner"); |
| i_lomacl = PQfnumber(res, "lomacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| |
| ntups = PQntuples(res); |
| |
| /* |
| * Each large object has its own "BLOB" archive entry. |
| */ |
| loinfo = (LoInfo *) pg_malloc(ntups * sizeof(LoInfo)); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| loinfo[i].dobj.objType = DO_LARGE_OBJECT; |
| loinfo[i].dobj.catId.tableoid = LargeObjectRelationId; |
| loinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&loinfo[i].dobj); |
| |
| loinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_oid)); |
| loinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_lomacl)); |
| loinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| loinfo[i].dacl.privtype = 0; |
| loinfo[i].dacl.initprivs = NULL; |
| loinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_lomowner)); |
| |
| /* LOs have data */ |
| loinfo[i].dobj.components |= DUMP_COMPONENT_DATA; |
| |
| /* Mark whether LO has an ACL */ |
| if (!PQgetisnull(res, i, i_lomacl)) |
| loinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| |
| /* |
| * In binary-upgrade mode for LOs, we do *not* dump out the LO data, |
| * as it will be copied by pg_upgrade, which simply copies the |
| * pg_largeobject table. We *do* however dump out anything but the |
| * data, as pg_upgrade copies just pg_largeobject, but not |
| * pg_largeobject_metadata, after the dump is restored. |
| */ |
| if (dopt->binary_upgrade) |
| loinfo[i].dobj.dump &= ~DUMP_COMPONENT_DATA; |
| } |
| |
| /* |
| * If we have any large objects, a "BLOBS" archive entry is needed. This |
| * is just a placeholder for sorting; it carries no data now. |
| */ |
| if (ntups > 0) |
| { |
| lodata = (DumpableObject *) pg_malloc(sizeof(DumpableObject)); |
| lodata->objType = DO_LARGE_OBJECT_DATA; |
| lodata->catId = nilCatalogId; |
| AssignDumpId(lodata); |
| lodata->name = pg_strdup("BLOBS"); |
| lodata->components |= DUMP_COMPONENT_DATA; |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(loQry); |
| } |
| |
| /* |
| * dumpLO |
| * |
| * dump the definition (metadata) of the given large object |
| */ |
| static void |
| dumpLO(Archive *fout, const LoInfo *loinfo) |
| { |
| PQExpBuffer cquery = createPQExpBuffer(); |
| PQExpBuffer dquery = createPQExpBuffer(); |
| |
| appendPQExpBuffer(cquery, |
| "SELECT pg_catalog.lo_create('%s');\n", |
| loinfo->dobj.name); |
| |
| appendPQExpBuffer(dquery, |
| "SELECT pg_catalog.lo_unlink('%s');\n", |
| loinfo->dobj.name); |
| |
| if (loinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, loinfo->dobj.catId, loinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = loinfo->dobj.name, |
| .owner = loinfo->rolname, |
| .description = "BLOB", |
| .section = SECTION_PRE_DATA, |
| .createStmt = cquery->data, |
| .dropStmt = dquery->data)); |
| |
| /* Dump comment if any */ |
| if (loinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "LARGE OBJECT", loinfo->dobj.name, |
| NULL, loinfo->rolname, |
| loinfo->dobj.catId, 0, loinfo->dobj.dumpId); |
| |
| /* Dump security label if any */ |
| if (loinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "LARGE OBJECT", loinfo->dobj.name, |
| NULL, loinfo->rolname, |
| loinfo->dobj.catId, 0, loinfo->dobj.dumpId); |
| |
| /* Dump ACL if any */ |
| if (loinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, loinfo->dobj.dumpId, InvalidDumpId, "LARGE OBJECT", |
| loinfo->dobj.name, NULL, |
| NULL, loinfo->rolname, &loinfo->dacl); |
| |
| destroyPQExpBuffer(cquery); |
| destroyPQExpBuffer(dquery); |
| } |
| |
| /* |
| * dumpLOs: |
| * dump the data contents of all large objects |
| */ |
| static int |
| dumpLOs(Archive *fout, const void *arg) |
| { |
| const char *loQry; |
| const char *loFetchQry; |
| PGconn *conn = GetConnection(fout); |
| PGresult *res; |
| char buf[LOBBUFSIZE]; |
| int ntups; |
| int i; |
| int cnt; |
| |
| pg_log_info("saving large objects"); |
| |
| /* |
| * Currently, we re-fetch all LO OIDs using a cursor. Consider scanning |
| * the already-in-memory dumpable objects instead... |
| */ |
| loQry = |
| "DECLARE looid CURSOR FOR " |
| "SELECT oid FROM pg_largeobject_metadata ORDER BY 1"; |
| |
| ExecuteSqlStatement(fout, loQry); |
| |
| /* Command to fetch from cursor */ |
| loFetchQry = "FETCH 1000 IN looid"; |
| |
| do |
| { |
| /* Do a fetch */ |
| res = ExecuteSqlQuery(fout, loFetchQry, PGRES_TUPLES_OK); |
| |
| /* Process the tuples, if any */ |
| ntups = PQntuples(res); |
| for (i = 0; i < ntups; i++) |
| { |
| Oid loOid; |
| int loFd; |
| |
| loOid = atooid(PQgetvalue(res, i, 0)); |
| /* Open the LO */ |
| loFd = lo_open(conn, loOid, INV_READ); |
| if (loFd == -1) |
| pg_fatal("could not open large object %u: %s", |
| loOid, PQerrorMessage(conn)); |
| |
| StartLO(fout, loOid); |
| |
| /* Now read it in chunks, sending data to archive */ |
| do |
| { |
| cnt = lo_read(conn, loFd, buf, LOBBUFSIZE); |
| if (cnt < 0) |
| pg_fatal("error reading large object %u: %s", |
| loOid, PQerrorMessage(conn)); |
| |
| WriteData(fout, buf, cnt); |
| } while (cnt > 0); |
| |
| lo_close(conn, loFd); |
| |
| EndLO(fout, loOid); |
| } |
| |
| PQclear(res); |
| } while (ntups > 0); |
| |
| return 1; |
| } |
| |
| /* |
| * getPolicies |
| * get information about all RLS policies on dumpable tables. |
| */ |
| void |
| getPolicies(Archive *fout, TableInfo tblinfo[], int numTables) |
| { |
| PQExpBuffer query; |
| PQExpBuffer tbloids; |
| PGresult *res; |
| PolicyInfo *polinfo; |
| int i_oid; |
| int i_tableoid; |
| int i_polrelid; |
| int i_polname; |
| int i_polcmd; |
| int i_polpermissive; |
| int i_polroles; |
| int i_polqual; |
| int i_polwithcheck; |
| int i, |
| j, |
| ntups; |
| |
| /* No policies before 9.5 */ |
| if (fout->remoteVersion < 90500) |
| return; |
| |
| query = createPQExpBuffer(); |
| tbloids = createPQExpBuffer(); |
| |
| /* |
| <<<<<<< HEAD |
| * First, check which tables have RLS enabled. We represent RLS being |
| * enabled on a table by creating a PolicyInfo object with null polname. |
| */ |
| ======= |
| * Identify tables of interest, and check which ones have RLS enabled. |
| */ |
| appendPQExpBufferChar(tbloids, '{'); |
| >>>>>>> REL_16_9 |
| for (i = 0; i < numTables; i++) |
| { |
| TableInfo *tbinfo = &tblinfo[i]; |
| |
| /* Ignore row security on tables not to be dumped */ |
| if (!(tbinfo->dobj.dump & DUMP_COMPONENT_POLICY)) |
| continue; |
| |
| <<<<<<< HEAD |
| ======= |
| /* It can't have RLS or policies if it's not a table */ |
| if (tbinfo->relkind != RELKIND_RELATION && |
| tbinfo->relkind != RELKIND_PARTITIONED_TABLE) |
| continue; |
| |
| /* Add it to the list of table OIDs to be probed below */ |
| if (tbloids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(tbloids, ','); |
| appendPQExpBuffer(tbloids, "%u", tbinfo->dobj.catId.oid); |
| |
| /* Is RLS enabled? (That's separate from whether it has policies) */ |
| >>>>>>> REL_16_9 |
| if (tbinfo->rowsec) |
| { |
| tbinfo->dobj.components |= DUMP_COMPONENT_POLICY; |
| |
| /* |
| * We represent RLS being enabled on a table by creating a |
| * PolicyInfo object with null polname. |
| * |
| * Note: use tableoid 0 so that this object won't be mistaken for |
| * something that pg_depend entries apply to. |
| */ |
| polinfo = pg_malloc(sizeof(PolicyInfo)); |
| polinfo->dobj.objType = DO_POLICY; |
| polinfo->dobj.catId.tableoid = 0; |
| polinfo->dobj.catId.oid = tbinfo->dobj.catId.oid; |
| AssignDumpId(&polinfo->dobj); |
| polinfo->dobj.namespace = tbinfo->dobj.namespace; |
| polinfo->dobj.name = pg_strdup(tbinfo->dobj.name); |
| polinfo->poltable = tbinfo; |
| polinfo->polname = NULL; |
| polinfo->polcmd = '\0'; |
| polinfo->polpermissive = 0; |
| polinfo->polroles = NULL; |
| polinfo->polqual = NULL; |
| polinfo->polwithcheck = NULL; |
| } |
| } |
| <<<<<<< HEAD |
| |
| /* |
| * Now, read all RLS policies, and create PolicyInfo objects for all those |
| * that are of interest. |
| ======= |
| appendPQExpBufferChar(tbloids, '}'); |
| |
| /* |
| * Now, read all RLS policies belonging to the tables of interest, and |
| * create PolicyInfo objects for them. (Note that we must filter the |
| * results server-side not locally, because we dare not apply pg_get_expr |
| * to tables we don't have lock on.) |
| >>>>>>> REL_16_9 |
| */ |
| pg_log_info("reading row-level security policies"); |
| |
| printfPQExpBuffer(query, |
| <<<<<<< HEAD |
| "SELECT oid, tableoid, pol.polrelid, pol.polname, pol.polcmd, "); |
| if (fout->remoteVersion >= 100000) |
| appendPQExpBuffer(query, "pol.polpermissive, "); |
| else |
| appendPQExpBuffer(query, "'t' as polpermissive, "); |
| ======= |
| "SELECT pol.oid, pol.tableoid, pol.polrelid, pol.polname, pol.polcmd, "); |
| if (fout->remoteVersion >= 100000) |
| appendPQExpBufferStr(query, "pol.polpermissive, "); |
| else |
| appendPQExpBufferStr(query, "'t' as polpermissive, "); |
| >>>>>>> REL_16_9 |
| appendPQExpBuffer(query, |
| "CASE WHEN pol.polroles = '{0}' THEN NULL ELSE " |
| " pg_catalog.array_to_string(ARRAY(SELECT pg_catalog.quote_ident(rolname) from pg_catalog.pg_roles WHERE oid = ANY(pol.polroles)), ', ') END AS polroles, " |
| "pg_catalog.pg_get_expr(pol.polqual, pol.polrelid) AS polqual, " |
| "pg_catalog.pg_get_expr(pol.polwithcheck, pol.polrelid) AS polwithcheck " |
| <<<<<<< HEAD |
| "FROM pg_catalog.pg_policy pol"); |
| ======= |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_policy pol ON (src.tbloid = pol.polrelid)", |
| tbloids->data); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| if (ntups > 0) |
| { |
| i_oid = PQfnumber(res, "oid"); |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_polrelid = PQfnumber(res, "polrelid"); |
| i_polname = PQfnumber(res, "polname"); |
| i_polcmd = PQfnumber(res, "polcmd"); |
| i_polpermissive = PQfnumber(res, "polpermissive"); |
| i_polroles = PQfnumber(res, "polroles"); |
| i_polqual = PQfnumber(res, "polqual"); |
| i_polwithcheck = PQfnumber(res, "polwithcheck"); |
| |
| polinfo = pg_malloc(ntups * sizeof(PolicyInfo)); |
| |
| for (j = 0; j < ntups; j++) |
| { |
| Oid polrelid = atooid(PQgetvalue(res, j, i_polrelid)); |
| TableInfo *tbinfo = findTableByOid(polrelid); |
| |
| <<<<<<< HEAD |
| /* |
| * Ignore row security on tables not to be dumped. (This will |
| * result in some harmless wasted slots in polinfo[].) |
| */ |
| if (!(tbinfo->dobj.dump & DUMP_COMPONENT_POLICY)) |
| continue; |
| ======= |
| tbinfo->dobj.components |= DUMP_COMPONENT_POLICY; |
| >>>>>>> REL_16_9 |
| |
| polinfo[j].dobj.objType = DO_POLICY; |
| polinfo[j].dobj.catId.tableoid = |
| atooid(PQgetvalue(res, j, i_tableoid)); |
| polinfo[j].dobj.catId.oid = atooid(PQgetvalue(res, j, i_oid)); |
| AssignDumpId(&polinfo[j].dobj); |
| polinfo[j].dobj.namespace = tbinfo->dobj.namespace; |
| polinfo[j].poltable = tbinfo; |
| polinfo[j].polname = pg_strdup(PQgetvalue(res, j, i_polname)); |
| polinfo[j].dobj.name = pg_strdup(polinfo[j].polname); |
| |
| polinfo[j].polcmd = *(PQgetvalue(res, j, i_polcmd)); |
| polinfo[j].polpermissive = *(PQgetvalue(res, j, i_polpermissive)) == 't'; |
| |
| if (PQgetisnull(res, j, i_polroles)) |
| polinfo[j].polroles = NULL; |
| else |
| polinfo[j].polroles = pg_strdup(PQgetvalue(res, j, i_polroles)); |
| |
| if (PQgetisnull(res, j, i_polqual)) |
| polinfo[j].polqual = NULL; |
| else |
| polinfo[j].polqual = pg_strdup(PQgetvalue(res, j, i_polqual)); |
| |
| if (PQgetisnull(res, j, i_polwithcheck)) |
| polinfo[j].polwithcheck = NULL; |
| else |
| polinfo[j].polwithcheck |
| = pg_strdup(PQgetvalue(res, j, i_polwithcheck)); |
| } |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(tbloids); |
| } |
| |
| /* |
| * dumpPolicy |
| * dump the definition of the given policy |
| */ |
| static void |
| dumpPolicy(Archive *fout, const PolicyInfo *polinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = polinfo->poltable; |
| PQExpBuffer query; |
| PQExpBuffer delqry; |
| PQExpBuffer polprefix; |
| char *qtabname; |
| const char *cmd; |
| char *tag; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* |
| * If polname is NULL, then this record is just indicating that ROW LEVEL |
| * SECURITY is enabled for the table. Dump as ALTER TABLE <table> ENABLE |
| * ROW LEVEL SECURITY. |
| */ |
| if (polinfo->polname == NULL) |
| { |
| query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "ALTER TABLE %s ENABLE ROW LEVEL SECURITY;", |
| fmtQualifiedDumpable(tbinfo)); |
| |
| /* |
| * We must emit the ROW SECURITY object's dependency on its table |
| * explicitly, because it will not match anything in pg_depend (unlike |
| * the case for other PolicyInfo objects). |
| */ |
| if (polinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, polinfo->dobj.catId, polinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = polinfo->dobj.name, |
| .namespace = polinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "ROW SECURITY", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1)); |
| |
| destroyPQExpBuffer(query); |
| return; |
| } |
| |
| if (polinfo->polcmd == '*') |
| cmd = ""; |
| else if (polinfo->polcmd == 'r') |
| cmd = " FOR SELECT"; |
| else if (polinfo->polcmd == 'a') |
| cmd = " FOR INSERT"; |
| else if (polinfo->polcmd == 'w') |
| cmd = " FOR UPDATE"; |
| else if (polinfo->polcmd == 'd') |
| cmd = " FOR DELETE"; |
| else |
| pg_fatal("unexpected policy command type: %c", |
| polinfo->polcmd); |
| |
| query = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| polprefix = createPQExpBuffer(); |
| |
| qtabname = pg_strdup(fmtId(tbinfo->dobj.name)); |
| |
| appendPQExpBuffer(query, "CREATE POLICY %s", fmtId(polinfo->polname)); |
| |
| appendPQExpBuffer(query, " ON %s%s%s", fmtQualifiedDumpable(tbinfo), |
| !polinfo->polpermissive ? " AS RESTRICTIVE" : "", cmd); |
| |
| if (polinfo->polroles != NULL) |
| appendPQExpBuffer(query, " TO %s", polinfo->polroles); |
| |
| if (polinfo->polqual != NULL) |
| appendPQExpBuffer(query, " USING (%s)", polinfo->polqual); |
| |
| if (polinfo->polwithcheck != NULL) |
| appendPQExpBuffer(query, " WITH CHECK (%s)", polinfo->polwithcheck); |
| |
| appendPQExpBufferStr(query, ";\n"); |
| |
| appendPQExpBuffer(delqry, "DROP POLICY %s", fmtId(polinfo->polname)); |
| appendPQExpBuffer(delqry, " ON %s;\n", fmtQualifiedDumpable(tbinfo)); |
| |
| appendPQExpBuffer(polprefix, "POLICY %s ON", |
| fmtId(polinfo->polname)); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, polinfo->dobj.name); |
| |
| if (polinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, polinfo->dobj.catId, polinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = polinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "POLICY", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data, |
| .dropStmt = delqry->data)); |
| |
| if (polinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, polprefix->data, qtabname, |
| tbinfo->dobj.namespace->dobj.name, tbinfo->rolname, |
| polinfo->dobj.catId, 0, polinfo->dobj.dumpId); |
| |
| free(tag); |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(delqry); |
| destroyPQExpBuffer(polprefix); |
| free(qtabname); |
| } |
| |
| /* |
| * getPublications |
| * get information about publications |
| */ |
| PublicationInfo * |
| getPublications(Archive *fout, int *numPublications) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PGresult *res; |
| PublicationInfo *pubinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_pubname; |
| int i_pubowner; |
| int i_puballtables; |
| int i_pubinsert; |
| int i_pubupdate; |
| int i_pubdelete; |
| int i_pubtruncate; |
| int i_pubviaroot; |
| int i, |
| ntups; |
| |
| if (dopt->no_publications || fout->remoteVersion < 100000) |
| { |
| *numPublications = 0; |
| return NULL; |
| } |
| |
| query = createPQExpBuffer(); |
| |
| resetPQExpBuffer(query); |
| |
| /* Get the publications. */ |
| if (fout->remoteVersion >= 130000) |
| appendPQExpBufferStr(query, |
| "SELECT p.tableoid, p.oid, p.pubname, " |
| "p.pubowner, " |
| "p.puballtables, p.pubinsert, p.pubupdate, p.pubdelete, p.pubtruncate, p.pubviaroot " |
| "FROM pg_publication p"); |
| else if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(query, |
| "SELECT p.tableoid, p.oid, p.pubname, " |
| "p.pubowner, " |
| "p.puballtables, p.pubinsert, p.pubupdate, p.pubdelete, p.pubtruncate, false AS pubviaroot " |
| "FROM pg_publication p"); |
| else |
| appendPQExpBufferStr(query, |
| "SELECT p.tableoid, p.oid, p.pubname, " |
| "p.pubowner, " |
| "p.puballtables, p.pubinsert, p.pubupdate, p.pubdelete, false AS pubtruncate, false AS pubviaroot " |
| "FROM pg_publication p"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_pubname = PQfnumber(res, "pubname"); |
| i_pubowner = PQfnumber(res, "pubowner"); |
| i_puballtables = PQfnumber(res, "puballtables"); |
| i_pubinsert = PQfnumber(res, "pubinsert"); |
| i_pubupdate = PQfnumber(res, "pubupdate"); |
| i_pubdelete = PQfnumber(res, "pubdelete"); |
| i_pubtruncate = PQfnumber(res, "pubtruncate"); |
| i_pubviaroot = PQfnumber(res, "pubviaroot"); |
| |
| pubinfo = pg_malloc(ntups * sizeof(PublicationInfo)); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| pubinfo[i].dobj.objType = DO_PUBLICATION; |
| pubinfo[i].dobj.catId.tableoid = |
| atooid(PQgetvalue(res, i, i_tableoid)); |
| pubinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&pubinfo[i].dobj); |
| pubinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_pubname)); |
| pubinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_pubowner)); |
| pubinfo[i].puballtables = |
| (strcmp(PQgetvalue(res, i, i_puballtables), "t") == 0); |
| pubinfo[i].pubinsert = |
| (strcmp(PQgetvalue(res, i, i_pubinsert), "t") == 0); |
| pubinfo[i].pubupdate = |
| (strcmp(PQgetvalue(res, i, i_pubupdate), "t") == 0); |
| pubinfo[i].pubdelete = |
| (strcmp(PQgetvalue(res, i, i_pubdelete), "t") == 0); |
| pubinfo[i].pubtruncate = |
| (strcmp(PQgetvalue(res, i, i_pubtruncate), "t") == 0); |
| pubinfo[i].pubviaroot = |
| (strcmp(PQgetvalue(res, i, i_pubviaroot), "t") == 0); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(pubinfo[i].dobj), fout); |
| } |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| *numPublications = ntups; |
| return pubinfo; |
| } |
| |
| /* |
| * dumpPublication |
| * dump the definition of the given publication |
| */ |
| static void |
| dumpPublication(Archive *fout, const PublicationInfo *pubinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| char *qpubname; |
| bool first = true; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| qpubname = pg_strdup(fmtId(pubinfo->dobj.name)); |
| |
| appendPQExpBuffer(delq, "DROP PUBLICATION %s;\n", |
| qpubname); |
| |
| appendPQExpBuffer(query, "CREATE PUBLICATION %s", |
| qpubname); |
| |
| if (pubinfo->puballtables) |
| appendPQExpBufferStr(query, " FOR ALL TABLES"); |
| |
| appendPQExpBufferStr(query, " WITH (publish = '"); |
| if (pubinfo->pubinsert) |
| { |
| appendPQExpBufferStr(query, "insert"); |
| first = false; |
| } |
| |
| if (pubinfo->pubupdate) |
| { |
| if (!first) |
| appendPQExpBufferStr(query, ", "); |
| |
| appendPQExpBufferStr(query, "update"); |
| first = false; |
| } |
| |
| if (pubinfo->pubdelete) |
| { |
| if (!first) |
| appendPQExpBufferStr(query, ", "); |
| |
| appendPQExpBufferStr(query, "delete"); |
| first = false; |
| } |
| |
| if (pubinfo->pubtruncate) |
| { |
| if (!first) |
| appendPQExpBufferStr(query, ", "); |
| |
| appendPQExpBufferStr(query, "truncate"); |
| first = false; |
| } |
| |
| appendPQExpBufferChar(query, '\''); |
| |
| if (pubinfo->pubviaroot) |
| appendPQExpBufferStr(query, ", publish_via_partition_root = true"); |
| |
| appendPQExpBufferStr(query, ");\n"); |
| |
| if (pubinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, pubinfo->dobj.catId, pubinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = pubinfo->dobj.name, |
| .owner = pubinfo->rolname, |
| .description = "PUBLICATION", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data, |
| .dropStmt = delq->data)); |
| |
| if (pubinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "PUBLICATION", qpubname, |
| NULL, pubinfo->rolname, |
| pubinfo->dobj.catId, 0, pubinfo->dobj.dumpId); |
| |
| if (pubinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "PUBLICATION", qpubname, |
| NULL, pubinfo->rolname, |
| pubinfo->dobj.catId, 0, pubinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qpubname); |
| } |
| |
| /* |
| * getPublicationNamespaces |
| * get information about publication membership for dumpable schemas. |
| */ |
| void |
| getPublicationNamespaces(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| PublicationSchemaInfo *pubsinfo; |
| DumpOptions *dopt = fout->dopt; |
| int i_tableoid; |
| int i_oid; |
| int i_pnpubid; |
| int i_pnnspid; |
| int i, |
| j, |
| ntups; |
| |
| if (dopt->no_publications || fout->remoteVersion < 150000) |
| return; |
| |
| query = createPQExpBuffer(); |
| |
| /* Collect all publication membership info. */ |
| appendPQExpBufferStr(query, |
| "SELECT tableoid, oid, pnpubid, pnnspid " |
| "FROM pg_catalog.pg_publication_namespace"); |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_pnpubid = PQfnumber(res, "pnpubid"); |
| i_pnnspid = PQfnumber(res, "pnnspid"); |
| |
| /* this allocation may be more than we need */ |
| pubsinfo = pg_malloc(ntups * sizeof(PublicationSchemaInfo)); |
| j = 0; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| Oid pnpubid = atooid(PQgetvalue(res, i, i_pnpubid)); |
| Oid pnnspid = atooid(PQgetvalue(res, i, i_pnnspid)); |
| PublicationInfo *pubinfo; |
| NamespaceInfo *nspinfo; |
| |
| /* |
| * Ignore any entries for which we aren't interested in either the |
| * publication or the rel. |
| */ |
| pubinfo = findPublicationByOid(pnpubid); |
| if (pubinfo == NULL) |
| continue; |
| nspinfo = findNamespaceByOid(pnnspid); |
| if (nspinfo == NULL) |
| continue; |
| |
| /* |
| * We always dump publication namespaces unless the corresponding |
| * namespace is excluded from the dump. |
| */ |
| if (nspinfo->dobj.dump == DUMP_COMPONENT_NONE) |
| continue; |
| |
| /* OK, make a DumpableObject for this relationship */ |
| pubsinfo[j].dobj.objType = DO_PUBLICATION_TABLE_IN_SCHEMA; |
| pubsinfo[j].dobj.catId.tableoid = |
| atooid(PQgetvalue(res, i, i_tableoid)); |
| pubsinfo[j].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&pubsinfo[j].dobj); |
| pubsinfo[j].dobj.namespace = nspinfo->dobj.namespace; |
| pubsinfo[j].dobj.name = nspinfo->dobj.name; |
| pubsinfo[j].publication = pubinfo; |
| pubsinfo[j].pubschema = nspinfo; |
| |
| /* Decide whether we want to dump it */ |
| selectDumpablePublicationObject(&(pubsinfo[j].dobj), fout); |
| |
| j++; |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getPublicationTables |
| * get information about publication membership for dumpable tables. |
| */ |
| void |
| getPublicationTables(Archive *fout, TableInfo tblinfo[], int numTables) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| PublicationRelInfo *pubrinfo; |
| DumpOptions *dopt = fout->dopt; |
| int i_tableoid; |
| int i_oid; |
| int i_prpubid; |
| int i_prrelid; |
| int i_prrelqual; |
| int i_prattrs; |
| int i, |
| j, |
| ntups; |
| |
| if (dopt->no_publications || fout->remoteVersion < 100000) |
| return; |
| |
| query = createPQExpBuffer(); |
| |
| /* Collect all publication membership info. */ |
| if (fout->remoteVersion >= 150000) |
| appendPQExpBufferStr(query, |
| "SELECT tableoid, oid, prpubid, prrelid, " |
| "pg_catalog.pg_get_expr(prqual, prrelid) AS prrelqual, " |
| "(CASE\n" |
| " WHEN pr.prattrs IS NOT NULL THEN\n" |
| " (SELECT array_agg(attname)\n" |
| " FROM\n" |
| " pg_catalog.generate_series(0, pg_catalog.array_upper(pr.prattrs::pg_catalog.int2[], 1)) s,\n" |
| " pg_catalog.pg_attribute\n" |
| " WHERE attrelid = pr.prrelid AND attnum = prattrs[s])\n" |
| " ELSE NULL END) prattrs " |
| "FROM pg_catalog.pg_publication_rel pr"); |
| else |
| appendPQExpBufferStr(query, |
| "SELECT tableoid, oid, prpubid, prrelid, " |
| "NULL AS prrelqual, NULL AS prattrs " |
| "FROM pg_catalog.pg_publication_rel"); |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_prpubid = PQfnumber(res, "prpubid"); |
| i_prrelid = PQfnumber(res, "prrelid"); |
| i_prrelqual = PQfnumber(res, "prrelqual"); |
| i_prattrs = PQfnumber(res, "prattrs"); |
| |
| /* this allocation may be more than we need */ |
| pubrinfo = pg_malloc(ntups * sizeof(PublicationRelInfo)); |
| j = 0; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| Oid prpubid = atooid(PQgetvalue(res, i, i_prpubid)); |
| Oid prrelid = atooid(PQgetvalue(res, i, i_prrelid)); |
| PublicationInfo *pubinfo; |
| TableInfo *tbinfo; |
| |
| /* |
| * Ignore any entries for which we aren't interested in either the |
| * publication or the rel. |
| */ |
| pubinfo = findPublicationByOid(prpubid); |
| if (pubinfo == NULL) |
| continue; |
| tbinfo = findTableByOid(prrelid); |
| if (tbinfo == NULL) |
| continue; |
| |
| /* |
| * Ignore publication membership of tables whose definitions are not |
| * to be dumped. |
| */ |
| if (!(tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION)) |
| continue; |
| |
| /* OK, make a DumpableObject for this relationship */ |
| pubrinfo[j].dobj.objType = DO_PUBLICATION_REL; |
| pubrinfo[j].dobj.catId.tableoid = |
| atooid(PQgetvalue(res, i, i_tableoid)); |
| pubrinfo[j].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&pubrinfo[j].dobj); |
| pubrinfo[j].dobj.namespace = tbinfo->dobj.namespace; |
| pubrinfo[j].dobj.name = tbinfo->dobj.name; |
| pubrinfo[j].publication = pubinfo; |
| pubrinfo[j].pubtable = tbinfo; |
| if (PQgetisnull(res, i, i_prrelqual)) |
| pubrinfo[j].pubrelqual = NULL; |
| else |
| pubrinfo[j].pubrelqual = pg_strdup(PQgetvalue(res, i, i_prrelqual)); |
| |
| if (!PQgetisnull(res, i, i_prattrs)) |
| { |
| char **attnames; |
| int nattnames; |
| PQExpBuffer attribs; |
| |
| if (!parsePGArray(PQgetvalue(res, i, i_prattrs), |
| &attnames, &nattnames)) |
| pg_fatal("could not parse %s array", "prattrs"); |
| attribs = createPQExpBuffer(); |
| for (int k = 0; k < nattnames; k++) |
| { |
| if (k > 0) |
| appendPQExpBufferStr(attribs, ", "); |
| |
| appendPQExpBufferStr(attribs, fmtId(attnames[k])); |
| } |
| pubrinfo[j].pubrattrs = attribs->data; |
| } |
| else |
| pubrinfo[j].pubrattrs = NULL; |
| |
| /* Decide whether we want to dump it */ |
| selectDumpablePublicationObject(&(pubrinfo[j].dobj), fout); |
| |
| j++; |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpPublicationNamespace |
| * dump the definition of the given publication schema mapping. |
| */ |
| static void |
| dumpPublicationNamespace(Archive *fout, const PublicationSchemaInfo *pubsinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| NamespaceInfo *schemainfo = pubsinfo->pubschema; |
| PublicationInfo *pubinfo = pubsinfo->publication; |
| PQExpBuffer query; |
| char *tag; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| tag = psprintf("%s %s", pubinfo->dobj.name, schemainfo->dobj.name); |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "ALTER PUBLICATION %s ", fmtId(pubinfo->dobj.name)); |
| appendPQExpBuffer(query, "ADD TABLES IN SCHEMA %s;\n", fmtId(schemainfo->dobj.name)); |
| |
| /* |
| * There is no point in creating drop query as the drop is done by schema |
| * drop. |
| */ |
| if (pubsinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, pubsinfo->dobj.catId, pubsinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = schemainfo->dobj.name, |
| .owner = pubinfo->rolname, |
| .description = "PUBLICATION TABLES IN SCHEMA", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data)); |
| |
| /* These objects can't currently have comments or seclabels */ |
| |
| free(tag); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpPublicationTable |
| * dump the definition of the given publication table mapping |
| */ |
| static void |
| dumpPublicationTable(Archive *fout, const PublicationRelInfo *pubrinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PublicationInfo *pubinfo = pubrinfo->publication; |
| TableInfo *tbinfo = pubrinfo->pubtable; |
| PQExpBuffer query; |
| char *tag; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| tag = psprintf("%s %s", pubinfo->dobj.name, tbinfo->dobj.name); |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "ALTER PUBLICATION %s ADD TABLE ONLY", |
| fmtId(pubinfo->dobj.name)); |
| appendPQExpBuffer(query, " %s", |
| fmtQualifiedDumpable(tbinfo)); |
| |
| if (pubrinfo->pubrattrs) |
| appendPQExpBuffer(query, " (%s)", pubrinfo->pubrattrs); |
| |
| if (pubrinfo->pubrelqual) |
| { |
| /* |
| * It's necessary to add parentheses around the expression because |
| * pg_get_expr won't supply the parentheses for things like WHERE |
| * TRUE. |
| */ |
| appendPQExpBuffer(query, " WHERE (%s)", pubrinfo->pubrelqual); |
| } |
| appendPQExpBufferStr(query, ";\n"); |
| |
| /* |
| * There is no point in creating a drop query as the drop is done by table |
| * drop. (If you think to change this, see also _printTocEntry().) |
| * Although this object doesn't really have ownership as such, set the |
| * owner field anyway to ensure that the command is run by the correct |
| * role at restore time. |
| */ |
| if (pubrinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, pubrinfo->dobj.catId, pubrinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = pubinfo->rolname, |
| .description = "PUBLICATION TABLE", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data)); |
| |
| /* These objects can't currently have comments or seclabels */ |
| |
| free(tag); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * Is the currently connected user a superuser? |
| */ |
| static bool |
| is_superuser(Archive *fout) |
| { |
| ArchiveHandle *AH = (ArchiveHandle *) fout; |
| const char *val; |
| |
| val = PQparameterStatus(AH->connection, "is_superuser"); |
| |
| if (val && strcmp(val, "on") == 0) |
| return true; |
| |
| return false; |
| } |
| |
| /* |
| * Set the given value to restrict_nonsystem_relation_kind value. Since |
| * restrict_nonsystem_relation_kind is introduced in minor version releases, |
| * the setting query is effective only where available. |
| */ |
| static void |
| set_restrict_relation_kind(Archive *AH, const char *value) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| |
| appendPQExpBuffer(query, |
| "SELECT set_config(name, '%s', false) " |
| "FROM pg_settings " |
| "WHERE name = 'restrict_nonsystem_relation_kind'", |
| value); |
| res = ExecuteSqlQuery(AH, query->data, PGRES_TUPLES_OK); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getSubscriptions |
| * get information about subscriptions |
| */ |
| void |
| getSubscriptions(Archive *fout) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PGresult *res; |
| SubscriptionInfo *subinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_subname; |
| int i_subowner; |
| int i_subbinary; |
| int i_substream; |
| int i_subtwophasestate; |
| int i_subdisableonerr; |
| int i_subpasswordrequired; |
| int i_subrunasowner; |
| int i_subconninfo; |
| int i_subslotname; |
| int i_subsynccommit; |
| int i_subpublications; |
| int i_suborigin; |
| int i, |
| ntups; |
| |
| if (dopt->no_subscriptions || fout->remoteVersion < 100000) |
| return; |
| |
| if (!is_superuser(fout)) |
| { |
| int n; |
| |
| res = ExecuteSqlQuery(fout, |
| "SELECT count(*) FROM pg_subscription " |
| "WHERE subdbid = (SELECT oid FROM pg_database" |
| " WHERE datname = current_database())", |
| PGRES_TUPLES_OK); |
| n = atoi(PQgetvalue(res, 0, 0)); |
| if (n > 0) |
| pg_log_warning("subscriptions not dumped because current user is not a superuser"); |
| PQclear(res); |
| return; |
| } |
| |
| query = createPQExpBuffer(); |
| |
| /* Get the subscriptions in current database. */ |
| appendPQExpBufferStr(query, |
| "SELECT s.tableoid, s.oid, s.subname,\n" |
| " s.subowner,\n" |
| " s.subconninfo, s.subslotname, s.subsynccommit,\n" |
| " s.subpublications,\n"); |
| |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(query, " s.subbinary,\n"); |
| else |
| appendPQExpBufferStr(query, " false AS subbinary,\n"); |
| |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(query, " s.substream,\n"); |
| else |
| appendPQExpBufferStr(query, " 'f' AS substream,\n"); |
| |
| if (fout->remoteVersion >= 150000) |
| appendPQExpBufferStr(query, |
| " s.subtwophasestate,\n" |
| " s.subdisableonerr,\n"); |
| else |
| appendPQExpBuffer(query, |
| " '%c' AS subtwophasestate,\n" |
| " false AS subdisableonerr,\n", |
| LOGICALREP_TWOPHASE_STATE_DISABLED); |
| |
| if (fout->remoteVersion >= 160000) |
| appendPQExpBufferStr(query, |
| " s.subpasswordrequired,\n" |
| " s.subrunasowner,\n" |
| " s.suborigin\n"); |
| else |
| appendPQExpBuffer(query, |
| " 't' AS subpasswordrequired,\n" |
| " 't' AS subrunasowner,\n" |
| " '%s' AS suborigin\n", |
| LOGICALREP_ORIGIN_ANY); |
| |
| appendPQExpBufferStr(query, |
| "FROM pg_subscription s\n" |
| "WHERE s.subdbid = (SELECT oid FROM pg_database\n" |
| " WHERE datname = current_database())"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| /* |
| * Get subscription fields. We don't include subskiplsn in the dump as |
| * after restoring the dump this value may no longer be relevant. |
| */ |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_subname = PQfnumber(res, "subname"); |
| i_subowner = PQfnumber(res, "subowner"); |
| i_subbinary = PQfnumber(res, "subbinary"); |
| i_substream = PQfnumber(res, "substream"); |
| i_subtwophasestate = PQfnumber(res, "subtwophasestate"); |
| i_subdisableonerr = PQfnumber(res, "subdisableonerr"); |
| i_subpasswordrequired = PQfnumber(res, "subpasswordrequired"); |
| i_subrunasowner = PQfnumber(res, "subrunasowner"); |
| i_subconninfo = PQfnumber(res, "subconninfo"); |
| i_subslotname = PQfnumber(res, "subslotname"); |
| i_subsynccommit = PQfnumber(res, "subsynccommit"); |
| i_subpublications = PQfnumber(res, "subpublications"); |
| i_suborigin = PQfnumber(res, "suborigin"); |
| |
| subinfo = pg_malloc(ntups * sizeof(SubscriptionInfo)); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| subinfo[i].dobj.objType = DO_SUBSCRIPTION; |
| subinfo[i].dobj.catId.tableoid = |
| atooid(PQgetvalue(res, i, i_tableoid)); |
| subinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&subinfo[i].dobj); |
| subinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_subname)); |
| subinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_subowner)); |
| |
| subinfo[i].subbinary = |
| pg_strdup(PQgetvalue(res, i, i_subbinary)); |
| subinfo[i].substream = |
| pg_strdup(PQgetvalue(res, i, i_substream)); |
| subinfo[i].subtwophasestate = |
| pg_strdup(PQgetvalue(res, i, i_subtwophasestate)); |
| subinfo[i].subdisableonerr = |
| pg_strdup(PQgetvalue(res, i, i_subdisableonerr)); |
| subinfo[i].subpasswordrequired = |
| pg_strdup(PQgetvalue(res, i, i_subpasswordrequired)); |
| subinfo[i].subrunasowner = |
| pg_strdup(PQgetvalue(res, i, i_subrunasowner)); |
| subinfo[i].subconninfo = |
| pg_strdup(PQgetvalue(res, i, i_subconninfo)); |
| if (PQgetisnull(res, i, i_subslotname)) |
| subinfo[i].subslotname = NULL; |
| else |
| subinfo[i].subslotname = |
| pg_strdup(PQgetvalue(res, i, i_subslotname)); |
| subinfo[i].subsynccommit = |
| pg_strdup(PQgetvalue(res, i, i_subsynccommit)); |
| subinfo[i].subpublications = |
| pg_strdup(PQgetvalue(res, i, i_subpublications)); |
| subinfo[i].suborigin = pg_strdup(PQgetvalue(res, i, i_suborigin)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(subinfo[i].dobj), fout); |
| } |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpSubscription |
| * dump the definition of the given subscription |
| */ |
| static void |
| dumpSubscription(Archive *fout, const SubscriptionInfo *subinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| PQExpBuffer publications; |
| char *qsubname; |
| char **pubnames = NULL; |
| int npubnames = 0; |
| int i; |
| char two_phase_disabled[] = {LOGICALREP_TWOPHASE_STATE_DISABLED, '\0'}; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| qsubname = pg_strdup(fmtId(subinfo->dobj.name)); |
| |
| appendPQExpBuffer(delq, "DROP SUBSCRIPTION %s;\n", |
| qsubname); |
| |
| appendPQExpBuffer(query, "CREATE SUBSCRIPTION %s CONNECTION ", |
| qsubname); |
| appendStringLiteralAH(query, subinfo->subconninfo, fout); |
| |
| /* Build list of quoted publications and append them to query. */ |
| if (!parsePGArray(subinfo->subpublications, &pubnames, &npubnames)) |
| pg_fatal("could not parse %s array", "subpublications"); |
| |
| publications = createPQExpBuffer(); |
| for (i = 0; i < npubnames; i++) |
| { |
| if (i > 0) |
| appendPQExpBufferStr(publications, ", "); |
| |
| appendPQExpBufferStr(publications, fmtId(pubnames[i])); |
| } |
| |
| appendPQExpBuffer(query, " PUBLICATION %s WITH (connect = false, slot_name = ", publications->data); |
| if (subinfo->subslotname) |
| appendStringLiteralAH(query, subinfo->subslotname, fout); |
| else |
| appendPQExpBufferStr(query, "NONE"); |
| |
| if (strcmp(subinfo->subbinary, "t") == 0) |
| appendPQExpBufferStr(query, ", binary = true"); |
| |
| if (strcmp(subinfo->substream, "t") == 0) |
| appendPQExpBufferStr(query, ", streaming = on"); |
| else if (strcmp(subinfo->substream, "p") == 0) |
| appendPQExpBufferStr(query, ", streaming = parallel"); |
| |
| if (strcmp(subinfo->subtwophasestate, two_phase_disabled) != 0) |
| appendPQExpBufferStr(query, ", two_phase = on"); |
| |
| if (strcmp(subinfo->subdisableonerr, "t") == 0) |
| appendPQExpBufferStr(query, ", disable_on_error = true"); |
| |
| if (strcmp(subinfo->subpasswordrequired, "t") != 0) |
| appendPQExpBuffer(query, ", password_required = false"); |
| |
| if (strcmp(subinfo->subrunasowner, "t") == 0) |
| appendPQExpBufferStr(query, ", run_as_owner = true"); |
| |
| if (strcmp(subinfo->subsynccommit, "off") != 0) |
| appendPQExpBuffer(query, ", synchronous_commit = %s", fmtId(subinfo->subsynccommit)); |
| |
| if (pg_strcasecmp(subinfo->suborigin, LOGICALREP_ORIGIN_ANY) != 0) |
| appendPQExpBuffer(query, ", origin = %s", subinfo->suborigin); |
| |
| appendPQExpBufferStr(query, ");\n"); |
| |
| if (subinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, subinfo->dobj.catId, subinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = subinfo->dobj.name, |
| .owner = subinfo->rolname, |
| .description = "SUBSCRIPTION", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data, |
| .dropStmt = delq->data)); |
| |
| if (subinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "SUBSCRIPTION", qsubname, |
| NULL, subinfo->rolname, |
| subinfo->dobj.catId, 0, subinfo->dobj.dumpId); |
| |
| if (subinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "SUBSCRIPTION", qsubname, |
| NULL, subinfo->rolname, |
| subinfo->dobj.catId, 0, subinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(publications); |
| free(pubnames); |
| |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qsubname); |
| } |
| |
| /* |
| * Given a "create query", append as many ALTER ... DEPENDS ON EXTENSION as |
| * the object needs. |
| */ |
| static void |
| append_depends_on_extension(Archive *fout, |
| PQExpBuffer create, |
| const DumpableObject *dobj, |
| const char *catalog, |
| const char *keyword, |
| const char *objname) |
| { |
| if (dobj->depends_on_ext) |
| { |
| char *nm; |
| PGresult *res; |
| PQExpBuffer query; |
| int ntups; |
| int i_extname; |
| int i; |
| |
| /* dodge fmtId() non-reentrancy */ |
| nm = pg_strdup(objname); |
| |
| query = createPQExpBuffer(); |
| appendPQExpBuffer(query, |
| "SELECT e.extname " |
| "FROM pg_catalog.pg_depend d, pg_catalog.pg_extension e " |
| "WHERE d.refobjid = e.oid AND classid = '%s'::pg_catalog.regclass " |
| "AND objid = '%u'::pg_catalog.oid AND deptype = 'x' " |
| "AND refclassid = 'pg_catalog.pg_extension'::pg_catalog.regclass", |
| catalog, |
| dobj->catId.oid); |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| ntups = PQntuples(res); |
| i_extname = PQfnumber(res, "extname"); |
| for (i = 0; i < ntups; i++) |
| { |
| appendPQExpBuffer(create, "ALTER %s %s DEPENDS ON EXTENSION %s;\n", |
| keyword, nm, |
| fmtId(PQgetvalue(res, i, i_extname))); |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| pg_free(nm); |
| } |
| } |
| |
| static Oid |
| get_next_possible_free_pg_type_oid(Archive *fout, PQExpBuffer upgrade_query) |
| { |
| /* |
| * If the old version didn't assign an array type, but the new version |
| * does, we must select an unused type OID to assign. This currently only |
| * happens for domains, when upgrading pre-v11 to v11 and up. |
| * |
| * Note: local state here is kind of ugly, but we must have some, since we |
| * mustn't choose the same unused OID more than once. |
| */ |
| static Oid next_possible_free_oid = FirstNormalObjectId; |
| PGresult *res; |
| bool is_dup; |
| |
| do |
| { |
| ++next_possible_free_oid; |
| printfPQExpBuffer(upgrade_query, |
| "SELECT EXISTS(SELECT 1 " |
| "FROM pg_catalog.pg_type " |
| "WHERE oid = '%u'::pg_catalog.oid);", |
| next_possible_free_oid); |
| res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| is_dup = (PQgetvalue(res, 0, 0)[0] == 't'); |
| PQclear(res); |
| } while (is_dup); |
| |
| return next_possible_free_oid; |
| } |
| |
| static void |
| binary_upgrade_set_namespace_oid(Archive *fout, PQExpBuffer upgrade_buffer, |
| Oid pg_namespace_oid) |
| { |
| PQExpBuffer upgrade_query = createPQExpBuffer(); |
| PGresult *upgrade_res; |
| char *pg_nspname; |
| |
| appendPQExpBuffer(upgrade_query, |
| "SELECT nspname " |
| "FROM pg_catalog.pg_namespace " |
| "WHERE oid = '%u'::pg_catalog.oid;", pg_namespace_oid); |
| upgrade_res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| pg_nspname = PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "nspname")); |
| |
| simple_oid_list_append(&preassigned_oids, pg_namespace_oid); |
| appendPQExpBuffer(upgrade_buffer, "\n-- For binary upgrade, must preserve pg_namespace oid\n"); |
| |
| /* |
| * In PostgreSQL, the creation of the public schema is not dumped, because |
| * it's assumed to exist in a new cluster. But in GPDB, we need to re-create |
| * it, so that we can set its OID. This change is required because Oid |
| * preassignment during upgrade requires the namespace. If the public |
| * schema is not re-created, Oid preassignments in public will be |
| * mismatched between old and new cluster if the schema had been |
| * dropped and recreated on the old cluster. |
| */ |
| if (strcmp(pg_nspname, "public") == 0) |
| { |
| appendPQExpBuffer(upgrade_buffer, |
| "DROP SCHEMA IF EXISTS public;\n"); |
| } |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT binary_upgrade_set_next_pg_namespace_oid('%u'::pg_catalog.oid, " |
| "$$%s$$::text);\n\n", |
| pg_namespace_oid, pg_nspname); |
| PQclear(upgrade_res); |
| destroyPQExpBuffer(upgrade_query); |
| } |
| |
| static void |
| binary_upgrade_set_type_oids_by_type_oid(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_type_oid, Oid pg_type_ns_oid, |
| char *pg_type_name, |
| bool force_array_type, |
| bool include_multirange_type) |
| { |
| PQExpBuffer upgrade_query = createPQExpBuffer(); |
| PGresult *res; |
| Oid pg_type_array_oid; |
| Oid pg_type_multirange_oid; |
| Oid pg_type_multirange_array_oid; |
| Oid pg_type_array_nsoid; |
| Oid pg_type_multirange_nsoid; |
| Oid pg_type_multirange_array_nsoid; |
| char *pg_type_array_name; |
| |
| simple_oid_list_append(&preassigned_oids, pg_type_oid); |
| appendPQExpBufferStr(upgrade_buffer, "\n-- For binary upgrade, must preserve pg_type oid\n"); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_pg_type_oid('%u'::pg_catalog.oid, " |
| "'%u'::pg_catalog.oid, " |
| "$$%s$$::text);\n\n", |
| pg_type_oid, pg_type_ns_oid, pg_type_name); |
| |
| appendPQExpBuffer(upgrade_query, |
| "SELECT t.typarray, a.typname, a.typnamespace " |
| "FROM pg_catalog.pg_type t " |
| " LEFT OUTER JOIN pg_catalog.pg_type a ON (t.typarray = a.oid) " |
| "WHERE t.oid = '%u'::pg_catalog.oid;", |
| pg_type_oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| |
| pg_type_array_oid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typarray"))); |
| pg_type_array_nsoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typnamespace"))); |
| pg_type_array_name = pstrdup(PQgetvalue(res, 0, PQfnumber(res, "typname"))); |
| |
| PQclear(res); |
| |
| if (!OidIsValid(pg_type_array_oid) && force_array_type) |
| { |
| pg_type_array_oid = get_next_possible_free_pg_type_oid(fout, upgrade_query); |
| pg_type_array_nsoid = pg_type_ns_oid; |
| pfree(pg_type_array_name); |
| pg_type_array_name = psprintf("_%s", pg_type_name); |
| |
| } |
| |
| if (OidIsValid(pg_type_array_oid)) |
| { |
| simple_oid_list_append(&preassigned_oids, pg_type_array_oid); |
| appendPQExpBufferStr(upgrade_buffer, |
| "\n-- For binary upgrade, must preserve pg_type array oid\n"); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_array_pg_type_oid('%u'::pg_catalog.oid, " |
| "'%u'::pg_catalog.oid, $$%s$$::text);\n\n", |
| pg_type_array_oid, pg_type_array_nsoid, |
| pg_type_array_name); |
| } |
| |
| /* |
| * Pre-set the multirange type oid and its own array type oid. |
| */ |
| if (include_multirange_type) |
| { |
| char *pg_type_multirange_typename; |
| char *pg_type_multirange_array_typename; |
| if (fout->remoteVersion >= 140000) |
| { |
| printfPQExpBuffer(upgrade_query, |
| "SELECT t.oid, t.typarray " |
| "FROM pg_catalog.pg_type t " |
| "JOIN pg_catalog.pg_range r " |
| "ON t.oid = r.rngmultitypid " |
| "WHERE r.rngtypid = '%u'::pg_catalog.oid;", |
| pg_type_oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| |
| pg_type_multirange_oid = atooid(PQgetvalue(res, 0, PQfnumber(res, "oid"))); |
| pg_type_multirange_array_oid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typarray"))); |
| PQclear(res); |
| |
| /* |
| * GPDB_14_MERGE_FIXME |
| * Fetch namespaceoid and typename of multirange and multirangearray. |
| * 1.Can pg_type_multirange_oid be invalid?(asume no). |
| * 2.Can pg_type_multirange_array_oid be invalid?(asume possible). |
| * Asset them both now, need to check. |
| */ |
| //Assert(!OidIsValid(pg_type_multirange_oid)); |
| //Assert(!OidIsValid(pg_type_multirange_array_oid)); |
| |
| printfPQExpBuffer(upgrade_query, |
| "SELECT t.typnamespace, t.typname " |
| "FROM pg_catalog.pg_type t " |
| "WHERE oid = '%u'::pg_catalog.oid;", |
| pg_type_multirange_oid); |
| res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| pg_type_multirange_nsoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typnamespace"))); |
| pg_type_multirange_typename = pstrdup(PQgetvalue(res, 0, PQfnumber(res, "typname"))); |
| PQclear(res); |
| |
| printfPQExpBuffer(upgrade_query, |
| "SELECT t.typnamespace, t.typname " |
| "FROM pg_catalog.pg_type t " |
| "WHERE oid = '%u'::pg_catalog.oid;", |
| pg_type_multirange_array_oid); |
| res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| pg_type_multirange_array_nsoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typnamespace"))); |
| pg_type_multirange_array_typename = pstrdup(PQgetvalue(res, 0, PQfnumber(res, "typname"))); |
| PQclear(res); |
| |
| } |
| else |
| { |
| pg_type_multirange_oid = get_next_possible_free_pg_type_oid(fout, upgrade_query); |
| pg_type_multirange_array_oid = get_next_possible_free_pg_type_oid(fout, upgrade_query); |
| /* |
| * GPDB_14_MERGE_FIXEME: |
| * namspaceoid: use param pg_type_ns_oid. |
| * typename: replace param pg_type_name's 'range' with 'multirange' with a '_" prefix |
| */ |
| pg_type_multirange_array_nsoid = pg_type_multirange_nsoid = pg_type_ns_oid; |
| |
| StringInfo strinfo = makeStringInfo(); |
| |
| appendStringInfoString(strinfo, pg_type_name); |
| replaceStringInfoString(strinfo, "range", "multirange"); |
| pg_type_multirange_typename = psprintf("_%s", strinfo->data); |
| |
| resetStringInfo(strinfo); |
| |
| appendStringInfoString(strinfo, pg_type_name); |
| replaceStringInfoString(strinfo, "range", "multirange_array"); |
| pg_type_multirange_array_typename = psprintf("_%s", strinfo->data); |
| |
| pfree(strinfo->data); |
| pfree(strinfo); |
| } |
| |
| appendPQExpBufferStr(upgrade_buffer, |
| "\n-- For binary upgrade, must preserve multirange pg_type oid\n"); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_multirange_pg_type_oid('%u'::pg_catalog.oid, " |
| "'%u'::pg_catalog.oid, $$%s$$::text);\n\n", |
| pg_type_multirange_oid, pg_type_multirange_nsoid, |
| pg_type_multirange_typename); |
| appendPQExpBufferStr(upgrade_buffer, |
| "\n-- For binary upgrade, must preserve multirange pg_type array oid\n"); |
| /* GPDB_14_MERGE_FIXME: new function here, check if multiple params are needed */ |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_multirange_array_pg_type_oid('%u'::pg_catalog.oid, " |
| "'%u'::pg_catalog.oid, $$%s$$::text);\n\n", |
| pg_type_multirange_array_oid, pg_type_multirange_array_nsoid, |
| pg_type_multirange_array_typename); |
| pfree(pg_type_multirange_typename); |
| pfree(pg_type_multirange_array_typename); |
| } |
| |
| pfree(pg_type_array_name); |
| destroyPQExpBuffer(upgrade_query); |
| } |
| |
| <<<<<<< HEAD |
| /* |
| * GPDB: the implementation of this function has moved below, to |
| * binary_upgrade_set_type_oids_by_rel_oid_impl(), so that we can handle some of |
| * the annoyances of append-only tables without having to modify upstream |
| * callers. |
| * |
| * When dumping the types for an AO auxiliary table during binary upgrade, use |
| * binary_upgrade_set_type_oids_for_ao() instead. |
| */ |
| static void |
| binary_upgrade_set_type_oids_by_rel_oid(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_rel_oid) |
| { |
| binary_upgrade_set_type_oids_by_rel_oid_impl(fout, upgrade_buffer, |
| pg_rel_oid, NULL); |
| } |
| |
| static void |
| binary_upgrade_set_type_oids_by_rel_oid_impl(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_rel_oid, |
| char *typname_override) |
| { |
| PQExpBuffer upgrade_query = createPQExpBuffer(); |
| PGresult *upgrade_res; |
| Oid pg_type_oid; |
| Oid pg_type_nsoid; |
| char *pg_type_name; |
| char *co_table_check = ""; |
| |
| /* |
| * Starting GPDB7 CO tables no longer have TOAST tables. Hence, ignore |
| * toast OIDs for CO tables to avoid upgrade failures. |
| */ |
| if (fout->remoteVersion < 120000) |
| co_table_check = " AND c.relstorage <> 'c'"; |
| |
| appendPQExpBuffer(upgrade_query, |
| "SELECT c.reltype AS crel, " |
| " ct.typnamespace, ct.typname, " |
| " tt.typnamespace AS trelns " |
| "FROM pg_catalog.pg_class c " |
| "LEFT JOIN pg_catalog.pg_class t ON " |
| " (c.reltoastrelid = t.oid AND c.relkind <> '%c'%s) " |
| "LEFT JOIN pg_catalog.pg_type ct ON (c.reltype = ct.oid) " |
| "LEFT JOIN pg_catalog.pg_type tt ON (t.reltype = tt.oid) " |
| "WHERE c.oid = '%u'::pg_catalog.oid;", |
| RELKIND_PARTITIONED_TABLE, co_table_check, pg_rel_oid); |
| |
| upgrade_res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| |
| pg_type_oid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "crel"))); |
| pg_type_nsoid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "typnamespace"))); |
| if (typname_override) |
| { |
| pg_type_name = typname_override; |
| } |
| else |
| { |
| pg_type_name = PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "typname")); |
| } |
| |
| if (OidIsValid(pg_type_oid)) |
| binary_upgrade_set_type_oids_by_type_oid(fout, upgrade_buffer, |
| pg_type_oid, pg_type_nsoid, |
| pg_type_name, false, false); |
| |
| PQclear(upgrade_res); |
| destroyPQExpBuffer(upgrade_query); |
| ======= |
| static void |
| binary_upgrade_set_type_oids_by_rel(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| const TableInfo *tbinfo) |
| { |
| Oid pg_type_oid = tbinfo->reltype; |
| |
| if (OidIsValid(pg_type_oid)) |
| binary_upgrade_set_type_oids_by_type_oid(fout, upgrade_buffer, |
| pg_type_oid, false, false); |
| >>>>>>> REL_16_9 |
| } |
| |
| /* |
| * A version of binary_upgrade_set_type_oids_by_rel_oid() that allows the caller |
| * to override the expected name of the AO auxiliary table. This is necessary |
| * because the new database will name the aux tables based on the owning table's |
| * OID, but the old database's aux table names may be out of sync with that OID |
| * due to certain DML commands, and there is no DDL that allows us to specify |
| * the names of the aux tables. |
| */ |
| static void |
| binary_upgrade_set_type_oids_for_ao(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_rel_oid, |
| char *ao_aux_typname) |
| { |
| if (!ao_aux_typname) |
| fatal("binary_upgrade_set_type_oids_for_ao() requires an AO auxiliary type name"); |
| |
| binary_upgrade_set_type_oids_by_rel_oid_impl(fout, upgrade_buffer, |
| pg_rel_oid, ao_aux_typname); |
| } |
| |
| static void |
| create_ao_relname(char *dst, size_t len, const char *prefix, Oid auxoid) |
| { |
| size_t actual = snprintf(dst, len, "%s_%u", prefix, auxoid); |
| |
| if (actual >= len) |
| fatal("create_ao_relname: destination buffer is too short"); |
| } |
| |
| static void |
| create_ao_idxname(char *dst, size_t len, const char *prefix, Oid auxoid) |
| { |
| size_t actual = snprintf(dst, len, "%s_%u_index", prefix, auxoid); |
| |
| if (actual >= len) |
| fatal("create_ao_idxname: destination buffer is too short"); |
| } |
| |
| /* |
| * GPDB: the implementation of this function has moved below, to |
| * binary_upgrade_set_pg_class_oids_impl(), so that we can handle some of the |
| * annoyances of append-only tables without having to modify upstream callers. |
| * |
| * When dumping an AO auxiliary table during binary upgrade, use |
| * binary_upgrade_set_pg_class_oids_for_ao() instead. |
| */ |
| static void |
| binary_upgrade_set_pg_class_oids(Archive *fout, |
| PQExpBuffer upgrade_buffer, Oid pg_class_oid, |
| bool is_index) |
| { |
| <<<<<<< HEAD |
| binary_upgrade_set_pg_class_oids_impl(fout, upgrade_buffer, pg_class_oid, |
| is_index, |
| /* this is not an AO table, use the old relname */ |
| NULL); |
| } |
| |
| static void |
| binary_upgrade_set_pg_class_oids_impl(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_class_oid, |
| bool is_index, |
| char *relname_override) |
| { |
| /* |
| * GPDB_14_MERGE_FIXME: |
| * try to keep the same with upstream as much as possible, |
| * but put variables here for more clear. |
| */ |
| PQExpBuffer upgrade_query = createPQExpBuffer(); |
| PGresult *upgrade_res; |
| Oid pg_class_reltoastrelid; |
| Oid pg_index_indexrelid; |
| Oid pg_class_relnamespace; |
| char *pg_class_relname; |
| Oid pg_class_reltoastnamespace; |
| char *pg_class_reltoastname; |
| char *pg_class_reltidxname; |
| Oid pg_class_bmoid; |
| Oid pg_class_bmidxoid; |
| Oid ao_segrelid = InvalidOid; |
| Oid ao_blkdirrelid = InvalidOid; |
| Oid ao_blkdiridxid = InvalidOid; |
| Oid ao_visimaprelid = InvalidOid; |
| Oid ao_visimapidxid = InvalidOid; |
| bool ao_columnstore = false; |
| char pg_class_relkind; |
| |
| /* GPDB_14_MERGE_FIXME: we must put this sql here for variables which will befetched later by other sqls */ |
| appendPQExpBuffer(upgrade_query, |
| "SELECT c.reltoastrelid, c.relkind, t.relnamespace AS toast_relnamespace, t.relname AS toast_relname, " |
| " c.relnamespace, c.relname, " |
| " i.indexrelid, ti.relname AS tidx_relname, " |
| " bi.oid AS bmoid, bidx.oid AS bmidxoid, " |
| " pgao.segrelid, pgao.columnstore, " |
| " pgao.blkdirrelid, pgao.blkdiridxid, " |
| " pgao.visimaprelid, pgao.visimapidxid " |
| "FROM pg_catalog.pg_class c " |
| "LEFT JOIN pg_catalog.pg_class t ON (c.reltoastrelid = t.oid) " |
| "LEFT JOIN pg_catalog.pg_index i ON (c.reltoastrelid = i.indrelid AND i.indisvalid) " |
| "LEFT JOIN pg_catalog.pg_class ti ON (i.indexrelid = ti.oid) " |
| "LEFT JOIN pg_catalog.pg_class bi ON (bi.relname = 'pg_bm_%u'::text) " |
| "LEFT JOIN pg_catalog.pg_class bidx ON (bidx.relname = 'pg_bm_%u_index'::text) " |
| "LEFT JOIN pg_catalog.pg_appendonly pgao ON (c.oid = pgao.relid) " |
| "WHERE c.oid = '%u'::pg_catalog.oid;", |
| pg_class_oid, pg_class_oid, pg_class_oid); |
| |
| upgrade_res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| |
| pg_class_reltoastrelid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "reltoastrelid"))); |
| pg_class_relkind = *PQgetvalue(upgrade_res, 0, |
| PQfnumber(upgrade_res, "relkind")); |
| pg_index_indexrelid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "indexrelid"))); |
| |
| /* Cloudberry specific values */ |
| pg_class_relnamespace = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "relnamespace"))); |
| if (relname_override) |
| { |
| pg_class_relname = relname_override; |
| } |
| else |
| { |
| pg_class_relname = PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "relname")); |
| } |
| pg_class_reltoastnamespace = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "toast_relnamespace"))); |
| pg_class_reltoastname = PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "toast_relname")); |
| pg_class_reltidxname = PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "tidx_relname")); |
| pg_class_bmoid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "bmoid"))); |
| pg_class_bmidxoid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "bmidxoid"))); |
| |
| if (!PQgetisnull(upgrade_res, 0, PQfnumber(upgrade_res, "segrelid"))) |
| { |
| ao_segrelid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "segrelid"))); |
| ao_columnstore = (PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "columnstore"))[0] == 't'); |
| ao_blkdirrelid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "blkdirrelid"))); |
| ao_blkdiridxid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "blkdiridxid"))); |
| ao_visimaprelid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "visimaprelid"))); |
| ao_visimapidxid = atooid(PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "visimapidxid"))); |
| } |
| |
| ======= |
| PQExpBuffer upgrade_query = createPQExpBuffer(); |
| PGresult *upgrade_res; |
| RelFileNumber relfilenumber; |
| Oid toast_oid; |
| RelFileNumber toast_relfilenumber; |
| char relkind; |
| Oid toast_index_oid; |
| RelFileNumber toast_index_relfilenumber; |
| |
| /* |
| * Preserve the OID and relfilenumber of the table, table's index, table's |
| * toast table and toast table's index if any. |
| * |
| * One complexity is that the current table definition might not require |
| * the creation of a TOAST table, but the old database might have a TOAST |
| * table that was created earlier, before some wide columns were dropped. |
| * By setting the TOAST oid we force creation of the TOAST heap and index |
| * by the new backend, so we can copy the files during binary upgrade |
| * without worrying about this case. |
| */ |
| appendPQExpBuffer(upgrade_query, |
| "SELECT c.relkind, c.relfilenode, c.reltoastrelid, ct.relfilenode AS toast_relfilenode, i.indexrelid, cti.relfilenode AS toast_index_relfilenode " |
| "FROM pg_catalog.pg_class c LEFT JOIN " |
| "pg_catalog.pg_index i ON (c.reltoastrelid = i.indrelid AND i.indisvalid) " |
| "LEFT JOIN pg_catalog.pg_class ct ON (c.reltoastrelid = ct.oid) " |
| "LEFT JOIN pg_catalog.pg_class AS cti ON (i.indexrelid = cti.oid) " |
| "WHERE c.oid = '%u'::pg_catalog.oid;", |
| pg_class_oid); |
| |
| upgrade_res = ExecuteSqlQueryForSingleRow(fout, upgrade_query->data); |
| |
| relkind = *PQgetvalue(upgrade_res, 0, PQfnumber(upgrade_res, "relkind")); |
| |
| relfilenumber = atooid(PQgetvalue(upgrade_res, 0, |
| PQfnumber(upgrade_res, "relfilenode"))); |
| toast_oid = atooid(PQgetvalue(upgrade_res, 0, |
| PQfnumber(upgrade_res, "reltoastrelid"))); |
| toast_relfilenumber = atooid(PQgetvalue(upgrade_res, 0, |
| PQfnumber(upgrade_res, "toast_relfilenode"))); |
| toast_index_oid = atooid(PQgetvalue(upgrade_res, 0, |
| PQfnumber(upgrade_res, "indexrelid"))); |
| toast_index_relfilenumber = atooid(PQgetvalue(upgrade_res, 0, |
| PQfnumber(upgrade_res, "toast_index_relfilenode"))); |
| >>>>>>> REL_16_9 |
| |
| appendPQExpBufferStr(upgrade_buffer, |
| "\n-- For binary upgrade, must preserve pg_class oids and relfilenodes\n"); |
| |
| if (!is_index) |
| { |
| <<<<<<< HEAD |
| ======= |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_heap_pg_class_oid('%u'::pg_catalog.oid);\n", |
| pg_class_oid); |
| |
| >>>>>>> REL_16_9 |
| /* |
| * Not every relation has storage. Also, in a pre-v12 database, |
| * partitioned tables have a relfilenumber, which should not be |
| * preserved when upgrading. |
| */ |
| <<<<<<< HEAD |
| |
| simple_oid_list_append(&preassigned_oids, pg_class_oid); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_heap_pg_class_oid('%u'::pg_catalog.oid, " |
| "'%u'::pg_catalog.oid, $$%s$$::text);\n", |
| pg_class_oid, pg_class_relnamespace, pg_class_relname); |
| ======= |
| if (RelFileNumberIsValid(relfilenumber) && relkind != RELKIND_PARTITIONED_TABLE) |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_heap_relfilenode('%u'::pg_catalog.oid);\n", |
| relfilenumber); |
| >>>>>>> REL_16_9 |
| |
| /* |
| * In a pre-v12 database, partitioned tables might be marked as having |
| * toast tables, but we should ignore them if so. |
| * |
| * One complexity is that the table definition might not require |
| * the creation of a TOAST table, and the TOAST table might have |
| * been created long after table creation, when the table was |
| * loaded with wide data. By setting the TOAST oid we force |
| * creation of the TOAST heap and TOAST index by the backend so we |
| * can cleanly copy the files during binary upgrade. |
| * |
| * GPDB: note that we compose the toast table name using the |
| * relation OID, rather than using whatever name was in the old |
| * cluster. Some operations can cause the old TOAST table name not |
| * to match its owner's OID, but the new cluster will be using the |
| * correct name, and it's the new cluster's name that we have to use |
| * in preassignment. |
| */ |
| <<<<<<< HEAD |
| if (OidIsValid(pg_class_reltoastrelid) && |
| pg_class_relkind != RELKIND_PARTITIONED_TABLE && !ao_columnstore) |
| { |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_toast_pg_class_oid('%u'::pg_catalog.oid, '%u'::pg_catalog.oid, $$pg_toast_%u$$::text);\n", |
| pg_class_reltoastrelid, pg_class_reltoastnamespace, pg_class_oid); |
| ======= |
| if (OidIsValid(toast_oid) && |
| relkind != RELKIND_PARTITIONED_TABLE) |
| { |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_toast_pg_class_oid('%u'::pg_catalog.oid);\n", |
| toast_oid); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_toast_relfilenode('%u'::pg_catalog.oid);\n", |
| toast_relfilenumber); |
| >>>>>>> REL_16_9 |
| |
| /* every toast table has an index */ |
| simple_oid_list_append(&preassigned_oids, pg_index_indexrelid); |
| appendPQExpBuffer(upgrade_buffer, |
| <<<<<<< HEAD |
| "SELECT pg_catalog.binary_upgrade_set_next_index_pg_class_oid('%u'::pg_catalog.oid , '%u'::pg_catalog.oid, $$pg_toast_%u_index$$::text);\n", |
| pg_index_indexrelid, pg_class_reltoastnamespace, pg_class_oid); |
| } |
| |
| /* Set up any AO auxiliary tables with preallocated OIDs as well. */ |
| if (OidIsValid(ao_segrelid)) |
| { |
| /* |
| * Adjust the names of all pg_aoseg aux tables to match what they |
| * will be in the new cluster, using the OID of the owning table. In |
| * many cases these will be the same as the ones in the old cluster, |
| * but not always. |
| */ |
| char ao_relname[64]; |
| const char *aoseg_prefix = ao_columnstore ? "pg_aocsseg" : "pg_aoseg"; |
| |
| create_ao_relname(ao_relname, sizeof(ao_relname), aoseg_prefix, pg_class_oid); |
| binary_upgrade_set_pg_class_oids_for_ao(fout, upgrade_buffer, ao_segrelid, false, ao_relname); |
| binary_upgrade_set_type_oids_for_ao(fout, upgrade_buffer, ao_segrelid, ao_relname); |
| |
| /* blkdir is optional. */ |
| if (OidIsValid(ao_blkdirrelid)) |
| { |
| create_ao_relname(ao_relname, sizeof(ao_relname), "pg_aoblkdir", pg_class_oid); |
| binary_upgrade_set_pg_class_oids_for_ao(fout, upgrade_buffer, ao_blkdirrelid, false, ao_relname); |
| binary_upgrade_set_type_oids_for_ao(fout, upgrade_buffer, ao_blkdirrelid, ao_relname); |
| |
| create_ao_idxname(ao_relname, sizeof(ao_relname), "pg_aoblkdir", pg_class_oid); |
| binary_upgrade_set_pg_class_oids_for_ao(fout, upgrade_buffer, ao_blkdiridxid, true, ao_relname); |
| } |
| |
| create_ao_relname(ao_relname, sizeof(ao_relname), "pg_aovisimap", pg_class_oid); |
| binary_upgrade_set_pg_class_oids_for_ao(fout, upgrade_buffer, ao_visimaprelid, false, ao_relname); |
| binary_upgrade_set_type_oids_for_ao(fout, upgrade_buffer, ao_visimaprelid, ao_relname); |
| |
| create_ao_idxname(ao_relname, sizeof(ao_relname), "pg_aovisimap", pg_class_oid); |
| binary_upgrade_set_pg_class_oids_for_ao(fout, upgrade_buffer, ao_visimapidxid, true, ao_relname); |
| ======= |
| "SELECT pg_catalog.binary_upgrade_set_next_index_pg_class_oid('%u'::pg_catalog.oid);\n", |
| toast_index_oid); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_index_relfilenode('%u'::pg_catalog.oid);\n", |
| toast_index_relfilenumber); |
| >>>>>>> REL_16_9 |
| } |
| } |
| else |
| { |
| <<<<<<< HEAD |
| simple_oid_list_append(&preassigned_oids, pg_class_oid); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_index_pg_class_oid('%u'::pg_catalog.oid, '%u'::pg_catalog.oid, $$%s$$::text);\n", |
| pg_class_oid, pg_class_relnamespace, pg_class_relname); |
| |
| /* |
| * If this is a bitmap index, we need to preserve the oid of the aux |
| * heap table pg_bm_<oid> as well. |
| */ |
| if (OidIsValid(pg_class_bmoid)) |
| { |
| binary_upgrade_set_pg_class_oids(fout, upgrade_buffer, pg_class_bmoid, false); |
| binary_upgrade_set_type_oids_by_rel_oid(fout, upgrade_buffer, pg_class_bmoid); |
| |
| binary_upgrade_set_pg_class_oids(fout, upgrade_buffer, pg_class_bmidxoid, true); |
| } |
| } |
| ======= |
| /* Preserve the OID and relfilenumber of the index */ |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_index_pg_class_oid('%u'::pg_catalog.oid);\n", |
| pg_class_oid); |
| appendPQExpBuffer(upgrade_buffer, |
| "SELECT pg_catalog.binary_upgrade_set_next_index_relfilenode('%u'::pg_catalog.oid);\n", |
| relfilenumber); |
| } |
| |
| PQclear(upgrade_res); |
| >>>>>>> REL_16_9 |
| |
| appendPQExpBufferChar(upgrade_buffer, '\n'); |
| |
| destroyPQExpBuffer(upgrade_query); |
| } |
| |
| /* |
| * A version of binary_upgrade_set_pg_class_oids() that allows the caller to |
| * override the expected name of the AO auxiliary table. See the notes for |
| * binary_upgrade_set_type_oids_for_ao() for the rationale. |
| * */ |
| static void |
| binary_upgrade_set_pg_class_oids_for_ao(Archive *fout, |
| PQExpBuffer upgrade_buffer, |
| Oid pg_class_oid, |
| bool is_index, |
| char *ao_aux_relname) |
| { |
| if (!ao_aux_relname) |
| fatal("binary_upgrade_set_pg_class_oids_for_ao() requires an AO auxiliary relname"); |
| |
| binary_upgrade_set_pg_class_oids_impl(fout, upgrade_buffer, pg_class_oid, |
| is_index, ao_aux_relname); |
| } |
| |
| /* |
| * If the DumpableObject is a member of an extension, add a suitable |
| * ALTER EXTENSION ADD command to the creation commands in upgrade_buffer. |
| * |
| * For somewhat historical reasons, objname should already be quoted, |
| * but not objnamespace (if any). |
| */ |
| static void |
| binary_upgrade_extension_member(PQExpBuffer upgrade_buffer, |
| const DumpableObject *dobj, |
| const char *objtype, |
| const char *objname, |
| const char *objnamespace) |
| { |
| DumpableObject *extobj = NULL; |
| int i; |
| |
| if (!dobj->ext_member) |
| return; |
| |
| /* |
| * Find the parent extension. We could avoid this search if we wanted to |
| * add a link field to DumpableObject, but the space costs of that would |
| * be considerable. We assume that member objects could only have a |
| * direct dependency on their own extension, not any others. |
| */ |
| for (i = 0; i < dobj->nDeps; i++) |
| { |
| extobj = findObjectByDumpId(dobj->dependencies[i]); |
| if (extobj && extobj->objType == DO_EXTENSION) |
| break; |
| extobj = NULL; |
| } |
| if (extobj == NULL) |
| pg_fatal("could not find parent extension for %s %s", |
| objtype, objname); |
| |
| appendPQExpBufferStr(upgrade_buffer, |
| "\n-- For binary upgrade, handle extension membership the hard way\n"); |
| appendPQExpBuffer(upgrade_buffer, "ALTER EXTENSION %s ADD %s ", |
| fmtId(extobj->name), |
| objtype); |
| if (objnamespace && *objnamespace) |
| appendPQExpBuffer(upgrade_buffer, "%s.", fmtId(objnamespace)); |
| appendPQExpBuffer(upgrade_buffer, "%s;\n", objname); |
| } |
| |
| /* |
| * getNamespaces: |
| * read all namespaces in the system catalogs and return them in the |
| * NamespaceInfo* structure |
| * |
| * numNamespaces is set to the number of namespaces read in |
| */ |
| NamespaceInfo * |
| getNamespaces(Archive *fout, int *numNamespaces) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| NamespaceInfo *nsinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_nspname; |
| int i_nspowner; |
| int i_nspacl; |
| int i_acldefault; |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * we fetch all namespaces including system ones, so that every object we |
| * read in can be linked to a containing namespace. |
| */ |
| appendPQExpBufferStr(query, "SELECT n.tableoid, n.oid, n.nspname, " |
| "n.nspowner, " |
| "n.nspacl, " |
| "acldefault('n', n.nspowner) AS acldefault " |
| "FROM pg_namespace n"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| nsinfo = (NamespaceInfo *) pg_malloc(ntups * sizeof(NamespaceInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_nspname = PQfnumber(res, "nspname"); |
| i_nspowner = PQfnumber(res, "nspowner"); |
| i_nspacl = PQfnumber(res, "nspacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| const char *nspowner; |
| |
| nsinfo[i].dobj.objType = DO_NAMESPACE; |
| nsinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| nsinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&nsinfo[i].dobj); |
| nsinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_nspname)); |
| nsinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_nspacl)); |
| nsinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| nsinfo[i].dacl.privtype = 0; |
| nsinfo[i].dacl.initprivs = NULL; |
| nspowner = PQgetvalue(res, i, i_nspowner); |
| nsinfo[i].nspowner = atooid(nspowner); |
| nsinfo[i].rolname = getRoleName(nspowner); |
| |
| /* Decide whether to dump this namespace */ |
| selectDumpableNamespace(&nsinfo[i], fout); |
| |
| /* Mark whether namespace has an ACL */ |
| if (!PQgetisnull(res, i, i_nspacl)) |
| nsinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| |
| /* |
| * We ignore any pg_init_privs.initprivs entry for the public schema |
| * and assume a predetermined default, for several reasons. First, |
| * dropping and recreating the schema removes its pg_init_privs entry, |
| * but an empty destination database starts with this ACL nonetheless. |
| * Second, we support dump/reload of public schema ownership changes. |
| * ALTER SCHEMA OWNER filters nspacl through aclnewowner(), but |
| * initprivs continues to reflect the initial owner. Hence, |
| * synthesize the value that nspacl will have after the restore's |
| * ALTER SCHEMA OWNER. Third, this makes the destination database |
| * match the source's ACL, even if the latter was an initdb-default |
| * ACL, which changed in v15. An upgrade pulls in changes to most |
| * system object ACLs that the DBA had not customized. We've made the |
| * public schema depart from that, because changing its ACL so easily |
| * breaks applications. |
| */ |
| if (strcmp(nsinfo[i].dobj.name, "public") == 0) |
| { |
| PQExpBuffer aclarray = createPQExpBuffer(); |
| PQExpBuffer aclitem = createPQExpBuffer(); |
| |
| /* Standard ACL as of v15 is {owner=UC/owner,=U/owner} */ |
| appendPQExpBufferChar(aclarray, '{'); |
| quoteAclUserName(aclitem, nsinfo[i].rolname); |
| appendPQExpBufferStr(aclitem, "=UC/"); |
| quoteAclUserName(aclitem, nsinfo[i].rolname); |
| appendPGArray(aclarray, aclitem->data); |
| resetPQExpBuffer(aclitem); |
| appendPQExpBufferStr(aclitem, "=U/"); |
| quoteAclUserName(aclitem, nsinfo[i].rolname); |
| appendPGArray(aclarray, aclitem->data); |
| appendPQExpBufferChar(aclarray, '}'); |
| |
| nsinfo[i].dacl.privtype = 'i'; |
| nsinfo[i].dacl.initprivs = pstrdup(aclarray->data); |
| nsinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| |
| destroyPQExpBuffer(aclarray); |
| destroyPQExpBuffer(aclitem); |
| } |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| *numNamespaces = ntups; |
| |
| return nsinfo; |
| } |
| |
| /* |
| * findNamespace: |
| * given a namespace OID, look up the info read by getNamespaces |
| */ |
| static NamespaceInfo * |
| findNamespace(Oid nsoid) |
| { |
| NamespaceInfo *nsinfo; |
| |
| nsinfo = findNamespaceByOid(nsoid); |
| if (nsinfo == NULL) |
| pg_fatal("schema with OID %u does not exist", nsoid); |
| return nsinfo; |
| } |
| |
| /* |
| * getExtensions: |
| * read all extensions in the system catalogs and return them in the |
| * ExtensionInfo* structure |
| * |
| * numExtensions is set to the number of extensions read in |
| */ |
| ExtensionInfo * |
| getExtensions(Archive *fout, int *numExtensions) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| ExtensionInfo *extinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_extname; |
| int i_nspname; |
| int i_extrelocatable; |
| int i_extversion; |
| int i_extconfig; |
| int i_extcondition; |
| |
| <<<<<<< HEAD |
| /* |
| * Before 8.3 (porting from PG 9.1), there are no extensions. |
| */ |
| if (fout->remoteVersion < 80300) |
| { |
| *numExtensions = 0; |
| return NULL; |
| } |
| |
| ======= |
| >>>>>>> REL_16_9 |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT x.tableoid, x.oid, " |
| "x.extname, n.nspname, x.extrelocatable, x.extversion, x.extconfig, x.extcondition " |
| "FROM pg_extension x " |
| "JOIN pg_namespace n ON n.oid = x.extnamespace"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| extinfo = (ExtensionInfo *) pg_malloc(ntups * sizeof(ExtensionInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_extname = PQfnumber(res, "extname"); |
| i_nspname = PQfnumber(res, "nspname"); |
| i_extrelocatable = PQfnumber(res, "extrelocatable"); |
| i_extversion = PQfnumber(res, "extversion"); |
| i_extconfig = PQfnumber(res, "extconfig"); |
| i_extcondition = PQfnumber(res, "extcondition"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| extinfo[i].dobj.objType = DO_EXTENSION; |
| extinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| extinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&extinfo[i].dobj); |
| extinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_extname)); |
| extinfo[i].namespace = pg_strdup(PQgetvalue(res, i, i_nspname)); |
| extinfo[i].relocatable = *(PQgetvalue(res, i, i_extrelocatable)) == 't'; |
| extinfo[i].extversion = pg_strdup(PQgetvalue(res, i, i_extversion)); |
| extinfo[i].extconfig = pg_strdup(PQgetvalue(res, i, i_extconfig)); |
| extinfo[i].extcondition = pg_strdup(PQgetvalue(res, i, i_extcondition)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableExtension(&(extinfo[i]), dopt); |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| *numExtensions = ntups; |
| |
| return extinfo; |
| } |
| |
| BinaryUpgradeInfo * |
| getBinaryUpgradeObjects(void) |
| { |
| BinaryUpgradeInfo *binfo; |
| |
| binfo = (BinaryUpgradeInfo *) pg_malloc0(sizeof(BinaryUpgradeInfo)); |
| |
| binfo->dobj.objType = DO_BINARY_UPGRADE; |
| AssignDumpId(&binfo->dobj); |
| binfo->dobj.name = pg_strdup("__binary_upgrade"); |
| |
| binary_upgrade_dumpid = binfo->dobj.dumpId; |
| |
| return binfo; |
| } |
| |
| /* |
| * getTypes: |
| * read all types in the system catalogs and return them in the |
| * TypeInfo* structure |
| * |
| * numTypes is set to the number of types read in |
| * |
| * NB: this must run after getFuncs() because we assume we can do |
| * findFuncByOid(). |
| */ |
| TypeInfo * |
| getTypes(Archive *fout, int *numTypes) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| TypeInfo *tyinfo; |
| ShellTypeInfo *stinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_typname; |
| int i_typnamespace; |
| int i_typacl; |
| int i_acldefault; |
| int i_typowner; |
| int i_typelem; |
| int i_typrelid; |
| int i_typrelkind; |
| int i_typtype; |
| int i_typisdefined; |
| int i_isarray; |
| |
| /* |
| * we include even the built-in types because those may be used as array |
| * elements by user-defined types |
| * |
| * we filter out the built-in types when we dump out the types |
| * |
| * same approach for undefined (shell) types and array types |
| * |
| * Note: as of 8.3 we can reliably detect whether a type is an |
| * auto-generated array type by checking the element type's typarray. |
| * (Before that the test is capable of generating false positives.) We |
| * still check for name beginning with '_', though, so as to avoid the |
| * cost of the subselect probe for all standard types. This would have to |
| * be revisited if the backend ever allows renaming of array types. |
| */ |
| <<<<<<< HEAD |
| |
| if (fout->remoteVersion >= 90600) |
| { |
| PQExpBuffer acl_subquery = createPQExpBuffer(); |
| PQExpBuffer racl_subquery = createPQExpBuffer(); |
| PQExpBuffer initacl_subquery = createPQExpBuffer(); |
| PQExpBuffer initracl_subquery = createPQExpBuffer(); |
| |
| buildACLQueries(acl_subquery, racl_subquery, initacl_subquery, |
| initracl_subquery, "t.typacl", "t.typowner", "'T'", |
| dopt->binary_upgrade); |
| |
| appendPQExpBuffer(query, "SELECT t.tableoid, t.oid, t.typname, " |
| "t.typnamespace, " |
| "%s AS typacl, " |
| "%s AS rtypacl, " |
| "%s AS inittypacl, " |
| "%s AS initrtypacl, " |
| "(%s t.typowner) AS rolname, " |
| "t.typelem, t.typrelid, " |
| "CASE WHEN t.typrelid = 0 THEN ' '::\"char\" " |
| "ELSE (SELECT relkind FROM pg_class WHERE oid = t.typrelid) END AS typrelkind, " |
| "t.typtype, t.typisdefined, " |
| "t.typname[0] = '_' AND t.typelem != 0 AND " |
| "(SELECT typarray FROM pg_type te WHERE oid = t.typelem) = t.oid AS isarray " |
| "FROM pg_type t " |
| "LEFT JOIN pg_init_privs pip ON " |
| "(t.oid = pip.objoid " |
| "AND pip.classoid = 'pg_type'::regclass " |
| "AND pip.objsubid = 0) ", |
| acl_subquery->data, |
| racl_subquery->data, |
| initacl_subquery->data, |
| initracl_subquery->data, |
| username_subquery); |
| |
| destroyPQExpBuffer(acl_subquery); |
| destroyPQExpBuffer(racl_subquery); |
| destroyPQExpBuffer(initacl_subquery); |
| destroyPQExpBuffer(initracl_subquery); |
| } |
| else if (fout->remoteVersion >= 90200) |
| { |
| appendPQExpBuffer(query, "SELECT tableoid, oid, typname, " |
| "typnamespace, typacl, NULL as rtypacl, " |
| "NULL AS inittypacl, NULL AS initrtypacl, " |
| "(%s typowner) AS rolname, " |
| "typelem, typrelid, " |
| "CASE WHEN typrelid = 0 THEN ' '::\"char\" " |
| "ELSE (SELECT relkind FROM pg_class WHERE oid = typrelid) END AS typrelkind, " |
| "typtype, typisdefined, " |
| "typname[0] = '_' AND typelem != 0 AND " |
| "(SELECT typarray FROM pg_type te WHERE oid = pg_type.typelem) = oid AS isarray " |
| "FROM pg_type", |
| username_subquery); |
| } |
| else |
| { |
| appendPQExpBuffer(query, "SELECT tableoid, oid, typname, " |
| "typnamespace, NULL AS typacl, NULL as rtypacl, " |
| "NULL AS inittypacl, NULL AS initrtypacl, " |
| "(%s typowner) AS rolname, " |
| "typelem, typrelid, " |
| "CASE WHEN typrelid = 0 THEN ' '::\"char\" " |
| "ELSE (SELECT relkind FROM pg_class WHERE oid = typrelid) END AS typrelkind, " |
| "typtype, typisdefined, " |
| "typname[0] = '_' AND typelem != 0 AND " |
| "(SELECT typarray FROM pg_type te WHERE oid = pg_type.typelem) = oid AS isarray " |
| "FROM pg_type", |
| username_subquery); |
| } |
| ======= |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, typname, " |
| "typnamespace, typacl, " |
| "acldefault('T', typowner) AS acldefault, " |
| "typowner, " |
| "typelem, typrelid, " |
| "CASE WHEN typrelid = 0 THEN ' '::\"char\" " |
| "ELSE (SELECT relkind FROM pg_class WHERE oid = typrelid) END AS typrelkind, " |
| "typtype, typisdefined, " |
| "typname[0] = '_' AND typelem != 0 AND " |
| "(SELECT typarray FROM pg_type te WHERE oid = pg_type.typelem) = oid AS isarray " |
| "FROM pg_type"); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| tyinfo = (TypeInfo *) pg_malloc(ntups * sizeof(TypeInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_typname = PQfnumber(res, "typname"); |
| i_typnamespace = PQfnumber(res, "typnamespace"); |
| i_typacl = PQfnumber(res, "typacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| i_typowner = PQfnumber(res, "typowner"); |
| i_typelem = PQfnumber(res, "typelem"); |
| i_typrelid = PQfnumber(res, "typrelid"); |
| i_typrelkind = PQfnumber(res, "typrelkind"); |
| i_typtype = PQfnumber(res, "typtype"); |
| i_typisdefined = PQfnumber(res, "typisdefined"); |
| i_isarray = PQfnumber(res, "isarray"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| tyinfo[i].dobj.objType = DO_TYPE; |
| tyinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| tyinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&tyinfo[i].dobj); |
| tyinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_typname)); |
| tyinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_typnamespace))); |
| <<<<<<< HEAD |
| tyinfo[i].ftypname = NULL; /* may get filled later */ |
| tyinfo[i].rolname = pg_strdup(PQgetvalue(res, i, i_rolname)); |
| tyinfo[i].typacl = pg_strdup(PQgetvalue(res, i, i_typacl)); |
| tyinfo[i].rtypacl = pg_strdup(PQgetvalue(res, i, i_rtypacl)); |
| tyinfo[i].inittypacl = pg_strdup(PQgetvalue(res, i, i_inittypacl)); |
| tyinfo[i].initrtypacl = pg_strdup(PQgetvalue(res, i, i_initrtypacl)); |
| ======= |
| tyinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_typacl)); |
| tyinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| tyinfo[i].dacl.privtype = 0; |
| tyinfo[i].dacl.initprivs = NULL; |
| tyinfo[i].ftypname = NULL; /* may get filled later */ |
| tyinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_typowner)); |
| >>>>>>> REL_16_9 |
| tyinfo[i].typelem = atooid(PQgetvalue(res, i, i_typelem)); |
| tyinfo[i].typrelid = atooid(PQgetvalue(res, i, i_typrelid)); |
| tyinfo[i].typrelkind = *PQgetvalue(res, i, i_typrelkind); |
| tyinfo[i].typtype = *PQgetvalue(res, i, i_typtype); |
| tyinfo[i].shellType = NULL; |
| |
| if (strcmp(PQgetvalue(res, i, i_typisdefined), "t") == 0) |
| tyinfo[i].isDefined = true; |
| else |
| tyinfo[i].isDefined = false; |
| |
| if (strcmp(PQgetvalue(res, i, i_isarray), "t") == 0) |
| tyinfo[i].isArray = true; |
| else |
| tyinfo[i].isArray = false; |
| |
| if (tyinfo[i].typtype == 'm') |
| tyinfo[i].isMultirange = true; |
| else |
| tyinfo[i].isMultirange = false; |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableType(&tyinfo[i], fout); |
| |
| /* Mark whether type has an ACL */ |
| if (!PQgetisnull(res, i, i_typacl)) |
| tyinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| |
| /* |
| * If it's a domain, fetch info about its constraints, if any |
| */ |
| tyinfo[i].nDomChecks = 0; |
| tyinfo[i].domChecks = NULL; |
| if ((tyinfo[i].dobj.dump & DUMP_COMPONENT_DEFINITION) && |
| tyinfo[i].typtype == TYPTYPE_DOMAIN) |
| getDomainConstraints(fout, &(tyinfo[i])); |
| |
| /* |
| * If it's a base type, make a DumpableObject representing a shell |
| * definition of the type. We will need to dump that ahead of the I/O |
| * functions for the type. Similarly, range types need a shell |
| * definition in case they have a canonicalize function. |
| * |
| * Note: the shell type doesn't have a catId. You might think it |
| * should copy the base type's catId, but then it might capture the |
| * pg_depend entries for the type, which we don't want. |
| */ |
| if ((tyinfo[i].dobj.dump & DUMP_COMPONENT_DEFINITION) && |
| (tyinfo[i].typtype == TYPTYPE_BASE || |
| tyinfo[i].typtype == TYPTYPE_RANGE)) |
| { |
| stinfo = (ShellTypeInfo *) pg_malloc(sizeof(ShellTypeInfo)); |
| stinfo->dobj.objType = DO_SHELL_TYPE; |
| stinfo->dobj.catId = nilCatalogId; |
| AssignDumpId(&stinfo->dobj); |
| stinfo->dobj.name = pg_strdup(tyinfo[i].dobj.name); |
| stinfo->dobj.namespace = tyinfo[i].dobj.namespace; |
| stinfo->baseType = &(tyinfo[i]); |
| tyinfo[i].shellType = stinfo; |
| |
| /* |
| * Initially mark the shell type as not to be dumped. We'll only |
| * dump it if the I/O or canonicalize functions need to be dumped; |
| * this is taken care of while sorting dependencies. |
| */ |
| stinfo->dobj.dump = DUMP_COMPONENT_NONE; |
| } |
| } |
| |
| *numTypes = ntups; |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return tyinfo; |
| } |
| |
| |
| |
| /* |
| * getTypeStorageOptions: |
| * read all types with storage options in the system catalogs and return them in the |
| * TypeStorageOptions* structure |
| * |
| * numTypes is set to the number of types with storage options read in |
| * |
| */ |
| TypeStorageOptions * |
| getTypeStorageOptions(Archive *fout, int *numTypes) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| TypeStorageOptions *tstorageoptions; |
| int i_tableoid; |
| int i_oid; |
| int i_typname; |
| int i_typnamespace; |
| int i_typoptions; |
| int i_rolname; |
| |
| /* |
| * The following statement used format_type to resolve an internal name to its equivalent sql name. |
| * The format_type seems to do two things, it translates an internal type name (e.g. bpchar) into its |
| * sql equivalent (e.g. character), and it puts trailing "[]" on a type if it is an array. |
| * For any user defined type (ie. oid > 10000) or any type that might be an array (ie. starts with '_'), |
| * then we will call quote_ident. If the type is a system defined type (i.e. oid <= 10000) |
| * and can not possibly be an array (i.e. does not start with '_'), then call format_type to get the name. The |
| * reason we do not call format_type for arrays is that it will return a '[]' on the end, which can not be used |
| * when dumping the type. |
| */ |
| appendPQExpBuffer(query, "SELECT " |
| "CASE WHEN t.oid > 10000 OR substring(t.typname from 1 for 1) = '_' " |
| "THEN quote_ident(t.typname) " |
| "ELSE pg_catalog.format_type(t.oid, NULL) " |
| "END as typname, " |
| "t.tableoid as tableoid, " |
| "t.oid AS oid, " |
| "t.typnamespace AS typnamespace, " |
| "(%s typowner) as rolname, " |
| "array_to_string(a.typoptions, ', ') AS typoptions " |
| "FROM pg_type t " |
| "JOIN pg_catalog.pg_type_encoding a ON a.typid = t.oid " |
| "WHERE t.typisdefined = 't'", username_subquery); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| tstorageoptions = (TypeStorageOptions *) pg_malloc(ntups * sizeof(TypeStorageOptions)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_typname = PQfnumber(res, "typname"); |
| i_typnamespace = PQfnumber(res, "typnamespace"); |
| i_typoptions = PQfnumber(res, "typoptions"); |
| i_rolname = PQfnumber(res, "rolname"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| tstorageoptions[i].dobj.objType = DO_TYPE_STORAGE_OPTIONS; |
| tstorageoptions[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| tstorageoptions[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&tstorageoptions[i].dobj); |
| tstorageoptions[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_typname)); |
| tstorageoptions[i].dobj.namespace = findNamespace(atooid(PQgetvalue(res, i, i_typnamespace))); |
| tstorageoptions[i].typoptions = pg_strdup(PQgetvalue(res, i, i_typoptions)); |
| tstorageoptions[i].rolname = pg_strdup(PQgetvalue(res, i, i_rolname)); |
| } |
| |
| *numTypes = ntups; |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return tstorageoptions; |
| } |
| |
| |
| |
| /* |
| * getOperators: |
| * read all operators in the system catalogs and return them in the |
| * OprInfo* structure |
| * |
| * numOprs is set to the number of operators read in |
| */ |
| OprInfo * |
| getOperators(Archive *fout, int *numOprs) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| OprInfo *oprinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_oprname; |
| int i_oprnamespace; |
| int i_oprowner; |
| int i_oprkind; |
| int i_oprcode; |
| |
| /* |
| * find all operators, including builtin operators; we filter out |
| * system-defined operators at dump-out time. |
| */ |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, oprname, " |
| "oprnamespace, " |
| "oprowner, " |
| "oprkind, " |
| "oprcode::oid AS oprcode " |
| "FROM pg_operator"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numOprs = ntups; |
| |
| oprinfo = (OprInfo *) pg_malloc(ntups * sizeof(OprInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_oprname = PQfnumber(res, "oprname"); |
| i_oprnamespace = PQfnumber(res, "oprnamespace"); |
| i_oprowner = PQfnumber(res, "oprowner"); |
| i_oprkind = PQfnumber(res, "oprkind"); |
| i_oprcode = PQfnumber(res, "oprcode"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| oprinfo[i].dobj.objType = DO_OPERATOR; |
| oprinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| oprinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&oprinfo[i].dobj); |
| oprinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_oprname)); |
| oprinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_oprnamespace))); |
| oprinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_oprowner)); |
| oprinfo[i].oprkind = (PQgetvalue(res, i, i_oprkind))[0]; |
| oprinfo[i].oprcode = atooid(PQgetvalue(res, i, i_oprcode)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(oprinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return oprinfo; |
| } |
| |
| /* |
| * getCollations: |
| * read all collations in the system catalogs and return them in the |
| * CollInfo* structure |
| * |
| * numCollations is set to the number of collations read in |
| */ |
| CollInfo * |
| getCollations(Archive *fout, int *numCollations) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| CollInfo *collinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_collname; |
| int i_collnamespace; |
| int i_collowner; |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * find all collations, including builtin collations; we filter out |
| * system-defined collations at dump-out time. |
| */ |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, collname, " |
| "collnamespace, " |
| "collowner " |
| "FROM pg_collation"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numCollations = ntups; |
| |
| collinfo = (CollInfo *) pg_malloc(ntups * sizeof(CollInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_collname = PQfnumber(res, "collname"); |
| i_collnamespace = PQfnumber(res, "collnamespace"); |
| i_collowner = PQfnumber(res, "collowner"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| collinfo[i].dobj.objType = DO_COLLATION; |
| collinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| collinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&collinfo[i].dobj); |
| collinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_collname)); |
| collinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_collnamespace))); |
| collinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_collowner)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(collinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return collinfo; |
| } |
| |
| /* |
| * getConversions: |
| * read all conversions in the system catalogs and return them in the |
| * ConvInfo* structure |
| * |
| * numConversions is set to the number of conversions read in |
| */ |
| ConvInfo * |
| getConversions(Archive *fout, int *numConversions) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| ConvInfo *convinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_conname; |
| int i_connamespace; |
| int i_conowner; |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * find all conversions, including builtin conversions; we filter out |
| * system-defined conversions at dump-out time. |
| */ |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, conname, " |
| "connamespace, " |
| "conowner " |
| "FROM pg_conversion"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numConversions = ntups; |
| |
| convinfo = (ConvInfo *) pg_malloc(ntups * sizeof(ConvInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_conname = PQfnumber(res, "conname"); |
| i_connamespace = PQfnumber(res, "connamespace"); |
| i_conowner = PQfnumber(res, "conowner"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| convinfo[i].dobj.objType = DO_CONVERSION; |
| convinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| convinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&convinfo[i].dobj); |
| convinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_conname)); |
| convinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_connamespace))); |
| convinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_conowner)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(convinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return convinfo; |
| } |
| |
| /* |
| * getAccessMethods: |
| * read all user-defined access methods in the system catalogs and return |
| * them in the AccessMethodInfo* structure |
| * |
| * numAccessMethods is set to the number of access methods read in |
| */ |
| AccessMethodInfo * |
| getAccessMethods(Archive *fout, int *numAccessMethods) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| AccessMethodInfo *aminfo; |
| int i_tableoid; |
| int i_oid; |
| int i_amname; |
| int i_amhandler; |
| int i_amtype; |
| |
| /* Before 9.6, there are no user-defined access methods */ |
| if (fout->remoteVersion < 90600) |
| { |
| *numAccessMethods = 0; |
| return NULL; |
| } |
| |
| query = createPQExpBuffer(); |
| |
| /* Select all access methods from pg_am table */ |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, amname, amtype, " |
| "amhandler::pg_catalog.regproc AS amhandler " |
| "FROM pg_am"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numAccessMethods = ntups; |
| |
| aminfo = (AccessMethodInfo *) pg_malloc(ntups * sizeof(AccessMethodInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_amname = PQfnumber(res, "amname"); |
| i_amhandler = PQfnumber(res, "amhandler"); |
| i_amtype = PQfnumber(res, "amtype"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| aminfo[i].dobj.objType = DO_ACCESS_METHOD; |
| aminfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| aminfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&aminfo[i].dobj); |
| aminfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_amname)); |
| aminfo[i].dobj.namespace = NULL; |
| aminfo[i].amhandler = pg_strdup(PQgetvalue(res, i, i_amhandler)); |
| aminfo[i].amtype = *(PQgetvalue(res, i, i_amtype)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableAccessMethod(&(aminfo[i]), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return aminfo; |
| } |
| |
| |
| /* |
| * getOpclasses: |
| * read all opclasses in the system catalogs and return them in the |
| * OpclassInfo* structure |
| * |
| * numOpclasses is set to the number of opclasses read in |
| */ |
| OpclassInfo * |
| getOpclasses(Archive *fout, int *numOpclasses) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| OpclassInfo *opcinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_opcname; |
| int i_opcnamespace; |
| int i_opcowner; |
| |
| /* |
| * find all opclasses, including builtin opclasses; we filter out |
| * system-defined opclasses at dump-out time. |
| */ |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, opcname, " |
| "opcnamespace, " |
| "opcowner " |
| "FROM pg_opclass"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numOpclasses = ntups; |
| |
| opcinfo = (OpclassInfo *) pg_malloc(ntups * sizeof(OpclassInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_opcname = PQfnumber(res, "opcname"); |
| i_opcnamespace = PQfnumber(res, "opcnamespace"); |
| i_opcowner = PQfnumber(res, "opcowner"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| opcinfo[i].dobj.objType = DO_OPCLASS; |
| opcinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| opcinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&opcinfo[i].dobj); |
| opcinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_opcname)); |
| opcinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_opcnamespace))); |
| opcinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_opcowner)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(opcinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return opcinfo; |
| } |
| |
| /* |
| * getOpfamilies: |
| * read all opfamilies in the system catalogs and return them in the |
| * OpfamilyInfo* structure |
| * |
| * numOpfamilies is set to the number of opfamilies read in |
| */ |
| OpfamilyInfo * |
| getOpfamilies(Archive *fout, int *numOpfamilies) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| OpfamilyInfo *opfinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_opfname; |
| int i_opfnamespace; |
| int i_opfowner; |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * find all opfamilies, including builtin opfamilies; we filter out |
| * system-defined opfamilies at dump-out time. |
| */ |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, opfname, " |
| "opfnamespace, " |
| "opfowner " |
| "FROM pg_opfamily"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numOpfamilies = ntups; |
| |
| opfinfo = (OpfamilyInfo *) pg_malloc(ntups * sizeof(OpfamilyInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_opfname = PQfnumber(res, "opfname"); |
| i_opfnamespace = PQfnumber(res, "opfnamespace"); |
| i_opfowner = PQfnumber(res, "opfowner"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| opfinfo[i].dobj.objType = DO_OPFAMILY; |
| opfinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| opfinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&opfinfo[i].dobj); |
| opfinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_opfname)); |
| opfinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_opfnamespace))); |
| opfinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_opfowner)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(opfinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return opfinfo; |
| } |
| |
| /* |
| * getAggregates: |
| * read all the user-defined aggregates in the system catalogs and |
| * return them in the AggInfo* structure |
| * |
| * numAggs is set to the number of aggregates read in |
| */ |
| AggInfo * |
| getAggregates(Archive *fout, int *numAggs) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| AggInfo *agginfo; |
| int i_tableoid; |
| int i_oid; |
| int i_aggname; |
| int i_aggnamespace; |
| int i_pronargs; |
| int i_proargtypes; |
| int i_proowner; |
| int i_aggacl; |
| int i_acldefault; |
| |
| /* |
| * Find all interesting aggregates. See comment in getFuncs() for the |
| * rationale behind the filtering logic. |
| */ |
| if (fout->remoteVersion >= 90600) |
| { |
| const char *agg_check; |
| |
| agg_check = (fout->remoteVersion >= 110000 ? "p.prokind = 'a'" |
| : "p.proisagg"); |
| |
| appendPQExpBuffer(query, "SELECT p.tableoid, p.oid, " |
| "p.proname AS aggname, " |
| "p.pronamespace AS aggnamespace, " |
| "p.pronargs, p.proargtypes, " |
| "p.proowner, " |
| "p.proacl AS aggacl, " |
| "acldefault('f', p.proowner) AS acldefault " |
| "FROM pg_proc p " |
| "LEFT JOIN pg_init_privs pip ON " |
| "(p.oid = pip.objoid " |
| "AND pip.classoid = 'pg_proc'::regclass " |
| "AND pip.objsubid = 0) " |
| "WHERE %s AND (" |
| "p.pronamespace != " |
| "(SELECT oid FROM pg_namespace " |
| "WHERE nspname = 'pg_catalog') OR " |
| "p.proacl IS DISTINCT FROM pip.initprivs", |
| agg_check); |
| if (dopt->binary_upgrade) |
| appendPQExpBufferStr(query, |
| " OR EXISTS(SELECT 1 FROM pg_depend WHERE " |
| "classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND " |
| "refclassid = 'pg_extension'::regclass AND " |
| "deptype = 'e')"); |
| appendPQExpBufferChar(query, ')'); |
| <<<<<<< HEAD |
| |
| destroyPQExpBuffer(acl_subquery); |
| destroyPQExpBuffer(racl_subquery); |
| destroyPQExpBuffer(initacl_subquery); |
| destroyPQExpBuffer(initracl_subquery); |
| } |
| else |
| { |
| appendPQExpBuffer(query, "SELECT tableoid, oid, proname AS aggname, " |
| "pronamespace AS aggnamespace, " |
| "pronargs, proargtypes, " |
| "(%s proowner) AS rolname, " |
| "proacl AS aggacl, " |
| "NULL AS raggacl, " |
| "NULL AS initaggacl, NULL AS initraggacl " |
| "FROM pg_proc p " |
| "WHERE proisagg AND (" |
| "pronamespace != " |
| "(SELECT oid FROM pg_namespace " |
| "WHERE nspname = 'pg_catalog')", |
| username_subquery); |
| if (dopt->binary_upgrade && fout->remoteVersion >= 90100) |
| appendPQExpBufferStr(query, |
| " OR EXISTS(SELECT 1 FROM pg_depend WHERE " |
| "classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND " |
| "refclassid = 'pg_extension'::regclass AND " |
| "deptype = 'e')"); |
| appendPQExpBufferChar(query, ')'); |
| } |
| ======= |
| } |
| else |
| { |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, proname AS aggname, " |
| "pronamespace AS aggnamespace, " |
| "pronargs, proargtypes, " |
| "proowner, " |
| "proacl AS aggacl, " |
| "acldefault('f', proowner) AS acldefault " |
| "FROM pg_proc p " |
| "WHERE proisagg AND (" |
| "pronamespace != " |
| "(SELECT oid FROM pg_namespace " |
| "WHERE nspname = 'pg_catalog')"); |
| if (dopt->binary_upgrade) |
| appendPQExpBufferStr(query, |
| " OR EXISTS(SELECT 1 FROM pg_depend WHERE " |
| "classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND " |
| "refclassid = 'pg_extension'::regclass AND " |
| "deptype = 'e')"); |
| appendPQExpBufferChar(query, ')'); |
| } |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numAggs = ntups; |
| |
| agginfo = (AggInfo *) pg_malloc(ntups * sizeof(AggInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_aggname = PQfnumber(res, "aggname"); |
| i_aggnamespace = PQfnumber(res, "aggnamespace"); |
| i_pronargs = PQfnumber(res, "pronargs"); |
| i_proargtypes = PQfnumber(res, "proargtypes"); |
| i_proowner = PQfnumber(res, "proowner"); |
| i_aggacl = PQfnumber(res, "aggacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| agginfo[i].aggfn.dobj.objType = DO_AGG; |
| agginfo[i].aggfn.dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| agginfo[i].aggfn.dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&agginfo[i].aggfn.dobj); |
| agginfo[i].aggfn.dobj.name = pg_strdup(PQgetvalue(res, i, i_aggname)); |
| agginfo[i].aggfn.dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_aggnamespace))); |
| agginfo[i].aggfn.dacl.acl = pg_strdup(PQgetvalue(res, i, i_aggacl)); |
| agginfo[i].aggfn.dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| agginfo[i].aggfn.dacl.privtype = 0; |
| agginfo[i].aggfn.dacl.initprivs = NULL; |
| agginfo[i].aggfn.rolname = getRoleName(PQgetvalue(res, i, i_proowner)); |
| agginfo[i].aggfn.lang = InvalidOid; /* not currently interesting */ |
| agginfo[i].aggfn.prorettype = InvalidOid; /* not saved */ |
| agginfo[i].aggfn.nargs = atoi(PQgetvalue(res, i, i_pronargs)); |
| if (agginfo[i].aggfn.nargs == 0) |
| agginfo[i].aggfn.argtypes = NULL; |
| else |
| { |
| agginfo[i].aggfn.argtypes = (Oid *) pg_malloc(agginfo[i].aggfn.nargs * sizeof(Oid)); |
| parseOidArray(PQgetvalue(res, i, i_proargtypes), |
| agginfo[i].aggfn.argtypes, |
| agginfo[i].aggfn.nargs); |
| } |
| agginfo[i].aggfn.postponed_def = false; /* might get set during sort */ |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(agginfo[i].aggfn.dobj), fout); |
| |
| /* Mark whether aggregate has an ACL */ |
| if (!PQgetisnull(res, i, i_aggacl)) |
| agginfo[i].aggfn.dobj.components |= DUMP_COMPONENT_ACL; |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return agginfo; |
| } |
| |
| /* |
| * getExtProtocols: |
| * read all the user-defined protocols in the system catalogs and |
| * return them in the ExtProtInfo* structure |
| * |
| * numExtProtocols is set to the number of protocols read in |
| */ |
| /* Declared in pg_dump.h */ |
| ExtProtInfo * |
| getExtProtocols(Archive *fout, int *numExtProtocols) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| ExtProtInfo *ptcinfo; |
| int i_oid; |
| int i_tableoid; |
| int i_ptcname; |
| int i_rolname; |
| int i_ptcacl; |
| int i_ptcracl; |
| int i_ptcinitacl; |
| int i_ptcinitracl; |
| int i_ptctrusted; |
| int i_ptcreadid; |
| int i_ptcwriteid; |
| int i_ptcvalidid; |
| |
| /* find all user-defined external protocol */ |
| |
| |
| if (fout->remoteVersion >= 90600) |
| { |
| PQExpBuffer acl_subquery = createPQExpBuffer(); |
| PQExpBuffer racl_subquery = createPQExpBuffer(); |
| PQExpBuffer initacl_subquery = createPQExpBuffer(); |
| PQExpBuffer initracl_subquery = createPQExpBuffer(); |
| |
| buildACLQueries(acl_subquery, racl_subquery, initacl_subquery, |
| initracl_subquery, "ptc.ptcacl", "ptc.ptcowner", "'E'", |
| fout->dopt->binary_upgrade); |
| |
| appendPQExpBuffer(query, "SELECT ptc.tableoid as tableoid, " |
| " ptc.oid as oid, " |
| " ptc.ptcname as ptcname, " |
| " ptcreadfn as ptcreadoid, " |
| " ptcwritefn as ptcwriteoid, " |
| " ptcvalidatorfn as ptcvaloid, " |
| " (%s ptc.ptcowner) as rolname, " |
| " ptc.ptctrusted as ptctrusted, " |
| " %s AS ptcacl, " |
| " %s AS ptcracl, " |
| " %s AS ptcinitacl, " |
| " %s AS ptcinitracl " |
| "FROM pg_extprotocol ptc " |
| "LEFT JOIN pg_init_privs pip ON " |
| " (ptc.oid = pip.objoid " |
| " AND pip.classoid = 'pg_extprotocol'::regclass " |
| " AND pip.objsubid = 0)", |
| username_subquery, |
| acl_subquery->data, |
| racl_subquery->data, |
| initacl_subquery->data, |
| initracl_subquery->data); |
| |
| destroyPQExpBuffer(acl_subquery); |
| destroyPQExpBuffer(racl_subquery); |
| destroyPQExpBuffer(initacl_subquery); |
| destroyPQExpBuffer(initracl_subquery); |
| } |
| else |
| { |
| appendPQExpBuffer(query, "SELECT ptc.tableoid as tableoid, " |
| " ptc.oid as oid, " |
| " ptc.ptcname as ptcname, " |
| " ptcreadfn as ptcreadoid, " |
| " ptcwritefn as ptcwriteoid, " |
| " ptcvalidatorfn as ptcvaloid, " |
| " (%s ptc.ptcowner) as rolname, " |
| " ptc.ptctrusted as ptctrusted, " |
| " ptc.ptcacl as ptcacl, " |
| " NULL as ptcracl, " |
| " NULL as ptcinitacl, " |
| " NULL as ptcinitracl " |
| "FROM pg_extprotocol ptc", |
| username_subquery); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numExtProtocols = ntups; |
| |
| ptcinfo = (ExtProtInfo *) pg_malloc(ntups * sizeof(ExtProtInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_ptcname = PQfnumber(res, "ptcname"); |
| i_rolname = PQfnumber(res, "rolname"); |
| i_ptcacl = PQfnumber(res, "ptcacl"); |
| i_ptcracl = PQfnumber(res, "ptcracl"); |
| i_ptcinitacl = PQfnumber(res, "ptcinitacl"); |
| i_ptcinitracl = PQfnumber(res, "ptcinitracl"); |
| i_ptctrusted = PQfnumber(res, "ptctrusted"); |
| i_ptcreadid = PQfnumber(res, "ptcreadoid"); |
| i_ptcwriteid = PQfnumber(res, "ptcwriteoid"); |
| i_ptcvalidid = PQfnumber(res, "ptcvaloid"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| ptcinfo[i].dobj.objType = DO_EXTPROTOCOL; |
| ptcinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| ptcinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&ptcinfo[i].dobj); |
| ptcinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_ptcname)); |
| ptcinfo[i].dobj.namespace = NULL; |
| ptcinfo[i].ptcowner = pg_strdup(PQgetvalue(res, i, i_rolname)); |
| if (strlen(ptcinfo[i].ptcowner) == 0) |
| pg_log_warning("owner of external protocol \"%s\" appears to be invalid", |
| ptcinfo[i].dobj.name); |
| |
| if (PQgetisnull(res, i, i_ptcreadid)) |
| ptcinfo[i].ptcreadid = InvalidOid; |
| else |
| ptcinfo[i].ptcreadid = atooid(PQgetvalue(res, i, i_ptcreadid)); |
| |
| if (PQgetisnull(res, i, i_ptcwriteid)) |
| ptcinfo[i].ptcwriteid = InvalidOid; |
| else |
| ptcinfo[i].ptcwriteid = atooid(PQgetvalue(res, i, i_ptcwriteid)); |
| |
| if (PQgetisnull(res, i, i_ptcvalidid)) |
| ptcinfo[i].ptcvalidid = InvalidOid; |
| else |
| ptcinfo[i].ptcvalidid = atooid(PQgetvalue(res, i, i_ptcvalidid)); |
| |
| ptcinfo[i].ptcacl = pg_strdup(PQgetvalue(res, i, i_ptcacl)); |
| ptcinfo[i].rproacl = pg_strdup(PQgetvalue(res, i, i_ptcracl)); |
| ptcinfo[i].initproacl = pg_strdup(PQgetvalue(res, i, i_ptcinitacl)); |
| ptcinfo[i].initrproacl = pg_strdup(PQgetvalue(res, i, i_ptcinitracl)); |
| ptcinfo[i].ptctrusted = *(PQgetvalue(res, i, i_ptctrusted)) == 't'; |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(ptcinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return ptcinfo; |
| } |
| |
| /* |
| * getFuncs: |
| * read all the user-defined functions in the system catalogs and |
| * return them in the FuncInfo* structure |
| * |
| * numFuncs is set to the number of functions read in |
| */ |
| FuncInfo * |
| getFuncs(Archive *fout, int *numFuncs) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| FuncInfo *finfo; |
| int i_tableoid; |
| int i_oid; |
| int i_proname; |
| int i_pronamespace; |
| int i_proowner; |
| int i_prolang; |
| int i_pronargs; |
| int i_proargtypes; |
| int i_prorettype; |
| int i_proacl; |
| int i_acldefault; |
| |
| /* |
| * Find all interesting functions. This is a bit complicated: |
| * |
| * 1. Always exclude aggregates; those are handled elsewhere. |
| * |
| * 2. Always exclude functions that are internally dependent on something |
| * else, since presumably those will be created as a result of creating |
| * the something else. This currently acts only to suppress constructor |
| * functions for range types. Note this is OK only because the |
| * constructors don't have any dependencies the range type doesn't have; |
| * otherwise we might not get creation ordering correct. |
| * |
| * 3. Otherwise, we normally exclude functions in pg_catalog. However, if |
| * they're members of extensions and we are in binary-upgrade mode then |
| * include them, since we want to dump extension members individually in |
| * that mode. Also, if they are used by casts or transforms then we need |
| * to gather the information about them, though they won't be dumped if |
| * they are built-in. Also, in 9.6 and up, include functions in |
| * pg_catalog if they have an ACL different from what's shown in |
| * pg_init_privs (so we have to join to pg_init_privs; annoying). |
| */ |
| if (fout->remoteVersion >= 90600) |
| { |
| const char *not_agg_check; |
| |
| not_agg_check = (fout->remoteVersion >= 110000 ? "p.prokind <> 'a'" |
| : "NOT p.proisagg"); |
| |
| appendPQExpBuffer(query, |
| "SELECT p.tableoid, p.oid, p.proname, p.prolang, " |
| "p.pronargs, p.proargtypes, p.prorettype, " |
| "p.proacl, " |
| "acldefault('f', p.proowner) AS acldefault, " |
| "p.pronamespace, " |
| "p.proowner " |
| "FROM pg_proc p " |
| "LEFT JOIN pg_init_privs pip ON " |
| "(p.oid = pip.objoid " |
| "AND pip.classoid = 'pg_proc'::regclass " |
| "AND pip.objsubid = 0) " |
| "WHERE %s" |
| "\n AND NOT EXISTS (SELECT 1 FROM pg_depend " |
| "WHERE classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND deptype = 'i')" |
| "\n AND (" |
| "\n pronamespace != " |
| "(SELECT oid FROM pg_namespace " |
| "WHERE nspname = 'pg_catalog')" |
| "\n OR EXISTS (SELECT 1 FROM pg_cast" |
| "\n WHERE pg_cast.oid > %u " |
| "\n AND p.oid = pg_cast.castfunc)" |
| "\n OR EXISTS (SELECT 1 FROM pg_transform" |
| "\n WHERE pg_transform.oid > %u AND " |
| "\n (p.oid = pg_transform.trffromsql" |
| "\n OR p.oid = pg_transform.trftosql))", |
| not_agg_check, |
| g_last_builtin_oid, |
| g_last_builtin_oid); |
| if (dopt->binary_upgrade) |
| appendPQExpBufferStr(query, |
| "\n OR EXISTS(SELECT 1 FROM pg_depend WHERE " |
| "classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND " |
| "refclassid = 'pg_extension'::regclass AND " |
| "deptype = 'e')"); |
| appendPQExpBufferStr(query, |
| "\n OR p.proacl IS DISTINCT FROM pip.initprivs"); |
| appendPQExpBufferChar(query, ')'); |
| } |
| else |
| { |
| appendPQExpBuffer(query, |
| "SELECT tableoid, oid, proname, prolang, " |
| "pronargs, proargtypes, prorettype, proacl, " |
| "acldefault('f', proowner) AS acldefault, " |
| "pronamespace, " |
| "proowner " |
| "FROM pg_proc p " |
| "WHERE NOT proisagg" |
| "\n AND NOT EXISTS (SELECT 1 FROM pg_depend " |
| "WHERE classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND deptype = 'i')" |
| "\n AND (" |
| "\n pronamespace != " |
| "(SELECT oid FROM pg_namespace " |
| "WHERE nspname = 'pg_catalog')" |
| "\n OR EXISTS (SELECT 1 FROM pg_cast" |
| "\n WHERE pg_cast.oid > '%u'::oid" |
| "\n AND p.oid = pg_cast.castfunc)", |
| g_last_builtin_oid); |
| |
| if (fout->remoteVersion >= 90500) |
| appendPQExpBuffer(query, |
| "\n OR EXISTS (SELECT 1 FROM pg_transform" |
| "\n WHERE pg_transform.oid > '%u'::oid" |
| "\n AND (p.oid = pg_transform.trffromsql" |
| "\n OR p.oid = pg_transform.trftosql))", |
| g_last_builtin_oid); |
| |
| <<<<<<< HEAD |
| /* |
| * GPDB: Much of the extension machinery was backported into GPDB 5 from higher |
| * major versions. So include the clause if we are running against GPDB 5. |
| */ |
| #if 0 |
| if (dopt->binary_upgrade && fout->remoteVersion >= 90100) |
| #endif |
| if (dopt->binary_upgrade && fout->remoteVersion >= 80300) |
| ======= |
| if (dopt->binary_upgrade) |
| >>>>>>> REL_16_9 |
| appendPQExpBufferStr(query, |
| "\n OR EXISTS(SELECT 1 FROM pg_depend WHERE " |
| "classid = 'pg_proc'::regclass AND " |
| "objid = p.oid AND " |
| "refclassid = 'pg_extension'::regclass AND " |
| "deptype = 'e')"); |
| appendPQExpBufferChar(query, ')'); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numFuncs = ntups; |
| |
| finfo = (FuncInfo *) pg_malloc0(ntups * sizeof(FuncInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_proname = PQfnumber(res, "proname"); |
| i_pronamespace = PQfnumber(res, "pronamespace"); |
| i_proowner = PQfnumber(res, "proowner"); |
| i_prolang = PQfnumber(res, "prolang"); |
| i_pronargs = PQfnumber(res, "pronargs"); |
| i_proargtypes = PQfnumber(res, "proargtypes"); |
| i_prorettype = PQfnumber(res, "prorettype"); |
| i_proacl = PQfnumber(res, "proacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| finfo[i].dobj.objType = DO_FUNC; |
| finfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| finfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&finfo[i].dobj); |
| finfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_proname)); |
| finfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_pronamespace))); |
| finfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_proacl)); |
| finfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| finfo[i].dacl.privtype = 0; |
| finfo[i].dacl.initprivs = NULL; |
| finfo[i].rolname = getRoleName(PQgetvalue(res, i, i_proowner)); |
| finfo[i].lang = atooid(PQgetvalue(res, i, i_prolang)); |
| finfo[i].prorettype = atooid(PQgetvalue(res, i, i_prorettype)); |
| finfo[i].nargs = atoi(PQgetvalue(res, i, i_pronargs)); |
| if (finfo[i].nargs == 0) |
| finfo[i].argtypes = NULL; |
| else |
| { |
| finfo[i].argtypes = (Oid *) pg_malloc(finfo[i].nargs * sizeof(Oid)); |
| parseOidArray(PQgetvalue(res, i, i_proargtypes), |
| finfo[i].argtypes, finfo[i].nargs); |
| } |
| finfo[i].postponed_def = false; /* might get set during sort */ |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableFunction(&finfo[i]); |
| selectDumpableObject(&(finfo[i].dobj), fout); |
| |
| /* Mark whether function has an ACL */ |
| if (!PQgetisnull(res, i, i_proacl)) |
| finfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return finfo; |
| } |
| |
| /* |
| * getTables |
| * read all the tables (no indexes) in the system catalogs, |
| * and return them as an array of TableInfo structures |
| * |
| * *numTables is set to the number of tables read in |
| */ |
| TableInfo * |
| getTables(Archive *fout, int *numTables) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| TableInfo *tblinfo; |
| bool lockTableDumped = false; |
| int i_reltableoid; |
| int i_reloid; |
| int i_relname; |
| int i_relnamespace; |
| int i_relkind; |
| <<<<<<< HEAD |
| int i_relstorage; |
| int i_relacl; |
| int i_rrelacl; |
| int i_initrelacl; |
| int i_initrrelacl; |
| int i_rolname; |
| ======= |
| int i_reltype; |
| int i_relowner; |
| >>>>>>> REL_16_9 |
| int i_relchecks; |
| int i_relhasindex; |
| int i_relhasrules; |
| int i_relpages; |
| int i_toastpages; |
| int i_owning_tab; |
| int i_owning_col; |
| int i_reltablespace; |
| int i_relhasoids; |
| int i_relhastriggers; |
| int i_relpersistence; |
| int i_relispopulated; |
| int i_relreplident; |
| int i_relrowsec; |
| int i_relforcerowsec; |
| int i_relfrozenxid; |
| int i_toastfrozenxid; |
| int i_toastoid; |
| int i_relminmxid; |
| int i_toastminmxid; |
| int i_reloptions; |
| int i_checkoption; |
| int i_toastreloptions; |
| int i_reloftype; |
| <<<<<<< HEAD |
| int i_relpages; |
| int i_parrelid; |
| int i_parlevel; |
| ======= |
| >>>>>>> REL_16_9 |
| int i_foreignserver; |
| int i_amname; |
| <<<<<<< HEAD |
| int i_amoid; |
| int i_isivm; |
| int i_isdynamic; |
| ======= |
| int i_is_identity_sequence; |
| int i_relacl; |
| int i_acldefault; |
| int i_ispartition; |
| >>>>>>> REL_16_9 |
| |
| /* |
| * Find all the tables and table-like objects. |
| * |
| * We must fetch all tables in this phase because otherwise we cannot |
| * correctly identify inherited columns, owned sequences, etc. |
| * |
| * We include system catalogs, so that we can work if a user table is |
| * defined to inherit from a system catalog (pretty weird, but...) |
| * |
| * Note: in this phase we should collect only a minimal amount of |
| * information about each table, basically just enough to decide if it is |
| * interesting. In particular, since we do not yet have lock on any user |
| * table, we MUST NOT invoke any server-side data collection functions |
| * (for instance, pg_get_partkeydef()). Those are likely to fail or give |
| * wrong answers if any concurrent DDL is happening. |
| */ |
| |
| appendPQExpBufferStr(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relnamespace, c.relkind, c.reltype, " |
| "c.relowner, " |
| "c.relchecks, " |
| "c.relhasindex, c.relhasrules, c.relpages, " |
| "c.relhastriggers, " |
| "c.relpersistence, " |
| "c.reloftype, " |
| "c.relacl, " |
| "acldefault(CASE WHEN c.relkind = " CppAsString2(RELKIND_SEQUENCE) |
| " THEN 's'::\"char\" ELSE 'r'::\"char\" END, c.relowner) AS acldefault, " |
| "CASE WHEN c.relkind = " CppAsString2(RELKIND_FOREIGN_TABLE) " THEN " |
| "(SELECT ftserver FROM pg_catalog.pg_foreign_table WHERE ftrelid = c.oid) " |
| "ELSE 0 END AS foreignserver, " |
| "c.relfrozenxid, tc.relfrozenxid AS tfrozenxid, " |
| "tc.oid AS toid, " |
| "tc.relpages AS toastpages, " |
| "tc.reloptions AS toast_reloptions, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "tsp.spcname AS reltablespace, "); |
| |
| if (fout->remoteVersion >= 120000) |
| appendPQExpBufferStr(query, |
| "false AS relhasoids, "); |
| else |
| appendPQExpBufferStr(query, |
| "c.relhasoids, "); |
| |
| if (fout->remoteVersion >= 90300) |
| appendPQExpBufferStr(query, |
| "c.relispopulated, "); |
| else |
| appendPQExpBufferStr(query, |
| "'t' as relispopulated, "); |
| |
| if (fout->remoteVersion >= 90400) |
| appendPQExpBufferStr(query, |
| "c.relreplident, "); |
| else |
| appendPQExpBufferStr(query, |
| "'d' AS relreplident, "); |
| |
| if (fout->remoteVersion >= 90500) |
| appendPQExpBufferStr(query, |
| "c.relrowsecurity, c.relforcerowsecurity, "); |
| else |
| appendPQExpBufferStr(query, |
| "false AS relrowsecurity, " |
| "false AS relforcerowsecurity, "); |
| |
| if (fout->remoteVersion >= 90300) |
| appendPQExpBufferStr(query, |
| "c.relminmxid, tc.relminmxid AS tminmxid, "); |
| else |
| appendPQExpBufferStr(query, |
| "0 AS relminmxid, 0 AS tminmxid, "); |
| |
| if (fout->remoteVersion >= 90300) |
| appendPQExpBufferStr(query, |
| "array_remove(array_remove(c.reloptions,'check_option=local'),'check_option=cascaded') AS reloptions, " |
| "CASE WHEN 'check_option=local' = ANY (c.reloptions) THEN 'LOCAL'::text " |
| "WHEN 'check_option=cascaded' = ANY (c.reloptions) THEN 'CASCADED'::text ELSE NULL END AS checkoption, "); |
| else |
| appendPQExpBufferStr(query, |
| "c.reloptions, NULL AS checkoption, "); |
| |
| if (fout->remoteVersion >= 90600) |
| appendPQExpBufferStr(query, |
| "am.amname, "); |
| else |
| appendPQExpBufferStr(query, |
| "NULL AS amname, "); |
| |
| if (fout->remoteVersion >= 90600) |
| appendPQExpBufferStr(query, |
| "(d.deptype = 'i') IS TRUE AS is_identity_sequence, "); |
| else |
| appendPQExpBufferStr(query, |
| "false AS is_identity_sequence, "); |
| |
| if (fout->remoteVersion >= 100000) |
| appendPQExpBufferStr(query, |
| "c.relispartition AS ispartition "); |
| else |
| appendPQExpBufferStr(query, |
| "false AS ispartition "); |
| |
| /* |
| * Left join to pg_depend to pick up dependency info linking sequences to |
| * their owning column, if any (note this dependency is AUTO except for |
| * identity sequences, where it's INTERNAL). Also join to pg_tablespace to |
| * collect the spcname. |
| */ |
| appendPQExpBufferStr(query, |
| "\nFROM pg_class c\n" |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = " CppAsString2(RELKIND_SEQUENCE) " AND " |
| "d.classid = 'pg_class'::regclass AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = 'pg_class'::regclass AND d.deptype IN ('a', 'i'))\n" |
| "LEFT JOIN pg_tablespace tsp ON (tsp.oid = c.reltablespace)\n"); |
| |
| /* |
| * In 9.6 and up, left join to pg_am to pick up the amname. |
| */ |
| if (fout->remoteVersion >= 90600) |
| appendPQExpBufferStr(query, |
| "LEFT JOIN pg_am am ON (c.relam = am.oid)\n"); |
| |
| /* |
| * We purposefully ignore toast OIDs for partitioned tables; the reason is |
| * that versions 10 and 11 have them, but later versions do not, so |
| * emitting them causes the upgrade to fail. |
| */ |
| appendPQExpBufferStr(query, |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid" |
| " AND tc.relkind = " CppAsString2(RELKIND_TOASTVALUE) |
| " AND c.relkind <> " CppAsString2(RELKIND_PARTITIONED_TABLE) ")\n"); |
| |
| /* |
| * Restrict to interesting relkinds (in particular, not indexes). Not all |
| * relkinds are possible in older servers, but it's not worth the trouble |
| * to emit a version-dependent list. |
| * |
| * Composite-type table entries won't be dumped as such, but we have to |
| * make a DumpableObject for them so that we can track dependencies of the |
| * composite type (pg_depend entries for columns of the composite type |
| * link to the pg_class entry not the pg_type entry). |
| */ |
| <<<<<<< HEAD |
| |
| if (fout->remoteVersion >= 90600) |
| { |
| char *partkeydef = "NULL"; |
| char *ispartition = "false"; |
| char *partbound = "NULL"; |
| char *relhasoids = "c.relhasoids"; |
| |
| PQExpBuffer acl_subquery = createPQExpBuffer(); |
| PQExpBuffer racl_subquery = createPQExpBuffer(); |
| PQExpBuffer initacl_subquery = createPQExpBuffer(); |
| PQExpBuffer initracl_subquery = createPQExpBuffer(); |
| |
| PQExpBuffer attacl_subquery = createPQExpBuffer(); |
| PQExpBuffer attracl_subquery = createPQExpBuffer(); |
| PQExpBuffer attinitacl_subquery = createPQExpBuffer(); |
| PQExpBuffer attinitracl_subquery = createPQExpBuffer(); |
| |
| /* |
| * Collect the information about any partitioned tables, which were |
| * added in PG10. |
| */ |
| |
| if (fout->remoteVersion >= 100000) |
| { |
| partkeydef = "pg_get_partkeydef(c.oid)"; |
| ispartition = "c.relispartition"; |
| partbound = "pg_get_expr(c.relpartbound, c.oid)"; |
| } |
| |
| /* In PG12 upwards WITH OIDS does not exist anymore. */ |
| if (fout->remoteVersion >= 120000) |
| relhasoids = "'f'::bool"; |
| |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| * |
| * Left join to detect if any privileges are still as-set-at-init, in |
| * which case we won't dump out ACL commands for those. |
| */ |
| |
| buildACLQueries(acl_subquery, racl_subquery, initacl_subquery, |
| initracl_subquery, "c.relacl", "c.relowner", |
| "CASE WHEN c.relkind = " CppAsString2(RELKIND_SEQUENCE) |
| " THEN 's' ELSE 'r' END::\"char\"", |
| dopt->binary_upgrade); |
| |
| buildACLQueries(attacl_subquery, attracl_subquery, attinitacl_subquery, |
| attinitracl_subquery, "at.attacl", "c.relowner", "'c'", |
| dopt->binary_upgrade); |
| |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "'' AS relstorage, " |
| "%s AS relacl, %s as rrelacl, " |
| "%s AS initrelacl, %s as initrrelacl, " |
| "c.relkind, c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, %s AS relhasoids, " |
| "c.relrowsecurity, c.relforcerowsecurity, " |
| "c.relfrozenxid, c.relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "tc.relminmxid AS tminmxid, " |
| "c.relpersistence, c.relispopulated, " |
| "c.relreplident, c.relpages, am.amname, " |
| "am.oid AS amoid, " |
| "CASE WHEN c.relkind = 'f' THEN " |
| "(SELECT ftserver FROM pg_catalog.pg_foreign_table WHERE ftrelid = c.oid) " |
| "ELSE 0 END AS foreignserver, " |
| "CASE WHEN c.reloftype <> 0 THEN c.reloftype::pg_catalog.regtype ELSE NULL END AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "array_remove(array_remove(c.reloptions,'check_option=local'),'check_option=cascaded') AS reloptions, " |
| "CASE WHEN 'check_option=local' = ANY (c.reloptions) THEN 'LOCAL'::text " |
| "WHEN 'check_option=cascaded' = ANY (c.reloptions) THEN 'CASCADED'::text ELSE NULL END AS checkoption, " |
| "tc.reloptions AS toast_reloptions, " |
| "0 as parrelid, " |
| "0 as parlevel, " |
| "c.relkind = '%c' AND EXISTS (SELECT 1 FROM pg_depend WHERE classid = 'pg_class'::regclass AND objid = c.oid AND objsubid = 0 AND refclassid = 'pg_class'::regclass AND deptype = 'i') AS is_identity_sequence, " |
| "EXISTS (SELECT 1 FROM pg_attribute at LEFT JOIN pg_init_privs pip ON " |
| "(c.oid = pip.objoid " |
| "AND pip.classoid = 'pg_class'::regclass " |
| "AND pip.objsubid = at.attnum)" |
| "WHERE at.attrelid = c.oid AND (" |
| "%s IS NOT NULL " |
| "OR %s IS NOT NULL " |
| "OR %s IS NOT NULL " |
| "OR %s IS NOT NULL" |
| "))" |
| "AS changed_acl, " |
| "%s AS partkeydef, " |
| "%s AS ispartition, " |
| "%s AS partbound, " |
| "c.relisivm AS isivm, " |
| "c.relisdynamic AS isdynamic " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype IN ('a', 'i')) " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid AND c.relkind <> '%c') " |
| "LEFT JOIN pg_am am ON (c.relam = am.oid) " |
| "LEFT JOIN pg_init_privs pip ON " |
| "(c.oid = pip.objoid " |
| "AND pip.classoid = 'pg_class'::regclass " |
| "AND pip.objsubid = 0) " |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "ORDER BY c.oid", |
| acl_subquery->data, |
| racl_subquery->data, |
| initacl_subquery->data, |
| initracl_subquery->data, |
| username_subquery, |
| relhasoids, |
| RELKIND_SEQUENCE, |
| attacl_subquery->data, |
| attracl_subquery->data, |
| attinitacl_subquery->data, |
| attinitracl_subquery->data, |
| partkeydef, |
| ispartition, |
| partbound, |
| RELKIND_SEQUENCE, |
| RELKIND_PARTITIONED_TABLE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE, |
| RELKIND_MATVIEW, RELKIND_FOREIGN_TABLE, |
| RELKIND_PARTITIONED_TABLE); |
| |
| destroyPQExpBuffer(acl_subquery); |
| destroyPQExpBuffer(racl_subquery); |
| destroyPQExpBuffer(initacl_subquery); |
| destroyPQExpBuffer(initracl_subquery); |
| |
| destroyPQExpBuffer(attacl_subquery); |
| destroyPQExpBuffer(attracl_subquery); |
| destroyPQExpBuffer(attinitacl_subquery); |
| destroyPQExpBuffer(attinitracl_subquery); |
| } |
| else if (fout->remoteVersion >= 90500) |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "c.relrowsecurity, c.relforcerowsecurity, " |
| "c.relfrozenxid, c.relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "tc.relminmxid AS tminmxid, " |
| "c.relpersistence, c.relispopulated, " |
| "c.relreplident, c.relpages, " |
| "NULL AS amname, " |
| "CASE WHEN c.relkind = 'f' THEN " |
| "(SELECT ftserver FROM pg_catalog.pg_foreign_table WHERE ftrelid = c.oid) " |
| "ELSE 0 END AS foreignserver, " |
| "CASE WHEN c.reloftype <> 0 THEN c.reloftype::pg_catalog.regtype ELSE NULL END AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "array_remove(array_remove(c.reloptions,'check_option=local'),'check_option=cascaded') AS reloptions, " |
| "CASE WHEN 'check_option=local' = ANY (c.reloptions) THEN 'LOCAL'::text " |
| "WHEN 'check_option=cascaded' = ANY (c.reloptions) THEN 'CASCADED'::text ELSE NULL END AS checkoption, " |
| "tc.reloptions AS toast_reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid) " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0)" |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "AND c.oid NOT IN (SELECT p.parchildrelid FROM pg_partition_rule p LEFT " |
| "JOIN pg_exttable e ON p.parchildrelid=e.reloid WHERE e.reloid IS NULL)" |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE, |
| RELKIND_MATVIEW, RELKIND_FOREIGN_TABLE); |
| } |
| else if (fout->remoteVersion >= 90400) |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relstorage, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "'f'::bool AS relrowsecurity, " |
| "'f'::bool AS relforcerowsecurity, " |
| "c.relfrozenxid, c.relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "tc.relminmxid AS tminmxid, " |
| "c.relpersistence, c.relispopulated, " |
| "c.relreplident, c.relpages, " |
| "NULL AS amname, " |
| "CASE WHEN c.relkind = 'f' THEN " |
| "(SELECT ftserver FROM pg_catalog.pg_foreign_table WHERE ftrelid = c.oid) " |
| "ELSE 0 END AS foreignserver, " |
| "CASE WHEN c.reloftype <> 0 THEN c.reloftype::pg_catalog.regtype ELSE NULL END AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "array_remove(array_remove(c.reloptions,'check_option=local'),'check_option=cascaded') AS reloptions, " |
| "CASE WHEN 'check_option=local' = ANY (c.reloptions) THEN 'LOCAL'::text " |
| "WHEN 'check_option=cascaded' = ANY (c.reloptions) THEN 'CASCADED'::text ELSE NULL END AS checkoption, " |
| "tc.reloptions AS toast_reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid AND c.relstorage <> 'c') " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0)" |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "AND c.oid NOT IN (SELECT p.parchildrelid FROM pg_partition_rule p LEFT " |
| "JOIN pg_exttable e ON p.parchildrelid=e.reloid WHERE e.reloid IS NULL)" |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE, |
| RELKIND_MATVIEW, RELKIND_FOREIGN_TABLE); |
| } |
| else if (fout->remoteVersion >= 90300) |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "'f'::bool AS relrowsecurity, " |
| "'f'::bool AS relforcerowsecurity, " |
| "c.relfrozenxid, c.relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "tc.relminmxid AS tminmxid, " |
| "c.relpersistence, c.relispopulated, " |
| "'d' AS relreplident, c.relpages, " |
| "NULL AS amname, " |
| "CASE WHEN c.relkind = 'f' THEN " |
| "(SELECT ftserver FROM pg_catalog.pg_foreign_table WHERE ftrelid = c.oid) " |
| "ELSE 0 END AS foreignserver, " |
| "CASE WHEN c.reloftype <> 0 THEN c.reloftype::pg_catalog.regtype ELSE NULL END AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "array_remove(array_remove(c.reloptions,'check_option=local'),'check_option=cascaded') AS reloptions, " |
| "CASE WHEN 'check_option=local' = ANY (c.reloptions) THEN 'LOCAL'::text " |
| "WHEN 'check_option=cascaded' = ANY (c.reloptions) THEN 'CASCADED'::text ELSE NULL END AS checkoption, " |
| "tc.reloptions AS toast_reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid) " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0)" |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "AND c.oid NOT IN (SELECT p.parchildrelid FROM pg_partition_rule p LEFT " |
| "JOIN pg_exttable e ON p.parchildrelid=e.reloid WHERE e.reloid IS NULL)" |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE, |
| RELKIND_MATVIEW, RELKIND_FOREIGN_TABLE); |
| } |
| else if (fout->remoteVersion >= 90100) |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relstorage, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "'f'::bool AS relrowsecurity, " |
| "'f'::bool AS relforcerowsecurity, " |
| "c.relfrozenxid, 0 AS relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "0 AS tminmxid, " |
| "c.relpersistence, 't' as relispopulated, " |
| "'d' AS relreplident, c.relpages, " |
| "NULL AS amname, " |
| "CASE WHEN c.relkind = 'f' THEN " |
| "(SELECT ftserver FROM pg_catalog.pg_foreign_table WHERE ftrelid = c.oid) " |
| "ELSE 0 END AS foreignserver, " |
| "CASE WHEN c.reloftype <> 0 THEN c.reloftype::pg_catalog.regtype ELSE NULL END AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "c.reloptions AS reloptions, " |
| "tc.reloptions AS toast_reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid) " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0)" |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "AND c.oid NOT IN (SELECT p.parchildrelid FROM pg_partition_rule p LEFT " |
| "JOIN pg_exttable e ON p.parchildrelid=e.reloid WHERE e.reloid IS NULL)" |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE, |
| RELKIND_MATVIEW, RELKIND_FOREIGN_TABLE); |
| } |
| else if (fout->remoteVersion >= 90000) |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relstorage, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "'f'::bool AS relrowsecurity, " |
| "'f'::bool AS relforcerowsecurity, " |
| "c.relfrozenxid, 0 AS relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "0 AS tminmxid, " |
| "'p' AS relpersistence, 't' as relispopulated, " |
| "'d' AS relreplident, c.relpages, " |
| "NULL AS amname, " |
| "NULL AS foreignserver, " |
| "CASE WHEN c.reloftype <> 0 THEN c.reloftype::pg_catalog.regtype ELSE NULL END AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "c.reloptions AS reloptions, " |
| "tc.reloptions AS toast_reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid) " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0)" |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "AND c.oid NOT IN (SELECT p.parchildrelid FROM pg_partition_rule p LEFT " |
| "JOIN pg_exttable e ON p.parchildrelid=e.reloid WHERE e.reloid IS NULL)" |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE); |
| } |
| else if (fout->remoteVersion >= 80400) |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, c.relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "'f'::bool AS relrowsecurity, " |
| "'f'::bool AS relforcerowsecurity, " |
| "c.relfrozenxid, 0 AS relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "0 AS tminmxid, " |
| "'p' AS relpersistence, 't' as relispopulated, " |
| "'d' AS relreplident, c.relpages, " |
| "NULL AS amname, " |
| "NULL AS foreignserver, " |
| "NULL AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "c.reloptions AS reloptions, " |
| "tc.reloptions AS toast_reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid) " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0)" |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c') " |
| "AND c.relnamespace <> 7012 " /* BM_BITMAPINDEX_NAMESPACE */ |
| "AND c.oid NOT IN (SELECT p.parchildrelid FROM pg_partition_rule p LEFT " |
| "JOIN pg_exttable e ON p.parchildrelid=e.reloid WHERE e.reloid IS NULL)" |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE); |
| } |
| else |
| { |
| /* |
| * Left join to pick up dependency info linking sequences to their |
| * owning column, if any (note this dependency is AUTO as of 8.2) |
| */ |
| appendPQExpBuffer(query, |
| "SELECT c.tableoid, c.oid, c.relname, " |
| "c.relstorage, " |
| "c.relacl, NULL as rrelacl, " |
| "NULL AS initrelacl, NULL AS initrrelacl, " |
| "c.relkind, " |
| "c.relnamespace, " |
| "(%s c.relowner) AS rolname, " |
| "c.relchecks, (c.reltriggers <> 0) AS relhastriggers, " |
| "c.relhasindex, c.relhasrules, c.relhasoids, " |
| "'f'::bool AS relrowsecurity, " |
| "'f'::bool AS relforcerowsecurity, " |
| "c.relfrozenxid, 0 AS relminmxid, tc.oid AS toid, " |
| "tc.relfrozenxid AS tfrozenxid, " |
| "0 AS tminmxid, " |
| "'p' AS relpersistence, 't' as relispopulated, " |
| "'d' AS relreplident, c.relpages, " |
| "NULL AS amname, " |
| "NULL AS foreignserver, " |
| "NULL AS reloftype, " |
| "d.refobjid AS owning_tab, " |
| "d.refobjsubid AS owning_col, " |
| "(SELECT spcname FROM pg_tablespace t WHERE t.oid = c.reltablespace) AS reltablespace, " |
| "c.reloptions AS reloptions, " |
| "p.parrelid as parrelid, " |
| "pl.parlevel as parlevel, " |
| "NULL AS toast_reloptions, " |
| "NULL AS changed_acl, " |
| "NULL AS partkeydef, " |
| "false AS ispartition, " |
| "NULL AS partbound " |
| "FROM pg_class c " |
| "LEFT JOIN pg_depend d ON " |
| "(c.relkind = '%c' AND " |
| "d.classid = c.tableoid AND d.objid = c.oid AND " |
| "d.objsubid = 0 AND " |
| "d.refclassid = c.tableoid AND d.deptype = 'a') " |
| "LEFT JOIN pg_class tc ON (c.reltoastrelid = tc.oid) " |
| "LEFT JOIN pg_partition_rule pr ON c.oid = pr.parchildrelid " |
| "LEFT JOIN pg_partition p ON pr.paroid = p.oid " |
| "LEFT JOIN pg_partition pl ON (c.oid = pl.parrelid AND pl.parlevel = 0) " |
| "WHERE c.relkind in ('%c', '%c', '%c', '%c') " |
| "AND c.oid NOT IN (select p.parchildrelid from pg_partition_rule p " |
| "LEFT JOIN pg_exttable e on p.parchildrelid=e.reloid where e.reloid is null) " |
| "AND c.relnamespace <> 3012 " /* BM_BITMAPINDEX_NAMESPACE in GPDB 5 and below */ |
| "ORDER BY c.oid", |
| username_subquery, |
| RELKIND_SEQUENCE, |
| RELKIND_RELATION, RELKIND_SEQUENCE, |
| RELKIND_VIEW, RELKIND_COMPOSITE_TYPE); |
| } |
| ======= |
| appendPQExpBufferStr(query, |
| "WHERE c.relkind IN (" |
| CppAsString2(RELKIND_RELATION) ", " |
| CppAsString2(RELKIND_SEQUENCE) ", " |
| CppAsString2(RELKIND_VIEW) ", " |
| CppAsString2(RELKIND_COMPOSITE_TYPE) ", " |
| CppAsString2(RELKIND_MATVIEW) ", " |
| CppAsString2(RELKIND_FOREIGN_TABLE) ", " |
| CppAsString2(RELKIND_PARTITIONED_TABLE) ")\n" |
| "ORDER BY c.oid"); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numTables = ntups; |
| |
| /* |
| * Extract data from result and lock dumpable tables. We do the locking |
| * before anything else, to minimize the window wherein a table could |
| * disappear under us. |
| * |
| * Note that we have to save info about all tables here, even when dumping |
| * only one, because we don't yet know which tables might be inheritance |
| * ancestors of the target table. |
| */ |
| tblinfo = (TableInfo *) pg_malloc0(ntups * sizeof(TableInfo)); |
| |
| i_reltableoid = PQfnumber(res, "tableoid"); |
| i_reloid = PQfnumber(res, "oid"); |
| i_relname = PQfnumber(res, "relname"); |
| i_relnamespace = PQfnumber(res, "relnamespace"); |
| i_relkind = PQfnumber(res, "relkind"); |
| <<<<<<< HEAD |
| i_relstorage = PQfnumber(res, "relstorage"); |
| i_rolname = PQfnumber(res, "rolname"); |
| ======= |
| i_reltype = PQfnumber(res, "reltype"); |
| i_relowner = PQfnumber(res, "relowner"); |
| >>>>>>> REL_16_9 |
| i_relchecks = PQfnumber(res, "relchecks"); |
| i_relhasindex = PQfnumber(res, "relhasindex"); |
| i_relhasrules = PQfnumber(res, "relhasrules"); |
| i_relpages = PQfnumber(res, "relpages"); |
| i_toastpages = PQfnumber(res, "toastpages"); |
| i_owning_tab = PQfnumber(res, "owning_tab"); |
| i_owning_col = PQfnumber(res, "owning_col"); |
| i_reltablespace = PQfnumber(res, "reltablespace"); |
| i_relhasoids = PQfnumber(res, "relhasoids"); |
| i_relhastriggers = PQfnumber(res, "relhastriggers"); |
| i_relpersistence = PQfnumber(res, "relpersistence"); |
| i_relispopulated = PQfnumber(res, "relispopulated"); |
| i_relreplident = PQfnumber(res, "relreplident"); |
| i_relrowsec = PQfnumber(res, "relrowsecurity"); |
| i_relforcerowsec = PQfnumber(res, "relforcerowsecurity"); |
| i_relfrozenxid = PQfnumber(res, "relfrozenxid"); |
| i_toastfrozenxid = PQfnumber(res, "tfrozenxid"); |
| i_toastoid = PQfnumber(res, "toid"); |
| i_relminmxid = PQfnumber(res, "relminmxid"); |
| i_toastminmxid = PQfnumber(res, "tminmxid"); |
| i_reloptions = PQfnumber(res, "reloptions"); |
| i_checkoption = PQfnumber(res, "checkoption"); |
| i_toastreloptions = PQfnumber(res, "toast_reloptions"); |
| i_reloftype = PQfnumber(res, "reloftype"); |
| <<<<<<< HEAD |
| i_parrelid = PQfnumber(res, "parrelid"); |
| i_parlevel = PQfnumber(res, "parlevel"); |
| i_is_identity_sequence = PQfnumber(res, "is_identity_sequence"); |
| i_changed_acl = PQfnumber(res, "changed_acl"); |
| i_partkeydef = PQfnumber(res, "partkeydef"); |
| i_ispartition = PQfnumber(res, "ispartition"); |
| i_partbound = PQfnumber(res, "partbound"); |
| i_amname = PQfnumber(res, "amname"); |
| i_amoid = PQfnumber(res, "amoid"); |
| i_isivm = PQfnumber(res, "isivm"); |
| i_isdynamic = PQfnumber(res, "isdynamic"); |
| ======= |
| i_foreignserver = PQfnumber(res, "foreignserver"); |
| i_amname = PQfnumber(res, "amname"); |
| i_is_identity_sequence = PQfnumber(res, "is_identity_sequence"); |
| i_relacl = PQfnumber(res, "relacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| i_ispartition = PQfnumber(res, "ispartition"); |
| >>>>>>> REL_16_9 |
| |
| if (dopt->lockWaitTimeout) |
| { |
| /* |
| * Arrange to fail instead of waiting forever for a table lock. |
| * |
| * NB: this coding assumes that the only queries issued within the |
| * following loop are LOCK TABLEs; else the timeout may be undesirably |
| * applied to other things too. |
| */ |
| resetPQExpBuffer(query); |
| appendPQExpBufferStr(query, "SET statement_timeout = "); |
| appendStringLiteralConn(query, dopt->lockWaitTimeout, GetConnection(fout)); |
| ExecuteSqlStatement(fout, query->data); |
| } |
| |
| resetPQExpBuffer(query); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| tblinfo[i].dobj.objType = DO_TABLE; |
| tblinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_reltableoid)); |
| tblinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_reloid)); |
| AssignDumpId(&tblinfo[i].dobj); |
| tblinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_relname)); |
| tblinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_relnamespace))); |
| tblinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_relacl)); |
| tblinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| tblinfo[i].dacl.privtype = 0; |
| tblinfo[i].dacl.initprivs = NULL; |
| tblinfo[i].relkind = *(PQgetvalue(res, i, i_relkind)); |
| <<<<<<< HEAD |
| tblinfo[i].relstorage = *(PQgetvalue(res, i, i_relstorage)); |
| tblinfo[i].relpersistence = *(PQgetvalue(res, i, i_relpersistence)); |
| ======= |
| tblinfo[i].reltype = atooid(PQgetvalue(res, i, i_reltype)); |
| tblinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_relowner)); |
| tblinfo[i].ncheck = atoi(PQgetvalue(res, i, i_relchecks)); |
| >>>>>>> REL_16_9 |
| tblinfo[i].hasindex = (strcmp(PQgetvalue(res, i, i_relhasindex), "t") == 0); |
| tblinfo[i].hasrules = (strcmp(PQgetvalue(res, i, i_relhasrules), "t") == 0); |
| tblinfo[i].relpages = atoi(PQgetvalue(res, i, i_relpages)); |
| if (PQgetisnull(res, i, i_toastpages)) |
| tblinfo[i].toastpages = 0; |
| else |
| tblinfo[i].toastpages = atoi(PQgetvalue(res, i, i_toastpages)); |
| if (PQgetisnull(res, i, i_owning_tab)) |
| { |
| tblinfo[i].owning_tab = InvalidOid; |
| tblinfo[i].owning_col = 0; |
| } |
| else |
| { |
| tblinfo[i].owning_tab = atooid(PQgetvalue(res, i, i_owning_tab)); |
| tblinfo[i].owning_col = atoi(PQgetvalue(res, i, i_owning_col)); |
| } |
| tblinfo[i].reltablespace = pg_strdup(PQgetvalue(res, i, i_reltablespace)); |
| tblinfo[i].hasoids = (strcmp(PQgetvalue(res, i, i_relhasoids), "t") == 0); |
| tblinfo[i].hastriggers = (strcmp(PQgetvalue(res, i, i_relhastriggers), "t") == 0); |
| tblinfo[i].relpersistence = *(PQgetvalue(res, i, i_relpersistence)); |
| tblinfo[i].relispopulated = (strcmp(PQgetvalue(res, i, i_relispopulated), "t") == 0); |
| tblinfo[i].relreplident = *(PQgetvalue(res, i, i_relreplident)); |
| tblinfo[i].rowsec = (strcmp(PQgetvalue(res, i, i_relrowsec), "t") == 0); |
| tblinfo[i].forcerowsec = (strcmp(PQgetvalue(res, i, i_relforcerowsec), "t") == 0); |
| tblinfo[i].frozenxid = atooid(PQgetvalue(res, i, i_relfrozenxid)); |
| tblinfo[i].toast_frozenxid = atooid(PQgetvalue(res, i, i_toastfrozenxid)); |
| tblinfo[i].toast_oid = atooid(PQgetvalue(res, i, i_toastoid)); |
| tblinfo[i].minmxid = atooid(PQgetvalue(res, i, i_relminmxid)); |
| tblinfo[i].toast_minmxid = atooid(PQgetvalue(res, i, i_toastminmxid)); |
| tblinfo[i].reloptions = pg_strdup(PQgetvalue(res, i, i_reloptions)); |
| if (PQgetisnull(res, i, i_checkoption)) |
| tblinfo[i].checkoption = NULL; |
| else |
| tblinfo[i].checkoption = pg_strdup(PQgetvalue(res, i, i_checkoption)); |
| tblinfo[i].toast_reloptions = pg_strdup(PQgetvalue(res, i, i_toastreloptions)); |
| <<<<<<< HEAD |
| tblinfo[i].parrelid = atooid(PQgetvalue(res, i, i_parrelid)); |
| if (tblinfo[i].parrelid != 0) |
| { |
| /* |
| * Length of tmpStr is bigger than the sum of NAMEDATALEN |
| * and the length of EXT_PARTITION_NAME_POSTFIX |
| */ |
| char tmpStr[500]; |
| snprintf(tmpStr, sizeof(tmpStr), "%s%s", tblinfo[i].dobj.name, EXT_PARTITION_NAME_POSTFIX); |
| tblinfo[i].dobj.name = pg_strdup(tmpStr); |
| } |
| if (PQgetisnull(res, i, i_parlevel) || |
| atoi(PQgetvalue(res, i, i_parlevel)) > 0) |
| tblinfo[i].parparent = false; |
| else |
| tblinfo[i].parparent = true; |
| ======= |
| tblinfo[i].reloftype = atooid(PQgetvalue(res, i, i_reloftype)); |
| tblinfo[i].foreign_server = atooid(PQgetvalue(res, i, i_foreignserver)); |
| >>>>>>> REL_16_9 |
| if (PQgetisnull(res, i, i_amname)) |
| tblinfo[i].amname = NULL; |
| else |
| tblinfo[i].amname = pg_strdup(PQgetvalue(res, i, i_amname)); |
| tblinfo[i].is_identity_sequence = (strcmp(PQgetvalue(res, i, i_is_identity_sequence), "t") == 0); |
| tblinfo[i].ispartition = (strcmp(PQgetvalue(res, i, i_ispartition), "t") == 0); |
| |
| if (PQgetisnull(res, i, i_amoid)) |
| tblinfo[i].amoid = InvalidOid; |
| else |
| tblinfo[i].amoid = atooid(PQgetvalue(res, i, i_amoid)); |
| |
| /* other fields were zeroed above */ |
| |
| /* |
| * Decide whether we want to dump this table. |
| */ |
| if (tblinfo[i].relkind == RELKIND_COMPOSITE_TYPE) |
| tblinfo[i].dobj.dump = DUMP_COMPONENT_NONE; |
| else |
| selectDumpableTable(&tblinfo[i], fout); |
| |
| /* |
| * Now, consider the table "interesting" if we need to dump its |
| * definition or its data. Later on, we'll skip a lot of data |
| * collection for uninteresting tables. |
| * |
| * Note: the "interesting" flag will also be set by flagInhTables for |
| * parents of interesting tables, so that we collect necessary |
| * inheritance info even when the parents are not themselves being |
| * dumped. This is the main reason why we need an "interesting" flag |
| * that's separate from the components-to-dump bitmask. |
| */ |
| tblinfo[i].interesting = (tblinfo[i].dobj.dump & |
| (DUMP_COMPONENT_DEFINITION | |
| DUMP_COMPONENT_DATA)) != 0; |
| |
| tblinfo[i].dummy_view = false; /* might get set during sort */ |
| tblinfo[i].postponed_def = false; /* might get set during sort */ |
| |
| /* Tables have data */ |
| tblinfo[i].dobj.components |= DUMP_COMPONENT_DATA; |
| |
| <<<<<<< HEAD |
| /* Partition key string or NULL */ |
| tblinfo[i].partkeydef = pg_strdup(PQgetvalue(res, i, i_partkeydef)); |
| tblinfo[i].ispartition = (strcmp(PQgetvalue(res, i, i_ispartition), "t") == 0); |
| tblinfo[i].partbound = pg_strdup(PQgetvalue(res, i, i_partbound)); |
| tblinfo[i].isivm = (strcmp(PQgetvalue(res, i, i_isivm), "t") == 0); |
| tblinfo[i].isdynamic = (strcmp(PQgetvalue(res, i, i_isdynamic), "t") == 0); |
| |
| /* foreign server */ |
| tblinfo[i].foreign_server = atooid(PQgetvalue(res, i, i_foreignserver)); |
| ======= |
| /* Mark whether table has an ACL */ |
| if (!PQgetisnull(res, i, i_relacl)) |
| tblinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| tblinfo[i].hascolumnACLs = false; /* may get set later */ |
| >>>>>>> REL_16_9 |
| |
| /* |
| * Read-lock target tables to make sure they aren't DROPPED or altered |
| * in schema before we get around to dumping them. |
| * |
| * Note that we don't explicitly lock parents of the target tables; we |
| * assume our lock on the child is enough to prevent schema |
| * alterations to parent tables. |
| * |
| * NOTE: it'd be kinda nice to lock other relations too, not only |
| * plain or partitioned tables, but the backend doesn't presently |
| * allow that. |
| * |
| * We only need to lock the table for certain components; see |
| * pg_dump.h |
| * |
| * GPDB: Build a single LOCK TABLE statement to lock all interesting tables. |
| * This is more performant than issuing a separate LOCK TABLE statement for each table, |
| * with considerable savings in FE/BE overhead. It does come at the cost of some increased |
| * memory usage in both FE and BE, which we will be able to tolerate. |
| */ |
| <<<<<<< HEAD |
| /* GPDB_14_MERGE_FIXME: GPDB_96_MERGE_FIXME: Is the parrelid check still needed? */ |
| if (tblinfo[i].dobj.dump && |
| (tblinfo[i].relkind == RELKIND_RELATION || |
| tblinfo[i].relkind == RELKIND_PARTITIONED_TABLE) && |
| tblinfo[i].parrelid == 0 && |
| (tblinfo[i].dobj.dump & DUMP_COMPONENTS_REQUIRING_LOCK)) |
| { |
| if (!lockTableDumped) |
| appendPQExpBuffer(query, |
| "LOCK TABLE %s ", |
| fmtQualifiedDumpable(&tblinfo[i])); |
| else |
| appendPQExpBuffer(query, |
| ",%s ", |
| fmtQualifiedDumpable(&tblinfo[i])); |
| lockTableDumped = true; |
| } |
| ======= |
| if ((tblinfo[i].dobj.dump & DUMP_COMPONENTS_REQUIRING_LOCK) && |
| (tblinfo[i].relkind == RELKIND_RELATION || |
| tblinfo[i].relkind == RELKIND_PARTITIONED_TABLE)) |
| { |
| /* |
| * Tables are locked in batches. When dumping from a remote |
| * server this can save a significant amount of time by reducing |
| * the number of round trips. |
| */ |
| if (query->len == 0) |
| appendPQExpBuffer(query, "LOCK TABLE %s", |
| fmtQualifiedDumpable(&tblinfo[i])); |
| else |
| { |
| appendPQExpBuffer(query, ", %s", |
| fmtQualifiedDumpable(&tblinfo[i])); |
| >>>>>>> REL_16_9 |
| |
| /* Arbitrarily end a batch when query length reaches 100K. */ |
| if (query->len >= 100000) |
| { |
| /* Lock another batch of tables. */ |
| appendPQExpBufferStr(query, " IN ACCESS SHARE MODE"); |
| ExecuteSqlStatement(fout, query->data); |
| resetPQExpBuffer(query); |
| } |
| } |
| } |
| } |
| |
| if (query->len != 0) |
| { |
| /* Lock the tables in the last batch. */ |
| appendPQExpBufferStr(query, " IN ACCESS SHARE MODE"); |
| ExecuteSqlStatement(fout, query->data); |
| } |
| /* Are there any tables to lock? */ |
| if (lockTableDumped) |
| { |
| appendPQExpBuffer(query, |
| "IN ACCESS SHARE MODE"); |
| ExecuteSqlStatement(fout, query->data); |
| } |
| |
| if (dopt->lockWaitTimeout) |
| { |
| ExecuteSqlStatement(fout, "SET statement_timeout = 0"); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return tblinfo; |
| } |
| |
| /* |
| * getOwnedSeqs |
| * identify owned sequences and mark them as dumpable if owning table is |
| * |
| * We used to do this in getTables(), but it's better to do it after the |
| * index used by findTableByOid() has been set up. |
| */ |
| void |
| getOwnedSeqs(Archive *fout, TableInfo tblinfo[], int numTables) |
| { |
| int i; |
| |
| /* |
| * Force sequences that are "owned" by table columns to be dumped whenever |
| * their owning table is being dumped. |
| */ |
| for (i = 0; i < numTables; i++) |
| { |
| TableInfo *seqinfo = &tblinfo[i]; |
| TableInfo *owning_tab; |
| |
| if (!OidIsValid(seqinfo->owning_tab)) |
| continue; /* not an owned sequence */ |
| |
| owning_tab = findTableByOid(seqinfo->owning_tab); |
| |
| if (owning_tab == NULL) |
| pg_fatal("failed sanity check, parent table with OID %u of sequence with OID %u not found", |
| seqinfo->owning_tab, seqinfo->dobj.catId.oid); |
| |
| /* |
| * For an identity sequence, dump exactly the same components for the |
| * sequence as for the owning table. This is important because we |
| * treat the identity sequence as an integral part of the table. For |
| * example, there is not any DDL command that allows creation of such |
| * a sequence independently of the table. |
| * |
| * For other owned sequences such as serial sequences, we need to dump |
| * the components that are being dumped for the table and any |
| * components that the sequence is explicitly marked with. |
| * |
| * We can't simply use the set of components which are being dumped |
| * for the table as the table might be in an extension (and only the |
| * non-extension components, eg: ACLs if changed, security labels, and |
| * policies, are being dumped) while the sequence is not (and |
| * therefore the definition and other components should also be |
| * dumped). |
| * |
| * If the sequence is part of the extension then it should be properly |
| * marked by checkExtensionMembership() and this will be a no-op as |
| * the table will be equivalently marked. |
| */ |
| if (seqinfo->is_identity_sequence) |
| seqinfo->dobj.dump = owning_tab->dobj.dump; |
| else |
| seqinfo->dobj.dump |= owning_tab->dobj.dump; |
| |
| /* Make sure that necessary data is available if we're dumping it */ |
| if (seqinfo->dobj.dump != DUMP_COMPONENT_NONE) |
| { |
| seqinfo->interesting = true; |
| owning_tab->interesting = true; |
| } |
| } |
| } |
| |
| /* |
| * getInherits |
| * read all the inheritance information |
| * from the system catalogs return them in the InhInfo* structure |
| * |
| * numInherits is set to the number of pairs read in |
| */ |
| InhInfo * |
| getInherits(Archive *fout, int *numInherits) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| InhInfo *inhinfo; |
| |
| int i_inhrelid; |
| int i_inhparent; |
| |
| /* find all the inheritance information */ |
| appendPQExpBufferStr(query, "SELECT inhrelid, inhparent FROM pg_inherits"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numInherits = ntups; |
| |
| inhinfo = (InhInfo *) pg_malloc(ntups * sizeof(InhInfo)); |
| |
| i_inhrelid = PQfnumber(res, "inhrelid"); |
| i_inhparent = PQfnumber(res, "inhparent"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| inhinfo[i].inhrelid = atooid(PQgetvalue(res, i, i_inhrelid)); |
| inhinfo[i].inhparent = atooid(PQgetvalue(res, i, i_inhparent)); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return inhinfo; |
| } |
| |
| /* |
| * getPartitioningInfo |
| * get information about partitioning |
| * |
| * For the most part, we only collect partitioning info about tables we |
| * intend to dump. However, this function has to consider all partitioned |
| * tables in the database, because we need to know about parents of partitions |
| * we are going to dump even if the parents themselves won't be dumped. |
| * |
| * Specifically, what we need to know is whether each partitioned table |
| * has an "unsafe" partitioning scheme that requires us to force |
| * load-via-partition-root mode for its children. Currently the only case |
| * for which we force that is hash partitioning on enum columns, since the |
| * hash codes depend on enum value OIDs which won't be replicated across |
| * dump-and-reload. There are other cases in which load-via-partition-root |
| * might be necessary, but we expect users to cope with them. |
| */ |
| void |
| getPartitioningInfo(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| int ntups; |
| |
| /* hash partitioning didn't exist before v11 */ |
| if (fout->remoteVersion < 110000) |
| return; |
| /* needn't bother if schema-only dump */ |
| if (fout->dopt->schemaOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * Unsafe partitioning schemes are exactly those for which hash enum_ops |
| * appears among the partition opclasses. We needn't check partstrat. |
| * |
| * Note that this query may well retrieve info about tables we aren't |
| * going to dump and hence have no lock on. That's okay since we need not |
| * invoke any unsafe server-side functions. |
| */ |
| appendPQExpBufferStr(query, |
| "SELECT partrelid FROM pg_partitioned_table WHERE\n" |
| "(SELECT c.oid FROM pg_opclass c JOIN pg_am a " |
| "ON c.opcmethod = a.oid\n" |
| "WHERE opcname = 'enum_ops' " |
| "AND opcnamespace = 'pg_catalog'::regnamespace " |
| "AND amname = 'hash') = ANY(partclass)"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| for (int i = 0; i < ntups; i++) |
| { |
| Oid tabrelid = atooid(PQgetvalue(res, i, 0)); |
| TableInfo *tbinfo; |
| |
| tbinfo = findTableByOid(tabrelid); |
| if (tbinfo == NULL) |
| pg_fatal("failed sanity check, table OID %u appearing in pg_partitioned_table not found", |
| tabrelid); |
| tbinfo->unsafe_partitions = true; |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getIndexes |
| * get information about every index on a dumpable table |
| * |
| * Note: index data is not returned directly to the caller, but it |
| * does get entered into the DumpableObject tables. |
| */ |
| void |
| getIndexes(Archive *fout, TableInfo tblinfo[], int numTables) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PQExpBuffer tbloids = createPQExpBuffer(); |
| PGresult *res; |
| int ntups; |
| int curtblindx; |
| IndxInfo *indxinfo; |
| int i_tableoid, |
| i_oid, |
| i_indrelid, |
| i_indexname, |
| i_parentidx, |
| i_indexdef, |
| i_indnkeyatts, |
| i_indnatts, |
| i_indkey, |
| i_indisclustered, |
| i_indisreplident, |
| i_indnullsnotdistinct, |
| i_contype, |
| i_conname, |
| i_condeferrable, |
| i_condeferred, |
| i_contableoid, |
| i_conoid, |
| i_condef, |
| i_tablespace, |
| i_indreloptions, |
| i_indstatcols, |
| i_indstatvals; |
| |
| /* |
| * We want to perform just one query against pg_index. However, we |
| * mustn't try to select every row of the catalog and then sort it out on |
| * the client side, because some of the server-side functions we need |
| * would be unsafe to apply to tables we don't have lock on. Hence, we |
| * build an array of the OIDs of tables we care about (and now have lock |
| * on!), and use a WHERE clause to constrain which rows are selected. |
| */ |
| appendPQExpBufferChar(tbloids, '{'); |
| for (int i = 0; i < numTables; i++) |
| { |
| TableInfo *tbinfo = &tblinfo[i]; |
| |
| if (!tbinfo->hasindex) |
| continue; |
| |
| /* |
| * We can ignore indexes of uninteresting tables. |
| */ |
| if (!tbinfo->interesting) |
| continue; |
| |
| /* OK, we need info for this table */ |
| if (tbloids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(tbloids, ','); |
| appendPQExpBuffer(tbloids, "%u", tbinfo->dobj.catId.oid); |
| } |
| appendPQExpBufferChar(tbloids, '}'); |
| |
| appendPQExpBufferStr(query, |
| "SELECT t.tableoid, t.oid, i.indrelid, " |
| "t.relname AS indexname, " |
| "pg_catalog.pg_get_indexdef(i.indexrelid) AS indexdef, " |
| "i.indkey, i.indisclustered, " |
| "c.contype, c.conname, " |
| "c.condeferrable, c.condeferred, " |
| "c.tableoid AS contableoid, " |
| "c.oid AS conoid, " |
| "pg_catalog.pg_get_constraintdef(c.oid, false) AS condef, " |
| "(SELECT spcname FROM pg_catalog.pg_tablespace s WHERE s.oid = t.reltablespace) AS tablespace, " |
| "t.reloptions AS indreloptions, "); |
| |
| |
| if (fout->remoteVersion >= 90400) |
| appendPQExpBufferStr(query, |
| "i.indisreplident, "); |
| else |
| appendPQExpBufferStr(query, |
| "false AS indisreplident, "); |
| |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(query, |
| "inh.inhparent AS parentidx, " |
| "i.indnkeyatts AS indnkeyatts, " |
| "i.indnatts AS indnatts, " |
| "(SELECT pg_catalog.array_agg(attnum ORDER BY attnum) " |
| " FROM pg_catalog.pg_attribute " |
| " WHERE attrelid = i.indexrelid AND " |
| " attstattarget >= 0) AS indstatcols, " |
| "(SELECT pg_catalog.array_agg(attstattarget ORDER BY attnum) " |
| " FROM pg_catalog.pg_attribute " |
| " WHERE attrelid = i.indexrelid AND " |
| " attstattarget >= 0) AS indstatvals, "); |
| else |
| appendPQExpBufferStr(query, |
| "0 AS parentidx, " |
| "i.indnatts AS indnkeyatts, " |
| "i.indnatts AS indnatts, " |
| "'' AS indstatcols, " |
| "'' AS indstatvals, "); |
| |
| if (fout->remoteVersion >= 150000) |
| appendPQExpBufferStr(query, |
| "i.indnullsnotdistinct "); |
| else |
| appendPQExpBufferStr(query, |
| "false AS indnullsnotdistinct "); |
| |
| /* |
| * The point of the messy-looking outer join is to find a constraint that |
| * is related by an internal dependency link to the index. If we find one, |
| * create a CONSTRAINT entry linked to the INDEX entry. We assume an |
| * index won't have more than one internal dependency. |
| * |
| * Note: the check on conrelid is redundant, but useful because that |
| * column is indexed while conindid is not. |
| */ |
| if (fout->remoteVersion >= 110000) |
| { |
| appendPQExpBuffer(query, |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_index i ON (src.tbloid = i.indrelid) " |
| "JOIN pg_catalog.pg_class t ON (t.oid = i.indexrelid) " |
| "JOIN pg_catalog.pg_class t2 ON (t2.oid = i.indrelid) " |
| "LEFT JOIN pg_catalog.pg_constraint c " |
| "ON (i.indrelid = c.conrelid AND " |
| "i.indexrelid = c.conindid AND " |
| "c.contype IN ('p','u','x')) " |
| "LEFT JOIN pg_catalog.pg_inherits inh " |
| "ON (inh.inhrelid = indexrelid) " |
| "WHERE (i.indisvalid OR t2.relkind = 'p') " |
| "AND i.indisready " |
| "ORDER BY i.indrelid, indexname", |
| tbloids->data); |
| } |
| else |
| { |
| /* |
| * the test on indisready is necessary in 9.2, and harmless in |
| * earlier/later versions |
| */ |
| appendPQExpBuffer(query, |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_index i ON (src.tbloid = i.indrelid) " |
| "JOIN pg_catalog.pg_class t ON (t.oid = i.indexrelid) " |
| "LEFT JOIN pg_catalog.pg_constraint c " |
| "ON (i.indrelid = c.conrelid AND " |
| "i.indexrelid = c.conindid AND " |
| "c.contype IN ('p','u','x')) " |
| "WHERE i.indisvalid AND i.indisready " |
| "ORDER BY i.indrelid, indexname", |
| tbloids->data); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_indrelid = PQfnumber(res, "indrelid"); |
| i_indexname = PQfnumber(res, "indexname"); |
| i_parentidx = PQfnumber(res, "parentidx"); |
| i_indexdef = PQfnumber(res, "indexdef"); |
| i_indnkeyatts = PQfnumber(res, "indnkeyatts"); |
| i_indnatts = PQfnumber(res, "indnatts"); |
| i_indkey = PQfnumber(res, "indkey"); |
| i_indisclustered = PQfnumber(res, "indisclustered"); |
| i_indisreplident = PQfnumber(res, "indisreplident"); |
| i_indnullsnotdistinct = PQfnumber(res, "indnullsnotdistinct"); |
| i_contype = PQfnumber(res, "contype"); |
| i_conname = PQfnumber(res, "conname"); |
| i_condeferrable = PQfnumber(res, "condeferrable"); |
| i_condeferred = PQfnumber(res, "condeferred"); |
| i_contableoid = PQfnumber(res, "contableoid"); |
| i_conoid = PQfnumber(res, "conoid"); |
| i_condef = PQfnumber(res, "condef"); |
| i_tablespace = PQfnumber(res, "tablespace"); |
| i_indreloptions = PQfnumber(res, "indreloptions"); |
| i_indstatcols = PQfnumber(res, "indstatcols"); |
| i_indstatvals = PQfnumber(res, "indstatvals"); |
| |
| indxinfo = (IndxInfo *) pg_malloc(ntups * sizeof(IndxInfo)); |
| |
| /* |
| * Outer loop iterates once per table, not once per row. Incrementing of |
| * j is handled by the inner loop. |
| */ |
| curtblindx = -1; |
| for (int j = 0; j < ntups;) |
| { |
| Oid indrelid = atooid(PQgetvalue(res, j, i_indrelid)); |
| TableInfo *tbinfo = NULL; |
| int numinds; |
| |
| /* Count rows for this table */ |
| for (numinds = 1; numinds < ntups - j; numinds++) |
| if (atooid(PQgetvalue(res, j + numinds, i_indrelid)) != indrelid) |
| break; |
| |
| /* |
| * Locate the associated TableInfo; we rely on tblinfo[] being in OID |
| * order. |
| */ |
| while (++curtblindx < numTables) |
| { |
| <<<<<<< HEAD |
| appendPQExpBuffer(query, |
| "SELECT t.tableoid, t.oid, " |
| "t.relname AS indexname, " |
| "inh.inhparent AS parentidx, " |
| "pg_catalog.pg_get_indexdef(i.indexrelid) AS indexdef, " |
| "i.indnkeyatts AS indnkeyatts, " |
| "i.indnatts AS indnatts, " |
| "i.indkey, i.indisclustered, " |
| "i.indisreplident, " |
| "c.contype, c.conname, " |
| "c.condeferrable, c.condeferred, " |
| "c.tableoid AS contableoid, " |
| "c.oid AS conoid, " |
| "pg_catalog.pg_get_constraintdef(c.oid, false) AS condef, " |
| "(SELECT spcname FROM pg_catalog.pg_tablespace s WHERE s.oid = t.reltablespace) AS tablespace, " |
| "t.reloptions AS indreloptions, " |
| "(SELECT pg_catalog.array_agg(attnum ORDER BY attnum) " |
| " FROM pg_catalog.pg_attribute " |
| " WHERE attrelid = i.indexrelid AND " |
| " attstattarget >= 0) AS indstatcols," |
| "(SELECT pg_catalog.array_agg(attstattarget ORDER BY attnum) " |
| " FROM pg_catalog.pg_attribute " |
| " WHERE attrelid = i.indexrelid AND " |
| " attstattarget >= 0) AS indstatvals " |
| "FROM pg_catalog.pg_index i " |
| "JOIN pg_catalog.pg_class t ON (t.oid = i.indexrelid) " |
| "JOIN pg_catalog.pg_class t2 ON (t2.oid = i.indrelid) " |
| "LEFT JOIN pg_catalog.pg_constraint c " |
| "ON (i.indrelid = c.conrelid AND " |
| "i.indexrelid = c.conindid AND " |
| "c.contype IN ('p','u','x')) " |
| "LEFT JOIN pg_catalog.pg_inherits inh " |
| "ON (inh.inhrelid = indexrelid) " |
| "WHERE i.indrelid = '%u'::pg_catalog.oid " |
| "AND (i.indisvalid OR t2.relkind = 'p') " |
| "AND i.indisready " |
| "ORDER BY indexname", |
| tbinfo->dobj.catId.oid); |
| } |
| else if (fout->remoteVersion >= 90400) |
| { |
| /* |
| * the test on indisready is necessary in 9.2, and harmless in |
| * earlier/later versions |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tableoid, t.oid, " |
| "t.relname AS indexname, " |
| "0 AS parentidx, " |
| "pg_catalog.pg_get_indexdef(i.indexrelid) AS indexdef, " |
| "i.indnatts AS indnkeyatts, " |
| "i.indnatts AS indnatts, " |
| "i.indkey, i.indisclustered, " |
| "i.indisreplident, " |
| "c.contype, c.conname, " |
| "c.condeferrable, c.condeferred, " |
| "c.tableoid AS contableoid, " |
| "c.oid AS conoid, " |
| "pg_catalog.pg_get_constraintdef(c.oid, false) AS condef, " |
| "(SELECT spcname FROM pg_catalog.pg_tablespace s WHERE s.oid = t.reltablespace) AS tablespace, " |
| "t.reloptions AS indreloptions, " |
| "'' AS indstatcols, " |
| "'' AS indstatvals " |
| "FROM pg_catalog.pg_index i " |
| "JOIN pg_catalog.pg_class t ON (t.oid = i.indexrelid) " |
| "LEFT JOIN pg_catalog.pg_constraint c " |
| "ON (i.indrelid = c.conrelid AND " |
| "i.indexrelid = c.conindid AND " |
| "c.contype IN ('p','u','x')) " |
| "WHERE i.indrelid = '%u'::pg_catalog.oid " |
| "AND i.indisvalid AND i.indisready " |
| "ORDER BY indexname", |
| tbinfo->dobj.catId.oid); |
| } |
| else if (fout->remoteVersion >= 90000) |
| { |
| /* |
| * the test on indisready is necessary in 9.2, and harmless in |
| * earlier/later versions |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tableoid, t.oid, " |
| "t.relname AS indexname, " |
| "0 AS parentidx, " |
| "pg_catalog.pg_get_indexdef(i.indexrelid) AS indexdef, " |
| "i.indnatts AS indnkeyatts, " |
| "i.indnatts AS indnatts, " |
| "i.indkey, i.indisclustered, " |
| "false AS indisreplident, " |
| "c.contype, c.conname, " |
| "c.condeferrable, c.condeferred, " |
| "c.tableoid AS contableoid, " |
| "c.oid AS conoid, " |
| "pg_catalog.pg_get_constraintdef(c.oid, false) AS condef, " |
| "(SELECT spcname FROM pg_catalog.pg_tablespace s WHERE s.oid = t.reltablespace) AS tablespace, " |
| "t.reloptions AS indreloptions, " |
| "'' AS indstatcols, " |
| "'' AS indstatvals " |
| "FROM pg_catalog.pg_index i " |
| "JOIN pg_catalog.pg_class t ON (t.oid = i.indexrelid) " |
| "LEFT JOIN pg_catalog.pg_constraint c " |
| "ON (i.indrelid = c.conrelid AND " |
| "i.indexrelid = c.conindid AND " |
| "c.contype IN ('p','u','x')) " |
| "WHERE i.indrelid = '%u'::pg_catalog.oid " |
| "AND i.indisvalid AND i.indisready " |
| "ORDER BY indexname", |
| tbinfo->dobj.catId.oid); |
| } |
| else |
| { |
| appendPQExpBuffer(query, |
| "SELECT t.tableoid, t.oid, " |
| "t.relname AS indexname, " |
| "0 AS parentidx, " |
| "pg_catalog.pg_get_indexdef(i.indexrelid) AS indexdef, " |
| "i.indnatts AS indnkeyatts, " |
| "i.indnatts AS indnatts, " |
| "i.indkey, i.indisclustered, " |
| "false AS indisreplident, " |
| "c.contype, c.conname, " |
| "c.condeferrable, c.condeferred, " |
| "c.tableoid AS contableoid, " |
| "c.oid AS conoid, " |
| "null AS condef, " |
| "(SELECT spcname FROM pg_catalog.pg_tablespace s WHERE s.oid = t.reltablespace) AS tablespace, " |
| "t.reloptions AS indreloptions, " |
| "'' AS indstatcols, " |
| "'' AS indstatvals " |
| "FROM pg_catalog.pg_index i " |
| "JOIN pg_catalog.pg_class t ON (t.oid = i.indexrelid) " |
| "LEFT JOIN pg_catalog.pg_depend d " |
| "ON (d.classid = t.tableoid " |
| "AND d.objid = t.oid " |
| "AND d.deptype = 'i') " |
| "LEFT JOIN pg_catalog.pg_constraint c " |
| "ON (d.refclassid = c.tableoid " |
| "AND d.refobjid = c.oid) " |
| "WHERE i.indrelid = '%u'::pg_catalog.oid " |
| "AND i.indisvalid " |
| "ORDER BY indexname", |
| tbinfo->dobj.catId.oid); |
| } |
| ======= |
| tbinfo = &tblinfo[curtblindx]; |
| if (tbinfo->dobj.catId.oid == indrelid) |
| break; |
| } |
| if (curtblindx >= numTables) |
| pg_fatal("unrecognized table OID %u", indrelid); |
| /* cross-check that we only got requested tables */ |
| if (!tbinfo->hasindex || |
| !tbinfo->interesting) |
| pg_fatal("unexpected index data for table \"%s\"", |
| tbinfo->dobj.name); |
| >>>>>>> REL_16_9 |
| |
| /* Save data for this table */ |
| tbinfo->indexes = indxinfo + j; |
| tbinfo->numIndexes = numinds; |
| |
| <<<<<<< HEAD |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_indexname = PQfnumber(res, "indexname"); |
| i_parentidx = PQfnumber(res, "parentidx"); |
| i_indexdef = PQfnumber(res, "indexdef"); |
| i_indnkeyatts = PQfnumber(res, "indnkeyatts"); |
| i_indnatts = PQfnumber(res, "indnatts"); |
| i_indkey = PQfnumber(res, "indkey"); |
| i_indisclustered = PQfnumber(res, "indisclustered"); |
| i_indisreplident = PQfnumber(res, "indisreplident"); |
| i_contype = PQfnumber(res, "contype"); |
| i_conname = PQfnumber(res, "conname"); |
| i_condeferrable = PQfnumber(res, "condeferrable"); |
| i_condeferred = PQfnumber(res, "condeferred"); |
| i_contableoid = PQfnumber(res, "contableoid"); |
| i_conoid = PQfnumber(res, "conoid"); |
| i_condef = PQfnumber(res, "condef"); |
| i_tablespace = PQfnumber(res, "tablespace"); |
| i_indreloptions = PQfnumber(res, "indreloptions"); |
| i_indstatcols = PQfnumber(res, "indstatcols"); |
| i_indstatvals = PQfnumber(res, "indstatvals"); |
| |
| tbinfo->indexes = indxinfo = |
| (IndxInfo *) pg_malloc(ntups * sizeof(IndxInfo)); |
| tbinfo->numIndexes = ntups; |
| |
| for (j = 0; j < ntups; j++) |
| ======= |
| for (int c = 0; c < numinds; c++, j++) |
| >>>>>>> REL_16_9 |
| { |
| char contype; |
| |
| indxinfo[j].dobj.objType = DO_INDEX; |
| indxinfo[j].dobj.catId.tableoid = atooid(PQgetvalue(res, j, i_tableoid)); |
| indxinfo[j].dobj.catId.oid = atooid(PQgetvalue(res, j, i_oid)); |
| AssignDumpId(&indxinfo[j].dobj); |
| indxinfo[j].dobj.dump = tbinfo->dobj.dump; |
| indxinfo[j].dobj.name = pg_strdup(PQgetvalue(res, j, i_indexname)); |
| indxinfo[j].dobj.namespace = tbinfo->dobj.namespace; |
| indxinfo[j].indextable = tbinfo; |
| indxinfo[j].indexdef = pg_strdup(PQgetvalue(res, j, i_indexdef)); |
| indxinfo[j].indnkeyattrs = atoi(PQgetvalue(res, j, i_indnkeyatts)); |
| indxinfo[j].indnattrs = atoi(PQgetvalue(res, j, i_indnatts)); |
| indxinfo[j].tablespace = pg_strdup(PQgetvalue(res, j, i_tablespace)); |
| indxinfo[j].indreloptions = pg_strdup(PQgetvalue(res, j, i_indreloptions)); |
| indxinfo[j].indstatcols = pg_strdup(PQgetvalue(res, j, i_indstatcols)); |
| indxinfo[j].indstatvals = pg_strdup(PQgetvalue(res, j, i_indstatvals)); |
| indxinfo[j].indkeys = (Oid *) pg_malloc(indxinfo[j].indnattrs * sizeof(Oid)); |
| parseOidArray(PQgetvalue(res, j, i_indkey), |
| indxinfo[j].indkeys, indxinfo[j].indnattrs); |
| indxinfo[j].indisclustered = (PQgetvalue(res, j, i_indisclustered)[0] == 't'); |
| indxinfo[j].indisreplident = (PQgetvalue(res, j, i_indisreplident)[0] == 't'); |
| indxinfo[j].indnullsnotdistinct = (PQgetvalue(res, j, i_indnullsnotdistinct)[0] == 't'); |
| indxinfo[j].parentidx = atooid(PQgetvalue(res, j, i_parentidx)); |
| indxinfo[j].partattaches = (SimplePtrList) |
| { |
| NULL, NULL |
| }; |
| contype = *(PQgetvalue(res, j, i_contype)); |
| |
| if (contype == 'p' || contype == 'u' || contype == 'x') |
| { |
| /* |
| * If we found a constraint matching the index, create an |
| * entry for it. |
| */ |
| ConstraintInfo *constrinfo; |
| |
| constrinfo = (ConstraintInfo *) pg_malloc(sizeof(ConstraintInfo)); |
| constrinfo->dobj.objType = DO_CONSTRAINT; |
| constrinfo->dobj.catId.tableoid = atooid(PQgetvalue(res, j, i_contableoid)); |
| constrinfo->dobj.catId.oid = atooid(PQgetvalue(res, j, i_conoid)); |
| AssignDumpId(&constrinfo->dobj); |
| constrinfo->dobj.dump = tbinfo->dobj.dump; |
| constrinfo->dobj.name = pg_strdup(PQgetvalue(res, j, i_conname)); |
| constrinfo->dobj.namespace = tbinfo->dobj.namespace; |
| constrinfo->contable = tbinfo; |
| constrinfo->condomain = NULL; |
| constrinfo->contype = contype; |
| if (contype == 'x') |
| constrinfo->condef = pg_strdup(PQgetvalue(res, j, i_condef)); |
| else |
| constrinfo->condef = NULL; |
| constrinfo->confrelid = InvalidOid; |
| constrinfo->conindex = indxinfo[j].dobj.dumpId; |
| constrinfo->condeferrable = *(PQgetvalue(res, j, i_condeferrable)) == 't'; |
| constrinfo->condeferred = *(PQgetvalue(res, j, i_condeferred)) == 't'; |
| constrinfo->conislocal = true; |
| constrinfo->separate = true; |
| |
| indxinfo[j].indexconstraint = constrinfo->dobj.dumpId; |
| } |
| else |
| { |
| /* Plain secondary index */ |
| indxinfo[j].indexconstraint = 0; |
| } |
| } |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(tbloids); |
| } |
| |
| /* |
| * getExtendedStatistics |
| * get information about extended-statistics objects. |
| * |
| * Note: extended statistics data is not returned directly to the caller, but |
| * it does get entered into the DumpableObject tables. |
| */ |
| void |
| getExtendedStatistics(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| StatsExtInfo *statsextinfo; |
| int ntups; |
| int i_tableoid; |
| int i_oid; |
| int i_stxname; |
| int i_stxnamespace; |
| int i_stxowner; |
| int i_stxrelid; |
| int i_stattarget; |
| int i; |
| |
| /* Extended statistics were new in v10 */ |
| if (fout->remoteVersion < 100000) |
| return; |
| |
| query = createPQExpBuffer(); |
| |
| if (fout->remoteVersion < 130000) |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, stxname, " |
| "stxnamespace, stxowner, stxrelid, (-1) AS stxstattarget " |
| "FROM pg_catalog.pg_statistic_ext"); |
| else |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, stxname, " |
| "stxnamespace, stxowner, stxrelid, stxstattarget " |
| "FROM pg_catalog.pg_statistic_ext"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_stxname = PQfnumber(res, "stxname"); |
| i_stxnamespace = PQfnumber(res, "stxnamespace"); |
| i_stxowner = PQfnumber(res, "stxowner"); |
| i_stxrelid = PQfnumber(res, "stxrelid"); |
| i_stattarget = PQfnumber(res, "stxstattarget"); |
| |
| statsextinfo = (StatsExtInfo *) pg_malloc(ntups * sizeof(StatsExtInfo)); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| statsextinfo[i].dobj.objType = DO_STATSEXT; |
| statsextinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| statsextinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&statsextinfo[i].dobj); |
| statsextinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_stxname)); |
| statsextinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_stxnamespace))); |
| statsextinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_stxowner)); |
| statsextinfo[i].stattable = |
| findTableByOid(atooid(PQgetvalue(res, i, i_stxrelid))); |
| statsextinfo[i].stattarget = atoi(PQgetvalue(res, i, i_stattarget)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableStatisticsObject(&(statsextinfo[i]), fout); |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getConstraints |
| * |
| * Get info about constraints on dumpable tables. |
| * |
| * Currently handles foreign keys only. |
| * Unique and primary key constraints are handled with indexes, |
| * while check constraints are processed in getTableAttrs(). |
| */ |
| void |
| getConstraints(Archive *fout, TableInfo tblinfo[], int numTables) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PQExpBuffer tbloids = createPQExpBuffer(); |
| PGresult *res; |
| int ntups; |
| int curtblindx; |
| TableInfo *tbinfo = NULL; |
| ConstraintInfo *constrinfo; |
| int i_contableoid, |
| i_conoid, |
| i_conrelid, |
| i_conname, |
| i_confrelid, |
| i_conindid, |
| i_condef; |
| |
| /* |
| * We want to perform just one query against pg_constraint. However, we |
| * mustn't try to select every row of the catalog and then sort it out on |
| * the client side, because some of the server-side functions we need |
| * would be unsafe to apply to tables we don't have lock on. Hence, we |
| * build an array of the OIDs of tables we care about (and now have lock |
| * on!), and use a WHERE clause to constrain which rows are selected. |
| */ |
| appendPQExpBufferChar(tbloids, '{'); |
| for (int i = 0; i < numTables; i++) |
| { |
| TableInfo *tinfo = &tblinfo[i]; |
| |
| /* |
| * For partitioned tables, foreign keys have no triggers so they must |
| * be included anyway in case some foreign keys are defined. |
| */ |
| if ((!tinfo->hastriggers && |
| tinfo->relkind != RELKIND_PARTITIONED_TABLE) || |
| !(tinfo->dobj.dump & DUMP_COMPONENT_DEFINITION)) |
| continue; |
| |
| /* OK, we need info for this table */ |
| if (tbloids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(tbloids, ','); |
| appendPQExpBuffer(tbloids, "%u", tinfo->dobj.catId.oid); |
| } |
| appendPQExpBufferChar(tbloids, '}'); |
| |
| appendPQExpBufferStr(query, |
| "SELECT c.tableoid, c.oid, " |
| "conrelid, conname, confrelid, "); |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(query, "conindid, "); |
| else |
| appendPQExpBufferStr(query, "0 AS conindid, "); |
| appendPQExpBuffer(query, |
| "pg_catalog.pg_get_constraintdef(c.oid) AS condef\n" |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_constraint c ON (src.tbloid = c.conrelid)\n" |
| "WHERE contype = 'f' ", |
| tbloids->data); |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(query, |
| "AND conparentid = 0 "); |
| appendPQExpBufferStr(query, |
| "ORDER BY conrelid, conname"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_contableoid = PQfnumber(res, "tableoid"); |
| i_conoid = PQfnumber(res, "oid"); |
| i_conrelid = PQfnumber(res, "conrelid"); |
| i_conname = PQfnumber(res, "conname"); |
| i_confrelid = PQfnumber(res, "confrelid"); |
| i_conindid = PQfnumber(res, "conindid"); |
| i_condef = PQfnumber(res, "condef"); |
| |
| constrinfo = (ConstraintInfo *) pg_malloc(ntups * sizeof(ConstraintInfo)); |
| |
| curtblindx = -1; |
| for (int j = 0; j < ntups; j++) |
| { |
| Oid conrelid = atooid(PQgetvalue(res, j, i_conrelid)); |
| TableInfo *reftable; |
| |
| /* |
| * Locate the associated TableInfo; we rely on tblinfo[] being in OID |
| * order. |
| */ |
| if (tbinfo == NULL || tbinfo->dobj.catId.oid != conrelid) |
| { |
| while (++curtblindx < numTables) |
| { |
| tbinfo = &tblinfo[curtblindx]; |
| if (tbinfo->dobj.catId.oid == conrelid) |
| break; |
| } |
| if (curtblindx >= numTables) |
| pg_fatal("unrecognized table OID %u", conrelid); |
| } |
| |
| constrinfo[j].dobj.objType = DO_FK_CONSTRAINT; |
| constrinfo[j].dobj.catId.tableoid = atooid(PQgetvalue(res, j, i_contableoid)); |
| constrinfo[j].dobj.catId.oid = atooid(PQgetvalue(res, j, i_conoid)); |
| AssignDumpId(&constrinfo[j].dobj); |
| constrinfo[j].dobj.name = pg_strdup(PQgetvalue(res, j, i_conname)); |
| constrinfo[j].dobj.namespace = tbinfo->dobj.namespace; |
| constrinfo[j].contable = tbinfo; |
| constrinfo[j].condomain = NULL; |
| constrinfo[j].contype = 'f'; |
| constrinfo[j].condef = pg_strdup(PQgetvalue(res, j, i_condef)); |
| constrinfo[j].confrelid = atooid(PQgetvalue(res, j, i_confrelid)); |
| constrinfo[j].conindex = 0; |
| constrinfo[j].condeferrable = false; |
| constrinfo[j].condeferred = false; |
| constrinfo[j].conislocal = true; |
| constrinfo[j].separate = true; |
| |
| /* |
| * Restoring an FK that points to a partitioned table requires that |
| * all partition indexes have been attached beforehand. Ensure that |
| * happens by making the constraint depend on each index partition |
| * attach object. |
| */ |
| reftable = findTableByOid(constrinfo[j].confrelid); |
| if (reftable && reftable->relkind == RELKIND_PARTITIONED_TABLE) |
| { |
| Oid indexOid = atooid(PQgetvalue(res, j, i_conindid)); |
| |
| if (indexOid != InvalidOid) |
| { |
| for (int k = 0; k < reftable->numIndexes; k++) |
| { |
| IndxInfo *refidx; |
| |
| /* not our index? */ |
| if (reftable->indexes[k].dobj.catId.oid != indexOid) |
| continue; |
| |
| refidx = &reftable->indexes[k]; |
| addConstrChildIdxDeps(&constrinfo[j].dobj, refidx); |
| break; |
| } |
| } |
| } |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(tbloids); |
| } |
| |
| /* |
| * addConstrChildIdxDeps |
| * |
| * Recursive subroutine for getConstraints |
| * |
| * Given an object representing a foreign key constraint and an index on the |
| * partitioned table it references, mark the constraint object as dependent |
| * on the DO_INDEX_ATTACH object of each index partition, recursively |
| * drilling down to their partitions if any. This ensures that the FK is not |
| * restored until the index is fully marked valid. |
| */ |
| static void |
| addConstrChildIdxDeps(DumpableObject *dobj, const IndxInfo *refidx) |
| { |
| SimplePtrListCell *cell; |
| |
| Assert(dobj->objType == DO_FK_CONSTRAINT); |
| |
| for (cell = refidx->partattaches.head; cell; cell = cell->next) |
| { |
| IndexAttachInfo *attach = (IndexAttachInfo *) cell->ptr; |
| |
| addObjectDependency(dobj, attach->dobj.dumpId); |
| |
| if (attach->partitionIdx->partattaches.head != NULL) |
| addConstrChildIdxDeps(dobj, attach->partitionIdx); |
| } |
| } |
| |
| /* |
| * getDomainConstraints |
| * |
| * Get info about constraints on a domain. |
| */ |
| static void |
| getDomainConstraints(Archive *fout, TypeInfo *tyinfo) |
| { |
| int i; |
| ConstraintInfo *constrinfo; |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| int i_tableoid, |
| i_oid, |
| i_conname, |
| i_consrc; |
| int ntups; |
| |
| if (!fout->is_prepared[PREPQUERY_GETDOMAINCONSTRAINTS]) |
| { |
| /* Set up query for constraint-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE getDomainConstraints(pg_catalog.oid) AS\n" |
| "SELECT tableoid, oid, conname, " |
| "pg_catalog.pg_get_constraintdef(oid) AS consrc, " |
| "convalidated " |
| "FROM pg_catalog.pg_constraint " |
| "WHERE contypid = $1 " |
| "ORDER BY conname"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_GETDOMAINCONSTRAINTS] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE getDomainConstraints('%u')", |
| tyinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_conname = PQfnumber(res, "conname"); |
| i_consrc = PQfnumber(res, "consrc"); |
| |
| constrinfo = (ConstraintInfo *) pg_malloc(ntups * sizeof(ConstraintInfo)); |
| |
| tyinfo->nDomChecks = ntups; |
| tyinfo->domChecks = constrinfo; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| bool validated = PQgetvalue(res, i, 4)[0] == 't'; |
| |
| constrinfo[i].dobj.objType = DO_CONSTRAINT; |
| constrinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| constrinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&constrinfo[i].dobj); |
| constrinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_conname)); |
| constrinfo[i].dobj.namespace = tyinfo->dobj.namespace; |
| constrinfo[i].contable = NULL; |
| constrinfo[i].condomain = tyinfo; |
| constrinfo[i].contype = 'c'; |
| constrinfo[i].condef = pg_strdup(PQgetvalue(res, i, i_consrc)); |
| constrinfo[i].confrelid = InvalidOid; |
| constrinfo[i].conindex = 0; |
| constrinfo[i].condeferrable = false; |
| constrinfo[i].condeferred = false; |
| constrinfo[i].conislocal = true; |
| |
| constrinfo[i].separate = !validated; |
| |
| /* |
| * Make the domain depend on the constraint, ensuring it won't be |
| * output till any constraint dependencies are OK. If the constraint |
| * has not been validated, it's going to be dumped after the domain |
| * anyway, so this doesn't matter. |
| */ |
| if (validated) |
| addObjectDependency(&tyinfo->dobj, |
| constrinfo[i].dobj.dumpId); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getRules |
| * get basic information about every rule in the system |
| * |
| * numRules is set to the number of rules read in |
| */ |
| RuleInfo * |
| getRules(Archive *fout, int *numRules) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| RuleInfo *ruleinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_rulename; |
| int i_ruletable; |
| int i_ev_type; |
| int i_is_instead; |
| int i_ev_enabled; |
| |
| appendPQExpBufferStr(query, "SELECT " |
| "tableoid, oid, rulename, " |
| "ev_class AS ruletable, ev_type, is_instead, " |
| "ev_enabled " |
| "FROM pg_rewrite " |
| "ORDER BY oid"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numRules = ntups; |
| |
| ruleinfo = (RuleInfo *) pg_malloc(ntups * sizeof(RuleInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_rulename = PQfnumber(res, "rulename"); |
| i_ruletable = PQfnumber(res, "ruletable"); |
| i_ev_type = PQfnumber(res, "ev_type"); |
| i_is_instead = PQfnumber(res, "is_instead"); |
| i_ev_enabled = PQfnumber(res, "ev_enabled"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| Oid ruletableoid; |
| |
| ruleinfo[i].dobj.objType = DO_RULE; |
| ruleinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| ruleinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&ruleinfo[i].dobj); |
| ruleinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_rulename)); |
| ruletableoid = atooid(PQgetvalue(res, i, i_ruletable)); |
| ruleinfo[i].ruletable = findTableByOid(ruletableoid); |
| if (ruleinfo[i].ruletable == NULL) |
| pg_fatal("failed sanity check, parent table with OID %u of pg_rewrite entry with OID %u not found", |
| ruletableoid, ruleinfo[i].dobj.catId.oid); |
| ruleinfo[i].dobj.namespace = ruleinfo[i].ruletable->dobj.namespace; |
| ruleinfo[i].dobj.dump = ruleinfo[i].ruletable->dobj.dump; |
| ruleinfo[i].ev_type = *(PQgetvalue(res, i, i_ev_type)); |
| ruleinfo[i].is_instead = *(PQgetvalue(res, i, i_is_instead)) == 't'; |
| ruleinfo[i].ev_enabled = *(PQgetvalue(res, i, i_ev_enabled)); |
| if (ruleinfo[i].ruletable) |
| { |
| /* |
| * If the table is a view or materialized view, force its ON |
| * SELECT rule to be sorted before the view itself --- this |
| * ensures that any dependencies for the rule affect the table's |
| * positioning. Other rules are forced to appear after their |
| * table. |
| */ |
| if ((ruleinfo[i].ruletable->relkind == RELKIND_VIEW || |
| ruleinfo[i].ruletable->relkind == RELKIND_MATVIEW) && |
| ruleinfo[i].ev_type == '1' && ruleinfo[i].is_instead) |
| { |
| addObjectDependency(&ruleinfo[i].ruletable->dobj, |
| ruleinfo[i].dobj.dumpId); |
| /* We'll merge the rule into CREATE VIEW, if possible */ |
| ruleinfo[i].separate = false; |
| } |
| else |
| { |
| addObjectDependency(&ruleinfo[i].dobj, |
| ruleinfo[i].ruletable->dobj.dumpId); |
| ruleinfo[i].separate = true; |
| } |
| } |
| else |
| ruleinfo[i].separate = true; |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return ruleinfo; |
| } |
| |
| /* |
| * getTriggers |
| * get information about every trigger on a dumpable table |
| * |
| * Note: trigger data is not returned directly to the caller, but it |
| * does get entered into the DumpableObject tables. |
| */ |
| void |
| getTriggers(Archive *fout, TableInfo tblinfo[], int numTables) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PQExpBuffer tbloids = createPQExpBuffer(); |
| PGresult *res; |
| int ntups; |
| int curtblindx; |
| TriggerInfo *tginfo; |
| int i_tableoid, |
| i_oid, |
| i_tgrelid, |
| i_tgname, |
| i_tgfname, |
| i_tgtype, |
| i_tgnargs, |
| i_tgargs, |
| i_tgisconstraint, |
| i_tgconstrname, |
| i_tgconstrrelid, |
| i_tgconstrrelname, |
| i_tgenabled, |
| <<<<<<< HEAD |
| i_tgisinternal, |
| ======= |
| i_tgispartition, |
| >>>>>>> REL_16_9 |
| i_tgdeferrable, |
| i_tginitdeferred, |
| i_tgdef; |
| |
| /* |
| * We want to perform just one query against pg_trigger. However, we |
| * mustn't try to select every row of the catalog and then sort it out on |
| * the client side, because some of the server-side functions we need |
| * would be unsafe to apply to tables we don't have lock on. Hence, we |
| * build an array of the OIDs of tables we care about (and now have lock |
| * on!), and use a WHERE clause to constrain which rows are selected. |
| */ |
| appendPQExpBufferChar(tbloids, '{'); |
| for (int i = 0; i < numTables; i++) |
| { |
| TableInfo *tbinfo = &tblinfo[i]; |
| |
| if (!tbinfo->hastriggers || |
| !(tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION)) |
| continue; |
| |
| /* OK, we need info for this table */ |
| if (tbloids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(tbloids, ','); |
| appendPQExpBuffer(tbloids, "%u", tbinfo->dobj.catId.oid); |
| } |
| appendPQExpBufferChar(tbloids, '}'); |
| |
| <<<<<<< HEAD |
| resetPQExpBuffer(query); |
| if (fout->remoteVersion >= 130000) |
| { |
| /* |
| * NB: think not to use pretty=true in pg_get_triggerdef. It |
| * could result in non-forward-compatible dumps of WHEN clauses |
| * due to under-parenthesization. |
| * |
| * NB: We need to see tgisinternal triggers in partitions, in case |
| * the tgenabled flag has been changed from the parent. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, t.tableoid, t.oid, t.tgisinternal " |
| "FROM pg_catalog.pg_trigger t " |
| "LEFT JOIN pg_catalog.pg_trigger u ON u.oid = t.tgparentid " |
| "WHERE t.tgrelid = '%u'::pg_catalog.oid " |
| "AND (NOT t.tgisinternal OR t.tgenabled != u.tgenabled)", |
| tbinfo->dobj.catId.oid); |
| } |
| else if (fout->remoteVersion >= 110000) |
| { |
| /* |
| * NB: We need to see tgisinternal triggers in partitions, in case |
| * the tgenabled flag has been changed from the parent. No |
| * tgparentid in version 11-12, so we have to match them via |
| * pg_depend. |
| * |
| * See above about pretty=true in pg_get_triggerdef. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, t.tableoid, t.oid, t.tgisinternal " |
| "FROM pg_catalog.pg_trigger t " |
| "LEFT JOIN pg_catalog.pg_depend AS d ON " |
| " d.classid = 'pg_catalog.pg_trigger'::pg_catalog.regclass AND " |
| " d.refclassid = 'pg_catalog.pg_trigger'::pg_catalog.regclass AND " |
| " d.objid = t.oid " |
| "LEFT JOIN pg_catalog.pg_trigger AS pt ON pt.oid = refobjid " |
| "WHERE t.tgrelid = '%u'::pg_catalog.oid " |
| "AND (NOT t.tgisinternal%s)", |
| tbinfo->dobj.catId.oid, |
| tbinfo->ispartition ? |
| " OR t.tgenabled != pt.tgenabled" : ""); |
| } |
| else if (fout->remoteVersion >= 90000) |
| { |
| /* See above about pretty=true in pg_get_triggerdef */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, false as tgisinternal, " |
| "t.tableoid, t.oid " |
| "FROM pg_catalog.pg_trigger t " |
| "WHERE tgrelid = '%u'::pg_catalog.oid " |
| "AND NOT tgisinternal", |
| tbinfo->dobj.catId.oid); |
| } |
| else if (fout->remoteVersion >= 80300) |
| { |
| /* |
| * We ignore triggers that are tied to a foreign-key constraint |
| */ |
| appendPQExpBuffer(query, |
| "SELECT tgname, " |
| "tgfoid::pg_catalog.regproc AS tgfname, " |
| "tgtype, tgnargs, tgargs, tgenabled, " |
| "false as tgisinternal, " |
| "tgisconstraint, tgconstrname, tgdeferrable, " |
| "tgconstrrelid, tginitdeferred, tableoid, oid, " |
| "tgconstrrelid::pg_catalog.regclass AS tgconstrrelname " |
| "FROM pg_catalog.pg_trigger t " |
| "WHERE tgrelid = '%u'::pg_catalog.oid " |
| "AND tgconstraint = 0", |
| tbinfo->dobj.catId.oid); |
| } |
| else |
| { |
| /* |
| * We ignore triggers that are tied to a foreign-key constraint, |
| * but in these versions we have to grovel through pg_constraint |
| * to find out |
| */ |
| appendPQExpBuffer(query, |
| "SELECT tgname, " |
| "tgfoid::pg_catalog.regproc AS tgfname, " |
| "tgtype, tgnargs, tgargs, tgenabled, " |
| "false as tgisinternal, " |
| "tgisconstraint, tgconstrname, tgdeferrable, " |
| "tgconstrrelid, tginitdeferred, tableoid, oid, " |
| "tgconstrrelid::pg_catalog.regclass AS tgconstrrelname " |
| "FROM pg_catalog.pg_trigger t " |
| "WHERE tgrelid = '%u'::pg_catalog.oid " |
| "AND (NOT tgisconstraint " |
| " OR NOT EXISTS" |
| " (SELECT 1 FROM pg_catalog.pg_depend d " |
| " JOIN pg_catalog.pg_constraint c ON (d.refclassid = c.tableoid AND d.refobjid = c.oid) " |
| " WHERE d.classid = t.tableoid AND d.objid = t.oid AND d.deptype = 'i' AND c.contype = 'f'))", |
| tbinfo->dobj.catId.oid); |
| ======= |
| if (fout->remoteVersion >= 150000) |
| { |
| /* |
| * NB: think not to use pretty=true in pg_get_triggerdef. It could |
| * result in non-forward-compatible dumps of WHEN clauses due to |
| * under-parenthesization. |
| * |
| * NB: We need to see partition triggers in case the tgenabled flag |
| * has been changed from the parent. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgrelid, t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, t.tableoid, t.oid, " |
| "t.tgparentid <> 0 AS tgispartition\n" |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_trigger t ON (src.tbloid = t.tgrelid) " |
| "LEFT JOIN pg_catalog.pg_trigger u ON (u.oid = t.tgparentid) " |
| "WHERE ((NOT t.tgisinternal AND t.tgparentid = 0) " |
| "OR t.tgenabled != u.tgenabled) " |
| "ORDER BY t.tgrelid, t.tgname", |
| tbloids->data); |
| } |
| else if (fout->remoteVersion >= 130000) |
| { |
| /* |
| * NB: think not to use pretty=true in pg_get_triggerdef. It could |
| * result in non-forward-compatible dumps of WHEN clauses due to |
| * under-parenthesization. |
| * |
| * NB: We need to see tgisinternal triggers in partitions, in case the |
| * tgenabled flag has been changed from the parent. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgrelid, t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, t.tableoid, t.oid, t.tgisinternal as tgispartition\n" |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_trigger t ON (src.tbloid = t.tgrelid) " |
| "LEFT JOIN pg_catalog.pg_trigger u ON (u.oid = t.tgparentid) " |
| "WHERE (NOT t.tgisinternal OR t.tgenabled != u.tgenabled) " |
| "ORDER BY t.tgrelid, t.tgname", |
| tbloids->data); |
| } |
| else if (fout->remoteVersion >= 110000) |
| { |
| /* |
| * NB: We need to see tgisinternal triggers in partitions, in case the |
| * tgenabled flag has been changed from the parent. No tgparentid in |
| * version 11-12, so we have to match them via pg_depend. |
| * |
| * See above about pretty=true in pg_get_triggerdef. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgrelid, t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, t.tableoid, t.oid, t.tgisinternal as tgispartition " |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_trigger t ON (src.tbloid = t.tgrelid) " |
| "LEFT JOIN pg_catalog.pg_depend AS d ON " |
| " d.classid = 'pg_catalog.pg_trigger'::pg_catalog.regclass AND " |
| " d.refclassid = 'pg_catalog.pg_trigger'::pg_catalog.regclass AND " |
| " d.objid = t.oid " |
| "LEFT JOIN pg_catalog.pg_trigger AS pt ON pt.oid = refobjid " |
| "WHERE (NOT t.tgisinternal OR t.tgenabled != pt.tgenabled) " |
| "ORDER BY t.tgrelid, t.tgname", |
| tbloids->data); |
| } |
| else |
| { |
| /* See above about pretty=true in pg_get_triggerdef */ |
| appendPQExpBuffer(query, |
| "SELECT t.tgrelid, t.tgname, " |
| "t.tgfoid::pg_catalog.regproc AS tgfname, " |
| "pg_catalog.pg_get_triggerdef(t.oid, false) AS tgdef, " |
| "t.tgenabled, false as tgispartition, " |
| "t.tableoid, t.oid " |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_trigger t ON (src.tbloid = t.tgrelid) " |
| "WHERE NOT tgisinternal " |
| "ORDER BY t.tgrelid, t.tgname", |
| tbloids->data); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_tgrelid = PQfnumber(res, "tgrelid"); |
| i_tgname = PQfnumber(res, "tgname"); |
| i_tgfname = PQfnumber(res, "tgfname"); |
| i_tgtype = PQfnumber(res, "tgtype"); |
| i_tgnargs = PQfnumber(res, "tgnargs"); |
| i_tgargs = PQfnumber(res, "tgargs"); |
| i_tgisconstraint = PQfnumber(res, "tgisconstraint"); |
| i_tgconstrname = PQfnumber(res, "tgconstrname"); |
| i_tgconstrrelid = PQfnumber(res, "tgconstrrelid"); |
| i_tgconstrrelname = PQfnumber(res, "tgconstrrelname"); |
| i_tgenabled = PQfnumber(res, "tgenabled"); |
| i_tgispartition = PQfnumber(res, "tgispartition"); |
| i_tgdeferrable = PQfnumber(res, "tgdeferrable"); |
| i_tginitdeferred = PQfnumber(res, "tginitdeferred"); |
| i_tgdef = PQfnumber(res, "tgdef"); |
| |
| tginfo = (TriggerInfo *) pg_malloc(ntups * sizeof(TriggerInfo)); |
| |
| /* |
| * Outer loop iterates once per table, not once per row. Incrementing of |
| * j is handled by the inner loop. |
| */ |
| curtblindx = -1; |
| for (int j = 0; j < ntups;) |
| { |
| Oid tgrelid = atooid(PQgetvalue(res, j, i_tgrelid)); |
| TableInfo *tbinfo = NULL; |
| int numtrigs; |
| |
| /* Count rows for this table */ |
| for (numtrigs = 1; numtrigs < ntups - j; numtrigs++) |
| if (atooid(PQgetvalue(res, j + numtrigs, i_tgrelid)) != tgrelid) |
| break; |
| |
| /* |
| * Locate the associated TableInfo; we rely on tblinfo[] being in OID |
| * order. |
| */ |
| while (++curtblindx < numTables) |
| { |
| tbinfo = &tblinfo[curtblindx]; |
| if (tbinfo->dobj.catId.oid == tgrelid) |
| break; |
| >>>>>>> REL_16_9 |
| } |
| if (curtblindx >= numTables) |
| pg_fatal("unrecognized table OID %u", tgrelid); |
| |
| /* Save data for this table */ |
| tbinfo->triggers = tginfo + j; |
| tbinfo->numTriggers = numtrigs; |
| |
| <<<<<<< HEAD |
| ntups = PQntuples(res); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_tgname = PQfnumber(res, "tgname"); |
| i_tgfname = PQfnumber(res, "tgfname"); |
| i_tgtype = PQfnumber(res, "tgtype"); |
| i_tgnargs = PQfnumber(res, "tgnargs"); |
| i_tgargs = PQfnumber(res, "tgargs"); |
| i_tgisconstraint = PQfnumber(res, "tgisconstraint"); |
| i_tgconstrname = PQfnumber(res, "tgconstrname"); |
| i_tgconstrrelid = PQfnumber(res, "tgconstrrelid"); |
| i_tgconstrrelname = PQfnumber(res, "tgconstrrelname"); |
| i_tgenabled = PQfnumber(res, "tgenabled"); |
| i_tgisinternal = PQfnumber(res, "tgisinternal"); |
| i_tgdeferrable = PQfnumber(res, "tgdeferrable"); |
| i_tginitdeferred = PQfnumber(res, "tginitdeferred"); |
| i_tgdef = PQfnumber(res, "tgdef"); |
| |
| tginfo = (TriggerInfo *) pg_malloc(ntups * sizeof(TriggerInfo)); |
| |
| tbinfo->numTriggers = ntups; |
| tbinfo->triggers = tginfo; |
| |
| for (j = 0; j < ntups; j++) |
| ======= |
| for (int c = 0; c < numtrigs; c++, j++) |
| >>>>>>> REL_16_9 |
| { |
| tginfo[j].dobj.objType = DO_TRIGGER; |
| tginfo[j].dobj.catId.tableoid = atooid(PQgetvalue(res, j, i_tableoid)); |
| tginfo[j].dobj.catId.oid = atooid(PQgetvalue(res, j, i_oid)); |
| AssignDumpId(&tginfo[j].dobj); |
| tginfo[j].dobj.name = pg_strdup(PQgetvalue(res, j, i_tgname)); |
| tginfo[j].dobj.namespace = tbinfo->dobj.namespace; |
| tginfo[j].tgtable = tbinfo; |
| tginfo[j].tgenabled = *(PQgetvalue(res, j, i_tgenabled)); |
| <<<<<<< HEAD |
| tginfo[j].tgisinternal = *(PQgetvalue(res, j, i_tgisinternal)) == 't'; |
| ======= |
| tginfo[j].tgispartition = *(PQgetvalue(res, j, i_tgispartition)) == 't'; |
| >>>>>>> REL_16_9 |
| if (i_tgdef >= 0) |
| { |
| tginfo[j].tgdef = pg_strdup(PQgetvalue(res, j, i_tgdef)); |
| |
| /* remaining fields are not valid if we have tgdef */ |
| tginfo[j].tgfname = NULL; |
| tginfo[j].tgtype = 0; |
| tginfo[j].tgnargs = 0; |
| tginfo[j].tgargs = NULL; |
| tginfo[j].tgisconstraint = false; |
| tginfo[j].tgdeferrable = false; |
| tginfo[j].tginitdeferred = false; |
| tginfo[j].tgconstrname = NULL; |
| tginfo[j].tgconstrrelid = InvalidOid; |
| tginfo[j].tgconstrrelname = NULL; |
| } |
| else |
| { |
| tginfo[j].tgdef = NULL; |
| |
| tginfo[j].tgfname = pg_strdup(PQgetvalue(res, j, i_tgfname)); |
| tginfo[j].tgtype = atoi(PQgetvalue(res, j, i_tgtype)); |
| tginfo[j].tgnargs = atoi(PQgetvalue(res, j, i_tgnargs)); |
| tginfo[j].tgargs = pg_strdup(PQgetvalue(res, j, i_tgargs)); |
| tginfo[j].tgisconstraint = *(PQgetvalue(res, j, i_tgisconstraint)) == 't'; |
| tginfo[j].tgdeferrable = *(PQgetvalue(res, j, i_tgdeferrable)) == 't'; |
| tginfo[j].tginitdeferred = *(PQgetvalue(res, j, i_tginitdeferred)) == 't'; |
| |
| if (tginfo[j].tgisconstraint) |
| { |
| tginfo[j].tgconstrname = pg_strdup(PQgetvalue(res, j, i_tgconstrname)); |
| tginfo[j].tgconstrrelid = atooid(PQgetvalue(res, j, i_tgconstrrelid)); |
| if (OidIsValid(tginfo[j].tgconstrrelid)) |
| { |
| if (PQgetisnull(res, j, i_tgconstrrelname)) |
| pg_fatal("query produced null referenced table name for foreign key trigger \"%s\" on table \"%s\" (OID of table: %u)", |
| tginfo[j].dobj.name, |
| tbinfo->dobj.name, |
| tginfo[j].tgconstrrelid); |
| tginfo[j].tgconstrrelname = pg_strdup(PQgetvalue(res, j, i_tgconstrrelname)); |
| } |
| else |
| tginfo[j].tgconstrrelname = NULL; |
| } |
| else |
| { |
| tginfo[j].tgconstrname = NULL; |
| tginfo[j].tgconstrrelid = InvalidOid; |
| tginfo[j].tgconstrrelname = NULL; |
| } |
| } |
| } |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(tbloids); |
| } |
| |
| /* |
| * getEventTriggers |
| * get information about event triggers |
| */ |
| EventTriggerInfo * |
| getEventTriggers(Archive *fout, int *numEventTriggers) |
| { |
| int i; |
| PQExpBuffer query; |
| PGresult *res; |
| EventTriggerInfo *evtinfo; |
| int i_tableoid, |
| i_oid, |
| i_evtname, |
| i_evtevent, |
| i_evtowner, |
| i_evttags, |
| i_evtfname, |
| i_evtenabled; |
| int ntups; |
| |
| /* Before 9.3, there are no event triggers */ |
| if (fout->remoteVersion < 90300) |
| { |
| *numEventTriggers = 0; |
| return NULL; |
| } |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, |
| "SELECT e.tableoid, e.oid, evtname, evtenabled, " |
| "evtevent, evtowner, " |
| "array_to_string(array(" |
| "select quote_literal(x) " |
| " from unnest(evttags) as t(x)), ', ') as evttags, " |
| "e.evtfoid::regproc as evtfname " |
| "FROM pg_event_trigger e " |
| "ORDER BY e.oid"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numEventTriggers = ntups; |
| |
| evtinfo = (EventTriggerInfo *) pg_malloc(ntups * sizeof(EventTriggerInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_evtname = PQfnumber(res, "evtname"); |
| i_evtevent = PQfnumber(res, "evtevent"); |
| i_evtowner = PQfnumber(res, "evtowner"); |
| i_evttags = PQfnumber(res, "evttags"); |
| i_evtfname = PQfnumber(res, "evtfname"); |
| i_evtenabled = PQfnumber(res, "evtenabled"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| evtinfo[i].dobj.objType = DO_EVENT_TRIGGER; |
| evtinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| evtinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&evtinfo[i].dobj); |
| evtinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_evtname)); |
| evtinfo[i].evtname = pg_strdup(PQgetvalue(res, i, i_evtname)); |
| evtinfo[i].evtevent = pg_strdup(PQgetvalue(res, i, i_evtevent)); |
| evtinfo[i].evtowner = getRoleName(PQgetvalue(res, i, i_evtowner)); |
| evtinfo[i].evttags = pg_strdup(PQgetvalue(res, i, i_evttags)); |
| evtinfo[i].evtfname = pg_strdup(PQgetvalue(res, i, i_evtfname)); |
| evtinfo[i].evtenabled = *(PQgetvalue(res, i, i_evtenabled)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(evtinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return evtinfo; |
| } |
| |
| /* |
| * getProcLangs |
| * get basic information about every procedural language in the system |
| * |
| * numProcLangs is set to the number of langs read in |
| * |
| * NB: this must run after getFuncs() because we assume we can do |
| * findFuncByOid(). |
| */ |
| ProcLangInfo * |
| getProcLangs(Archive *fout, int *numProcLangs) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| ProcLangInfo *planginfo; |
| int i_tableoid; |
| int i_oid; |
| int i_lanname; |
| int i_lanpltrusted; |
| int i_lanplcallfoid; |
| int i_laninline; |
| int i_lanvalidator; |
| int i_lanacl; |
| int i_acldefault; |
| int i_lanowner; |
| |
| <<<<<<< HEAD |
| /* |
| * The laninline column was added in upstream 90000 but was backported to |
| * Cloudberry 5, so the check needs to go further back than 90000. |
| */ |
| |
| if (fout->remoteVersion >= 90600) |
| { |
| PQExpBuffer acl_subquery = createPQExpBuffer(); |
| PQExpBuffer racl_subquery = createPQExpBuffer(); |
| PQExpBuffer initacl_subquery = createPQExpBuffer(); |
| PQExpBuffer initracl_subquery = createPQExpBuffer(); |
| |
| buildACLQueries(acl_subquery, racl_subquery, initacl_subquery, |
| initracl_subquery, "l.lanacl", "l.lanowner", "'l'", |
| dopt->binary_upgrade); |
| |
| /* pg_language has a laninline column */ |
| appendPQExpBuffer(query, "SELECT l.tableoid, l.oid, " |
| "l.lanname, l.lanpltrusted, l.lanplcallfoid, " |
| "l.laninline, l.lanvalidator, " |
| "%s AS lanacl, " |
| "%s AS rlanacl, " |
| "%s AS initlanacl, " |
| "%s AS initrlanacl, " |
| "(%s l.lanowner) AS lanowner " |
| "FROM pg_language l " |
| "LEFT JOIN pg_init_privs pip ON " |
| "(l.oid = pip.objoid " |
| "AND pip.classoid = 'pg_language'::regclass " |
| "AND pip.objsubid = 0) " |
| "WHERE l.lanispl " |
| "ORDER BY l.oid", |
| acl_subquery->data, |
| racl_subquery->data, |
| initacl_subquery->data, |
| initracl_subquery->data, |
| username_subquery); |
| |
| destroyPQExpBuffer(acl_subquery); |
| destroyPQExpBuffer(racl_subquery); |
| destroyPQExpBuffer(initacl_subquery); |
| destroyPQExpBuffer(initracl_subquery); |
| } |
| else if (fout->remoteVersion >= 90000) |
| { |
| /* pg_language has a laninline column */ |
| appendPQExpBuffer(query, "SELECT tableoid, oid, " |
| "lanname, lanpltrusted, lanplcallfoid, " |
| "laninline, lanvalidator, lanacl, NULL AS rlanacl, " |
| "NULL AS initlanacl, NULL AS initrlanacl, " |
| "(%s lanowner) AS lanowner " |
| "FROM pg_language " |
| "WHERE lanispl " |
| "ORDER BY oid", |
| username_subquery); |
| } |
| else if (fout->remoteVersion >= 80300) |
| { |
| /* pg_language has a laninline column */ |
| /* pg_language has a lanowner column */ |
| appendPQExpBuffer(query, "SELECT tableoid, oid, " |
| "lanname, lanpltrusted, lanplcallfoid, " |
| "laninline, lanvalidator, lanacl, " |
| "0 AS laninline, lanvalidator, lanacl, " |
| "NULL AS rlanacl, " |
| "NULL AS initlanacl, NULL AS initrlanacl, " |
| "(%s lanowner) AS lanowner " |
| "FROM pg_language " |
| "WHERE lanispl " |
| "ORDER BY oid", |
| username_subquery); |
| } |
| else |
| { |
| /* Languages are owned by the bootstrap superuser, OID 10 */ |
| appendPQExpBuffer(query, "SELECT tableoid, oid, " |
| "lanname, lanpltrusted, lanplcallfoid, " |
| "0 AS laninline, lanvalidator, lanacl, " |
| "NULL AS rlanacl, " |
| "NULL AS initlanacl, NULL AS initrlanacl, " |
| "(%s '10') AS lanowner " |
| "FROM pg_language " |
| "WHERE lanispl " |
| "ORDER BY oid", |
| username_subquery); |
| } |
| ======= |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, " |
| "lanname, lanpltrusted, lanplcallfoid, " |
| "laninline, lanvalidator, " |
| "lanacl, " |
| "acldefault('l', lanowner) AS acldefault, " |
| "lanowner " |
| "FROM pg_language " |
| "WHERE lanispl " |
| "ORDER BY oid"); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numProcLangs = ntups; |
| |
| planginfo = (ProcLangInfo *) pg_malloc(ntups * sizeof(ProcLangInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_lanname = PQfnumber(res, "lanname"); |
| i_lanpltrusted = PQfnumber(res, "lanpltrusted"); |
| i_lanplcallfoid = PQfnumber(res, "lanplcallfoid"); |
| i_laninline = PQfnumber(res, "laninline"); |
| i_lanvalidator = PQfnumber(res, "lanvalidator"); |
| i_lanacl = PQfnumber(res, "lanacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| i_lanowner = PQfnumber(res, "lanowner"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| planginfo[i].dobj.objType = DO_PROCLANG; |
| planginfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| planginfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&planginfo[i].dobj); |
| |
| planginfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_lanname)); |
| planginfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_lanacl)); |
| planginfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| planginfo[i].dacl.privtype = 0; |
| planginfo[i].dacl.initprivs = NULL; |
| planginfo[i].lanpltrusted = *(PQgetvalue(res, i, i_lanpltrusted)) == 't'; |
| planginfo[i].lanplcallfoid = atooid(PQgetvalue(res, i, i_lanplcallfoid)); |
| planginfo[i].laninline = atooid(PQgetvalue(res, i, i_laninline)); |
| planginfo[i].lanvalidator = atooid(PQgetvalue(res, i, i_lanvalidator)); |
| planginfo[i].lanowner = getRoleName(PQgetvalue(res, i, i_lanowner)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableProcLang(&(planginfo[i]), fout); |
| |
| /* Mark whether language has an ACL */ |
| if (!PQgetisnull(res, i, i_lanacl)) |
| planginfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return planginfo; |
| } |
| |
| /* |
| * getCasts |
| * get basic information about most casts in the system |
| * |
| * numCasts is set to the number of casts read in |
| * |
| * Skip casts from a range to its multirange, since we'll create those |
| * automatically. |
| */ |
| CastInfo * |
| getCasts(Archive *fout, int *numCasts) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query = createPQExpBuffer(); |
| CastInfo *castinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_castsource; |
| int i_casttarget; |
| int i_castfunc; |
| int i_castcontext; |
| int i_castmethod; |
| |
| if (fout->remoteVersion >= 140000) |
| { |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, " |
| "castsource, casttarget, castfunc, castcontext, " |
| "castmethod " |
| "FROM pg_cast c " |
| "WHERE NOT EXISTS ( " |
| "SELECT 1 FROM pg_range r " |
| "WHERE c.castsource = r.rngtypid " |
| "AND c.casttarget = r.rngmultitypid " |
| ") " |
| "ORDER BY 3,4"); |
| } |
| else |
| { |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, " |
| "castsource, casttarget, castfunc, castcontext, " |
| "castmethod " |
| "FROM pg_cast ORDER BY 3,4"); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numCasts = ntups; |
| |
| castinfo = (CastInfo *) pg_malloc(ntups * sizeof(CastInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_castsource = PQfnumber(res, "castsource"); |
| i_casttarget = PQfnumber(res, "casttarget"); |
| i_castfunc = PQfnumber(res, "castfunc"); |
| i_castcontext = PQfnumber(res, "castcontext"); |
| i_castmethod = PQfnumber(res, "castmethod"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| PQExpBufferData namebuf; |
| TypeInfo *sTypeInfo; |
| TypeInfo *tTypeInfo; |
| |
| castinfo[i].dobj.objType = DO_CAST; |
| castinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| castinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&castinfo[i].dobj); |
| castinfo[i].castsource = atooid(PQgetvalue(res, i, i_castsource)); |
| castinfo[i].casttarget = atooid(PQgetvalue(res, i, i_casttarget)); |
| castinfo[i].castfunc = atooid(PQgetvalue(res, i, i_castfunc)); |
| castinfo[i].castcontext = *(PQgetvalue(res, i, i_castcontext)); |
| castinfo[i].castmethod = *(PQgetvalue(res, i, i_castmethod)); |
| |
| /* |
| * Try to name cast as concatenation of typnames. This is only used |
| * for purposes of sorting. If we fail to find either type, the name |
| * will be an empty string. |
| */ |
| initPQExpBuffer(&namebuf); |
| sTypeInfo = findTypeByOid(castinfo[i].castsource); |
| tTypeInfo = findTypeByOid(castinfo[i].casttarget); |
| if (sTypeInfo && tTypeInfo) |
| appendPQExpBuffer(&namebuf, "%s %s", |
| sTypeInfo->dobj.name, tTypeInfo->dobj.name); |
| castinfo[i].dobj.name = namebuf.data; |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableCast(&(castinfo[i]), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return castinfo; |
| } |
| |
| static char * |
| get_language_name(Archive *fout, Oid langid) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| char *lanname; |
| |
| query = createPQExpBuffer(); |
| appendPQExpBuffer(query, "SELECT lanname FROM pg_language WHERE oid = %u", langid); |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| lanname = pg_strdup(fmtId(PQgetvalue(res, 0, 0))); |
| destroyPQExpBuffer(query); |
| PQclear(res); |
| |
| return lanname; |
| } |
| |
| /* |
| * getTransforms |
| * get basic information about every transform in the system |
| * |
| * numTransforms is set to the number of transforms read in |
| */ |
| TransformInfo * |
| getTransforms(Archive *fout, int *numTransforms) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| TransformInfo *transforminfo; |
| int i_tableoid; |
| int i_oid; |
| int i_trftype; |
| int i_trflang; |
| int i_trffromsql; |
| int i_trftosql; |
| |
| /* Transforms didn't exist pre-9.5 */ |
| if (fout->remoteVersion < 90500) |
| { |
| *numTransforms = 0; |
| return NULL; |
| } |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, " |
| "trftype, trflang, trffromsql::oid, trftosql::oid " |
| "FROM pg_transform " |
| "ORDER BY 3,4"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| *numTransforms = ntups; |
| |
| transforminfo = (TransformInfo *) pg_malloc(ntups * sizeof(TransformInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_trftype = PQfnumber(res, "trftype"); |
| i_trflang = PQfnumber(res, "trflang"); |
| i_trffromsql = PQfnumber(res, "trffromsql"); |
| i_trftosql = PQfnumber(res, "trftosql"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| PQExpBufferData namebuf; |
| TypeInfo *typeInfo; |
| char *lanname; |
| |
| transforminfo[i].dobj.objType = DO_TRANSFORM; |
| transforminfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| transforminfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&transforminfo[i].dobj); |
| transforminfo[i].trftype = atooid(PQgetvalue(res, i, i_trftype)); |
| transforminfo[i].trflang = atooid(PQgetvalue(res, i, i_trflang)); |
| transforminfo[i].trffromsql = atooid(PQgetvalue(res, i, i_trffromsql)); |
| transforminfo[i].trftosql = atooid(PQgetvalue(res, i, i_trftosql)); |
| |
| /* |
| * Try to name transform as concatenation of type and language name. |
| * This is only used for purposes of sorting. If we fail to find |
| * either, the name will be an empty string. |
| */ |
| initPQExpBuffer(&namebuf); |
| typeInfo = findTypeByOid(transforminfo[i].trftype); |
| lanname = get_language_name(fout, transforminfo[i].trflang); |
| if (typeInfo && lanname) |
| appendPQExpBuffer(&namebuf, "%s %s", |
| typeInfo->dobj.name, lanname); |
| transforminfo[i].dobj.name = namebuf.data; |
| free(lanname); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(transforminfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return transforminfo; |
| } |
| |
| /* |
| * getTableAttrs - |
| * for each interesting table, read info about its attributes |
| * (names, types, default values, CHECK constraints, etc) |
| * |
| * modifies tblinfo |
| */ |
| void |
| getTableAttrs(Archive *fout, TableInfo *tblinfo, int numTables) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer tbloids = createPQExpBuffer(); |
| PQExpBuffer checkoids = createPQExpBuffer(); |
| PGresult *res; |
| int ntups; |
| int curtblindx; |
| int i_attrelid; |
| int i_attnum; |
| int i_attname; |
| int i_atttypname; |
| int i_attstattarget; |
| int i_attstorage; |
| int i_typstorage; |
| int i_attidentity; |
| int i_attgenerated; |
| int i_attisdropped; |
| int i_attlen; |
| int i_attalign; |
| int i_attislocal; |
| int i_attnotnull; |
| int i_attoptions; |
| int i_attcollation; |
| int i_attcompression; |
| int i_attfdwoptions; |
| int i_attmissingval; |
| int i_atthasdef; |
| |
| <<<<<<< HEAD |
| /* GPDB_14_MERGE_FIXME: GPDB specific column, need to keep this for easy to use*/ |
| int i_attencoding; |
| |
| ======= |
| /* |
| * We want to perform just one query against pg_attribute, and then just |
| * one against pg_attrdef (for DEFAULTs) and one against pg_constraint |
| * (for CHECK constraints). However, we mustn't try to select every row |
| * of those catalogs and then sort it out on the client side, because some |
| * of the server-side functions we need would be unsafe to apply to tables |
| * we don't have lock on. Hence, we build an array of the OIDs of tables |
| * we care about (and now have lock on!), and use a WHERE clause to |
| * constrain which rows are selected. |
| */ |
| appendPQExpBufferChar(tbloids, '{'); |
| appendPQExpBufferChar(checkoids, '{'); |
| >>>>>>> REL_16_9 |
| for (int i = 0; i < numTables; i++) |
| { |
| TableInfo *tbinfo = &tblinfo[i]; |
| |
| /* Don't bother to collect info for sequences */ |
| if (tbinfo->relkind == RELKIND_SEQUENCE) |
| continue; |
| |
| /* Don't bother with uninteresting tables, either */ |
| if (!tbinfo->interesting) |
| continue; |
| |
| /* OK, we need info for this table */ |
| if (tbloids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(tbloids, ','); |
| appendPQExpBuffer(tbloids, "%u", tbinfo->dobj.catId.oid); |
| |
| if (tbinfo->ncheck > 0) |
| { |
| /* Also make a list of the ones with check constraints */ |
| if (checkoids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(checkoids, ','); |
| appendPQExpBuffer(checkoids, "%u", tbinfo->dobj.catId.oid); |
| } |
| } |
| appendPQExpBufferChar(tbloids, '}'); |
| appendPQExpBufferChar(checkoids, '}'); |
| |
| /* |
| * Find all the user attributes and their types. |
| * |
| * Since we only want to dump COLLATE clauses for attributes whose |
| * collation is different from their type's default, we use a CASE here to |
| * suppress uninteresting attcollations cheaply. |
| */ |
| appendPQExpBufferStr(q, |
| "SELECT\n" |
| "a.attrelid,\n" |
| "a.attnum,\n" |
| "a.attname,\n" |
| "a.attstattarget,\n" |
| "a.attstorage,\n" |
| "t.typstorage,\n" |
| "a.attnotnull,\n" |
| "a.atthasdef,\n" |
| "a.attisdropped,\n" |
| "a.attlen,\n" |
| "a.attalign,\n" |
| "a.attislocal,\n" |
| "pg_catalog.format_type(t.oid, a.atttypmod) AS atttypname,\n" |
| "array_to_string(a.attoptions, ', ') AS attoptions,\n" |
| "CASE WHEN a.attcollation <> t.typcollation " |
| "THEN a.attcollation ELSE 0 END AS attcollation,\n" |
| "pg_catalog.array_to_string(ARRAY(" |
| "SELECT pg_catalog.quote_ident(option_name) || " |
| "' ' || pg_catalog.quote_literal(option_value) " |
| "FROM pg_catalog.pg_options_to_table(attfdwoptions) " |
| "ORDER BY option_name" |
| "), E',\n ') AS attfdwoptions,\n"); |
| |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(q, |
| "a.attcompression AS attcompression,\n"); |
| else |
| appendPQExpBufferStr(q, |
| "'' AS attcompression,\n"); |
| |
| if (fout->remoteVersion >= 100000) |
| appendPQExpBufferStr(q, |
| "a.attidentity,\n"); |
| else |
| appendPQExpBufferStr(q, |
| "'' AS attidentity,\n"); |
| |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(q, |
| "CASE WHEN a.atthasmissing AND NOT a.attisdropped " |
| "THEN a.attmissingval ELSE null END AS attmissingval,\n"); |
| else |
| appendPQExpBufferStr(q, |
| "NULL AS attmissingval,\n"); |
| |
| if (fout->remoteVersion >= 120000) |
| appendPQExpBufferStr(q, |
| "a.attgenerated\n"); |
| else |
| appendPQExpBufferStr(q, |
| "'' AS attgenerated\n"); |
| |
| /* need left join to pg_type to not fail on dropped columns ... */ |
| appendPQExpBuffer(q, |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_attribute a ON (src.tbloid = a.attrelid) " |
| "LEFT JOIN pg_catalog.pg_type t " |
| "ON (a.atttypid = t.oid)\n" |
| "WHERE a.attnum > 0::pg_catalog.int2\n" |
| "ORDER BY a.attrelid, a.attnum", |
| tbloids->data); |
| |
| res = ExecuteSqlQuery(fout, q->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_attrelid = PQfnumber(res, "attrelid"); |
| i_attnum = PQfnumber(res, "attnum"); |
| i_attname = PQfnumber(res, "attname"); |
| i_atttypname = PQfnumber(res, "atttypname"); |
| i_attstattarget = PQfnumber(res, "attstattarget"); |
| i_attstorage = PQfnumber(res, "attstorage"); |
| i_typstorage = PQfnumber(res, "typstorage"); |
| i_attidentity = PQfnumber(res, "attidentity"); |
| i_attgenerated = PQfnumber(res, "attgenerated"); |
| i_attisdropped = PQfnumber(res, "attisdropped"); |
| i_attlen = PQfnumber(res, "attlen"); |
| i_attalign = PQfnumber(res, "attalign"); |
| i_attislocal = PQfnumber(res, "attislocal"); |
| i_attnotnull = PQfnumber(res, "attnotnull"); |
| i_attoptions = PQfnumber(res, "attoptions"); |
| i_attcollation = PQfnumber(res, "attcollation"); |
| i_attcompression = PQfnumber(res, "attcompression"); |
| i_attfdwoptions = PQfnumber(res, "attfdwoptions"); |
| i_attmissingval = PQfnumber(res, "attmissingval"); |
| i_atthasdef = PQfnumber(res, "atthasdef"); |
| |
| /* Within the next loop, we'll accumulate OIDs of tables with defaults */ |
| resetPQExpBuffer(tbloids); |
| appendPQExpBufferChar(tbloids, '{'); |
| |
| /* |
| * Outer loop iterates once per table, not once per row. Incrementing of |
| * r is handled by the inner loop. |
| */ |
| curtblindx = -1; |
| for (int r = 0; r < ntups;) |
| { |
| Oid attrelid = atooid(PQgetvalue(res, r, i_attrelid)); |
| TableInfo *tbinfo = NULL; |
| int numatts; |
| bool hasdefaults; |
| |
| /* Count rows for this table */ |
| for (numatts = 1; numatts < ntups - r; numatts++) |
| if (atooid(PQgetvalue(res, r + numatts, i_attrelid)) != attrelid) |
| break; |
| |
| /* |
| * Locate the associated TableInfo; we rely on tblinfo[] being in OID |
| * order. |
| */ |
| while (++curtblindx < numTables) |
| { |
| tbinfo = &tblinfo[curtblindx]; |
| if (tbinfo->dobj.catId.oid == attrelid) |
| break; |
| } |
| if (curtblindx >= numTables) |
| pg_fatal("unrecognized table OID %u", attrelid); |
| /* cross-check that we only got requested tables */ |
| if (tbinfo->relkind == RELKIND_SEQUENCE || |
| !tbinfo->interesting) |
| pg_fatal("unexpected column data for table \"%s\"", |
| tbinfo->dobj.name); |
| |
| <<<<<<< HEAD |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBuffer(q, |
| "a.attcompression AS attcompression,\n"); |
| else |
| appendPQExpBuffer(q, |
| "'' AS attcompression,\n"); |
| |
| appendPQExpBuffer(q, |
| "pg_catalog.array_to_string(e.attoptions, ',') AS attencoding,\n"); |
| |
| if (fout->remoteVersion >= 90200) |
| appendPQExpBufferStr(q, |
| "pg_catalog.array_to_string(ARRAY(" |
| "SELECT pg_catalog.quote_ident(option_name) || " |
| "' ' || pg_catalog.quote_literal(option_value) " |
| "FROM pg_catalog.pg_options_to_table(attfdwoptions) " |
| "ORDER BY option_name" |
| "), E',\n ') AS attfdwoptions,\n"); |
| else |
| appendPQExpBufferStr(q, |
| "'' AS attfdwoptions,\n"); |
| |
| if (fout->remoteVersion >= 100000) |
| appendPQExpBufferStr(q, |
| "a.attidentity,\n"); |
| else |
| appendPQExpBufferStr(q, |
| "'' AS attidentity,\n"); |
| |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(q, |
| "CASE WHEN a.atthasmissing AND NOT a.attisdropped " |
| "THEN a.attmissingval ELSE null END AS attmissingval,\n"); |
| else |
| appendPQExpBufferStr(q, |
| "NULL AS attmissingval,\n"); |
| |
| if (fout->remoteVersion >= 120000) |
| appendPQExpBufferStr(q, |
| "a.attgenerated\n"); |
| else |
| appendPQExpBufferStr(q, |
| "'' AS attgenerated\n"); |
| |
| /* need left join here to not fail on dropped columns ... */ |
| appendPQExpBuffer(q, |
| "FROM pg_catalog.pg_attribute a LEFT JOIN pg_catalog.pg_type t " |
| "ON a.atttypid = t.oid\n" |
| "LEFT OUTER JOIN pg_catalog.pg_attribute_encoding e ON e.attrelid = a.attrelid AND e.attnum = a.attnum \n" |
| "WHERE a.attrelid = '%u'::pg_catalog.oid " |
| "AND a.attnum > 0::pg_catalog.int2\n" |
| "ORDER BY a.attnum", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, q->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_attencoding = PQfnumber(res, "attencoding"); |
| |
| tbinfo->numatts = ntups; |
| tbinfo->attnames = (char **) pg_malloc(ntups * sizeof(char *)); |
| tbinfo->atttypnames = (char **) pg_malloc(ntups * sizeof(char *)); |
| tbinfo->atttypmod = (int *) pg_malloc(ntups * sizeof(int)); |
| tbinfo->attstattarget = (int *) pg_malloc(ntups * sizeof(int)); |
| tbinfo->attstorage = (char *) pg_malloc(ntups * sizeof(char)); |
| tbinfo->typstorage = (char *) pg_malloc(ntups * sizeof(char)); |
| tbinfo->attidentity = (char *) pg_malloc(ntups * sizeof(char)); |
| tbinfo->attgenerated = (char *) pg_malloc(ntups * sizeof(char)); |
| tbinfo->attisdropped = (bool *) pg_malloc(ntups * sizeof(bool)); |
| tbinfo->attlen = (int *) pg_malloc(ntups * sizeof(int)); |
| tbinfo->attalign = (char *) pg_malloc(ntups * sizeof(char)); |
| tbinfo->attislocal = (bool *) pg_malloc(ntups * sizeof(bool)); |
| tbinfo->attoptions = (char **) pg_malloc(ntups * sizeof(char *)); |
| tbinfo->attcollation = (Oid *) pg_malloc(ntups * sizeof(Oid)); |
| tbinfo->attcompression = (char *) pg_malloc(ntups * sizeof(char)); |
| tbinfo->attfdwoptions = (char **) pg_malloc(ntups * sizeof(char *)); |
| tbinfo->attmissingval = (char **) pg_malloc(ntups * sizeof(char *)); |
| tbinfo->notnull = (bool *) pg_malloc(ntups * sizeof(bool)); |
| tbinfo->inhNotNull = (bool *) pg_malloc(ntups * sizeof(bool)); |
| tbinfo->attencoding = (char **) pg_malloc(ntups * sizeof(char *)); |
| tbinfo->attrdefs = (AttrDefInfo **) pg_malloc(ntups * sizeof(AttrDefInfo *)); |
| ======= |
| /* Save data for this table */ |
| tbinfo->numatts = numatts; |
| tbinfo->attnames = (char **) pg_malloc(numatts * sizeof(char *)); |
| tbinfo->atttypnames = (char **) pg_malloc(numatts * sizeof(char *)); |
| tbinfo->attstattarget = (int *) pg_malloc(numatts * sizeof(int)); |
| tbinfo->attstorage = (char *) pg_malloc(numatts * sizeof(char)); |
| tbinfo->typstorage = (char *) pg_malloc(numatts * sizeof(char)); |
| tbinfo->attidentity = (char *) pg_malloc(numatts * sizeof(char)); |
| tbinfo->attgenerated = (char *) pg_malloc(numatts * sizeof(char)); |
| tbinfo->attisdropped = (bool *) pg_malloc(numatts * sizeof(bool)); |
| tbinfo->attlen = (int *) pg_malloc(numatts * sizeof(int)); |
| tbinfo->attalign = (char *) pg_malloc(numatts * sizeof(char)); |
| tbinfo->attislocal = (bool *) pg_malloc(numatts * sizeof(bool)); |
| tbinfo->attoptions = (char **) pg_malloc(numatts * sizeof(char *)); |
| tbinfo->attcollation = (Oid *) pg_malloc(numatts * sizeof(Oid)); |
| tbinfo->attcompression = (char *) pg_malloc(numatts * sizeof(char)); |
| tbinfo->attfdwoptions = (char **) pg_malloc(numatts * sizeof(char *)); |
| tbinfo->attmissingval = (char **) pg_malloc(numatts * sizeof(char *)); |
| tbinfo->notnull = (bool *) pg_malloc(numatts * sizeof(bool)); |
| tbinfo->inhNotNull = (bool *) pg_malloc(numatts * sizeof(bool)); |
| tbinfo->attrdefs = (AttrDefInfo **) pg_malloc(numatts * sizeof(AttrDefInfo *)); |
| >>>>>>> REL_16_9 |
| hasdefaults = false; |
| |
| for (int j = 0; j < numatts; j++, r++) |
| { |
| if (j + 1 != atoi(PQgetvalue(res, r, i_attnum))) |
| pg_fatal("invalid column numbering in table \"%s\"", |
| tbinfo->dobj.name); |
| tbinfo->attnames[j] = pg_strdup(PQgetvalue(res, r, i_attname)); |
| tbinfo->atttypnames[j] = pg_strdup(PQgetvalue(res, r, i_atttypname)); |
| tbinfo->attstattarget[j] = atoi(PQgetvalue(res, r, i_attstattarget)); |
| tbinfo->attstorage[j] = *(PQgetvalue(res, r, i_attstorage)); |
| tbinfo->typstorage[j] = *(PQgetvalue(res, r, i_typstorage)); |
| tbinfo->attidentity[j] = *(PQgetvalue(res, r, i_attidentity)); |
| tbinfo->attgenerated[j] = *(PQgetvalue(res, r, i_attgenerated)); |
| tbinfo->needs_override = tbinfo->needs_override || (tbinfo->attidentity[j] == ATTRIBUTE_IDENTITY_ALWAYS); |
| tbinfo->attisdropped[j] = (PQgetvalue(res, r, i_attisdropped)[0] == 't'); |
| tbinfo->attlen[j] = atoi(PQgetvalue(res, r, i_attlen)); |
| tbinfo->attalign[j] = *(PQgetvalue(res, r, i_attalign)); |
| tbinfo->attislocal[j] = (PQgetvalue(res, r, i_attislocal)[0] == 't'); |
| tbinfo->notnull[j] = (PQgetvalue(res, r, i_attnotnull)[0] == 't'); |
| tbinfo->attoptions[j] = pg_strdup(PQgetvalue(res, r, i_attoptions)); |
| tbinfo->attcollation[j] = atooid(PQgetvalue(res, r, i_attcollation)); |
| tbinfo->attcompression[j] = *(PQgetvalue(res, r, i_attcompression)); |
| tbinfo->attfdwoptions[j] = pg_strdup(PQgetvalue(res, r, i_attfdwoptions)); |
| tbinfo->attmissingval[j] = pg_strdup(PQgetvalue(res, r, i_attmissingval)); |
| tbinfo->attrdefs[j] = NULL; /* fix below */ |
| if (PQgetvalue(res, r, i_atthasdef)[0] == 't') |
| hasdefaults = true; |
| /* these flags will be set in flagInhAttrs() */ |
| tbinfo->inhNotNull[j] = false; |
| |
| /* column storage attributes */ |
| if (!PQgetisnull(res, j, i_attencoding)) |
| tbinfo->attencoding[j] = pg_strdup(PQgetvalue(res, j, i_attencoding)); |
| else |
| tbinfo->attencoding[j] = NULL; |
| } |
| |
| if (hasdefaults) |
| { |
| /* Collect OIDs of interesting tables that have defaults */ |
| if (tbloids->len > 1) /* do we have more than the '{'? */ |
| appendPQExpBufferChar(tbloids, ','); |
| appendPQExpBuffer(tbloids, "%u", tbinfo->dobj.catId.oid); |
| } |
| } |
| |
| PQclear(res); |
| |
| /* |
| * Now get info about column defaults. This is skipped for a data-only |
| * dump, as it is only needed for table schemas. |
| */ |
| if (!dopt->dataOnly && tbloids->len > 1) |
| { |
| AttrDefInfo *attrdefs; |
| int numDefaults; |
| TableInfo *tbinfo = NULL; |
| |
| pg_log_info("finding table default expressions"); |
| |
| appendPQExpBufferChar(tbloids, '}'); |
| |
| printfPQExpBuffer(q, "SELECT a.tableoid, a.oid, adrelid, adnum, " |
| "pg_catalog.pg_get_expr(adbin, adrelid) AS adsrc\n" |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_attrdef a ON (src.tbloid = a.adrelid)\n" |
| "ORDER BY a.adrelid, a.adnum", |
| tbloids->data); |
| |
| res = ExecuteSqlQuery(fout, q->data, PGRES_TUPLES_OK); |
| |
| numDefaults = PQntuples(res); |
| attrdefs = (AttrDefInfo *) pg_malloc(numDefaults * sizeof(AttrDefInfo)); |
| |
| curtblindx = -1; |
| for (int j = 0; j < numDefaults; j++) |
| { |
| Oid adtableoid = atooid(PQgetvalue(res, j, 0)); |
| Oid adoid = atooid(PQgetvalue(res, j, 1)); |
| Oid adrelid = atooid(PQgetvalue(res, j, 2)); |
| int adnum = atoi(PQgetvalue(res, j, 3)); |
| char *adsrc = PQgetvalue(res, j, 4); |
| |
| /* |
| * Locate the associated TableInfo; we rely on tblinfo[] being in |
| * OID order. |
| */ |
| if (tbinfo == NULL || tbinfo->dobj.catId.oid != adrelid) |
| { |
| while (++curtblindx < numTables) |
| { |
| tbinfo = &tblinfo[curtblindx]; |
| if (tbinfo->dobj.catId.oid == adrelid) |
| break; |
| } |
| if (curtblindx >= numTables) |
| pg_fatal("unrecognized table OID %u", adrelid); |
| } |
| |
| if (adnum <= 0 || adnum > tbinfo->numatts) |
| pg_fatal("invalid adnum value %d for table \"%s\"", |
| adnum, tbinfo->dobj.name); |
| |
| /* |
| * dropped columns shouldn't have defaults, but just in case, |
| * ignore 'em |
| */ |
| if (tbinfo->attisdropped[adnum - 1]) |
| continue; |
| |
| attrdefs[j].dobj.objType = DO_ATTRDEF; |
| attrdefs[j].dobj.catId.tableoid = adtableoid; |
| attrdefs[j].dobj.catId.oid = adoid; |
| AssignDumpId(&attrdefs[j].dobj); |
| attrdefs[j].adtable = tbinfo; |
| attrdefs[j].adnum = adnum; |
| attrdefs[j].adef_expr = pg_strdup(adsrc); |
| |
| attrdefs[j].dobj.name = pg_strdup(tbinfo->dobj.name); |
| attrdefs[j].dobj.namespace = tbinfo->dobj.namespace; |
| |
| attrdefs[j].dobj.dump = tbinfo->dobj.dump; |
| |
| /* |
| * Figure out whether the default/generation expression should be |
| * dumped as part of the main CREATE TABLE (or similar) command or |
| * as a separate ALTER TABLE (or similar) command. The preference |
| * is to put it into the CREATE command, but in some cases that's |
| * not possible. |
| */ |
| if (tbinfo->attgenerated[adnum - 1]) |
| { |
| /* |
| * Column generation expressions cannot be dumped separately, |
| * because there is no syntax for it. By setting separate to |
| * false here we prevent the "default" from being processed as |
| * its own dumpable object. Later, flagInhAttrs() will mark |
| * it as not to be dumped at all, if possible (that is, if it |
| * can be inherited from a parent). |
| */ |
| attrdefs[j].separate = false; |
| } |
| else if (tbinfo->relkind == RELKIND_VIEW) |
| { |
| /* |
| * Defaults on a VIEW must always be dumped as separate ALTER |
| * TABLE commands. |
| */ |
| attrdefs[j].separate = true; |
| } |
| else if (!shouldPrintColumn(dopt, tbinfo, adnum - 1)) |
| { |
| /* column will be suppressed, print default separately */ |
| attrdefs[j].separate = true; |
| } |
| else |
| { |
| attrdefs[j].separate = false; |
| } |
| |
| if (!attrdefs[j].separate) |
| { |
| /* |
| * Mark the default as needing to appear before the table, so |
| * that any dependencies it has must be emitted before the |
| * CREATE TABLE. If this is not possible, we'll change to |
| * "separate" mode while sorting dependencies. |
| */ |
| addObjectDependency(&tbinfo->dobj, |
| attrdefs[j].dobj.dumpId); |
| } |
| |
| tbinfo->attrdefs[adnum - 1] = &attrdefs[j]; |
| } |
| |
| PQclear(res); |
| } |
| |
| /* |
| * Get info about table CHECK constraints. This is skipped for a |
| * data-only dump, as it is only needed for table schemas. |
| */ |
| if (!dopt->dataOnly && checkoids->len > 2) |
| { |
| ConstraintInfo *constrs; |
| int numConstrs; |
| int i_tableoid; |
| int i_oid; |
| int i_conrelid; |
| int i_conname; |
| int i_consrc; |
| int i_conislocal; |
| int i_convalidated; |
| |
| pg_log_info("finding table check constraints"); |
| |
| resetPQExpBuffer(q); |
| appendPQExpBuffer(q, |
| "SELECT c.tableoid, c.oid, conrelid, conname, " |
| "pg_catalog.pg_get_constraintdef(c.oid) AS consrc, " |
| "conislocal, convalidated " |
| "FROM unnest('%s'::pg_catalog.oid[]) AS src(tbloid)\n" |
| "JOIN pg_catalog.pg_constraint c ON (src.tbloid = c.conrelid)\n" |
| "WHERE contype = 'c' " |
| "ORDER BY c.conrelid, c.conname", |
| checkoids->data); |
| |
| res = ExecuteSqlQuery(fout, q->data, PGRES_TUPLES_OK); |
| |
| numConstrs = PQntuples(res); |
| constrs = (ConstraintInfo *) pg_malloc(numConstrs * sizeof(ConstraintInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_conrelid = PQfnumber(res, "conrelid"); |
| i_conname = PQfnumber(res, "conname"); |
| i_consrc = PQfnumber(res, "consrc"); |
| i_conislocal = PQfnumber(res, "conislocal"); |
| i_convalidated = PQfnumber(res, "convalidated"); |
| |
| /* As above, this loop iterates once per table, not once per row */ |
| curtblindx = -1; |
| for (int j = 0; j < numConstrs;) |
| { |
| Oid conrelid = atooid(PQgetvalue(res, j, i_conrelid)); |
| TableInfo *tbinfo = NULL; |
| int numcons; |
| |
| /* Count rows for this table */ |
| for (numcons = 1; numcons < numConstrs - j; numcons++) |
| if (atooid(PQgetvalue(res, j + numcons, i_conrelid)) != conrelid) |
| break; |
| |
| /* |
| * Locate the associated TableInfo; we rely on tblinfo[] being in |
| * OID order. |
| */ |
| while (++curtblindx < numTables) |
| { |
| tbinfo = &tblinfo[curtblindx]; |
| if (tbinfo->dobj.catId.oid == conrelid) |
| break; |
| } |
| if (curtblindx >= numTables) |
| pg_fatal("unrecognized table OID %u", conrelid); |
| |
| if (numcons != tbinfo->ncheck) |
| { |
| pg_log_error(ngettext("expected %d check constraint on table \"%s\" but found %d", |
| "expected %d check constraints on table \"%s\" but found %d", |
| tbinfo->ncheck), |
| tbinfo->ncheck, tbinfo->dobj.name, numcons); |
| pg_log_error_hint("The system catalogs might be corrupted."); |
| exit_nicely(1); |
| } |
| |
| tbinfo->checkexprs = constrs + j; |
| |
| for (int c = 0; c < numcons; c++, j++) |
| { |
| bool validated = PQgetvalue(res, j, i_convalidated)[0] == 't'; |
| |
| constrs[j].dobj.objType = DO_CONSTRAINT; |
| constrs[j].dobj.catId.tableoid = atooid(PQgetvalue(res, j, i_tableoid)); |
| constrs[j].dobj.catId.oid = atooid(PQgetvalue(res, j, i_oid)); |
| AssignDumpId(&constrs[j].dobj); |
| constrs[j].dobj.name = pg_strdup(PQgetvalue(res, j, i_conname)); |
| constrs[j].dobj.namespace = tbinfo->dobj.namespace; |
| constrs[j].contable = tbinfo; |
| constrs[j].condomain = NULL; |
| constrs[j].contype = 'c'; |
| constrs[j].condef = pg_strdup(PQgetvalue(res, j, i_consrc)); |
| constrs[j].confrelid = InvalidOid; |
| constrs[j].conindex = 0; |
| constrs[j].condeferrable = false; |
| constrs[j].condeferred = false; |
| constrs[j].conislocal = (PQgetvalue(res, j, i_conislocal)[0] == 't'); |
| |
| /* |
| * An unvalidated constraint needs to be dumped separately, so |
| * that potentially-violating existing data is loaded before |
| * the constraint. |
| */ |
| constrs[j].separate = !validated; |
| |
| constrs[j].dobj.dump = tbinfo->dobj.dump; |
| |
| /* |
| * Mark the constraint as needing to appear before the table |
| * --- this is so that any other dependencies of the |
| * constraint will be emitted before we try to create the |
| * table. If the constraint is to be dumped separately, it |
| * will be dumped after data is loaded anyway, so don't do it. |
| * (There's an automatic dependency in the opposite direction |
| * anyway, so don't need to add one manually here.) |
| */ |
| if (!constrs[j].separate) |
| addObjectDependency(&tbinfo->dobj, |
| constrs[j].dobj.dumpId); |
| |
| /* |
| * We will detect later whether the constraint must be split |
| * out from the table definition. |
| */ |
| } |
| } |
| |
| PQclear(res); |
| } |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(tbloids); |
| destroyPQExpBuffer(checkoids); |
| } |
| |
| /* |
| * Test whether a column should be printed as part of table's CREATE TABLE. |
| * Column number is zero-based. |
| * |
| * Normally this is always true, but it's false for dropped columns, as well |
| * as those that were inherited without any local definition. (If we print |
| * such a column it will mistakenly get pg_attribute.attislocal set to true.) |
| * For partitions, it's always true, because we want the partitions to be |
| * created independently and ATTACH PARTITION used afterwards. |
| * |
| * In binary_upgrade mode, we must print all columns and fix the attislocal/ |
| * attisdropped state later, so as to keep control of the physical column |
| * order. |
| * |
| * This function exists because there are scattered nonobvious places that |
| * must be kept in sync with this decision. |
| */ |
| bool |
| shouldPrintColumn(const DumpOptions *dopt, const TableInfo *tbinfo, int colno) |
| { |
| if (dopt->binary_upgrade) |
| return true; |
| if (tbinfo->attisdropped[colno]) |
| return false; |
| return (tbinfo->attislocal[colno] || tbinfo->ispartition); |
| } |
| |
| |
| /* |
| * getTSParsers: |
| * read all text search parsers in the system catalogs and return them |
| * in the TSParserInfo* structure |
| * |
| * numTSParsers is set to the number of parsers read in |
| */ |
| TSParserInfo * |
| getTSParsers(Archive *fout, int *numTSParsers) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| TSParserInfo *prsinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_prsname; |
| int i_prsnamespace; |
| int i_prsstart; |
| int i_prstoken; |
| int i_prsend; |
| int i_prsheadline; |
| int i_prslextype; |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * find all text search objects, including builtin ones; we filter out |
| * system-defined objects at dump-out time. |
| */ |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, prsname, prsnamespace, " |
| "prsstart::oid, prstoken::oid, " |
| "prsend::oid, prsheadline::oid, prslextype::oid " |
| "FROM pg_ts_parser"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numTSParsers = ntups; |
| |
| prsinfo = (TSParserInfo *) pg_malloc(ntups * sizeof(TSParserInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_prsname = PQfnumber(res, "prsname"); |
| i_prsnamespace = PQfnumber(res, "prsnamespace"); |
| i_prsstart = PQfnumber(res, "prsstart"); |
| i_prstoken = PQfnumber(res, "prstoken"); |
| i_prsend = PQfnumber(res, "prsend"); |
| i_prsheadline = PQfnumber(res, "prsheadline"); |
| i_prslextype = PQfnumber(res, "prslextype"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| prsinfo[i].dobj.objType = DO_TSPARSER; |
| prsinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| prsinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&prsinfo[i].dobj); |
| prsinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_prsname)); |
| prsinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_prsnamespace))); |
| prsinfo[i].prsstart = atooid(PQgetvalue(res, i, i_prsstart)); |
| prsinfo[i].prstoken = atooid(PQgetvalue(res, i, i_prstoken)); |
| prsinfo[i].prsend = atooid(PQgetvalue(res, i, i_prsend)); |
| prsinfo[i].prsheadline = atooid(PQgetvalue(res, i, i_prsheadline)); |
| prsinfo[i].prslextype = atooid(PQgetvalue(res, i, i_prslextype)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(prsinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return prsinfo; |
| } |
| |
| /* |
| * getTSDictionaries: |
| * read all text search dictionaries in the system catalogs and return them |
| * in the TSDictInfo* structure |
| * |
| * numTSDicts is set to the number of dictionaries read in |
| */ |
| TSDictInfo * |
| getTSDictionaries(Archive *fout, int *numTSDicts) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| TSDictInfo *dictinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_dictname; |
| int i_dictnamespace; |
| int i_dictowner; |
| int i_dicttemplate; |
| int i_dictinitoption; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, dictname, " |
| "dictnamespace, dictowner, " |
| "dicttemplate, dictinitoption " |
| "FROM pg_ts_dict"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numTSDicts = ntups; |
| |
| dictinfo = (TSDictInfo *) pg_malloc(ntups * sizeof(TSDictInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_dictname = PQfnumber(res, "dictname"); |
| i_dictnamespace = PQfnumber(res, "dictnamespace"); |
| i_dictowner = PQfnumber(res, "dictowner"); |
| i_dictinitoption = PQfnumber(res, "dictinitoption"); |
| i_dicttemplate = PQfnumber(res, "dicttemplate"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| dictinfo[i].dobj.objType = DO_TSDICT; |
| dictinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| dictinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&dictinfo[i].dobj); |
| dictinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_dictname)); |
| dictinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_dictnamespace))); |
| dictinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_dictowner)); |
| dictinfo[i].dicttemplate = atooid(PQgetvalue(res, i, i_dicttemplate)); |
| if (PQgetisnull(res, i, i_dictinitoption)) |
| dictinfo[i].dictinitoption = NULL; |
| else |
| dictinfo[i].dictinitoption = pg_strdup(PQgetvalue(res, i, i_dictinitoption)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(dictinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return dictinfo; |
| } |
| |
| /* |
| * getTSTemplates: |
| * read all text search templates in the system catalogs and return them |
| * in the TSTemplateInfo* structure |
| * |
| * numTSTemplates is set to the number of templates read in |
| */ |
| TSTemplateInfo * |
| getTSTemplates(Archive *fout, int *numTSTemplates) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| TSTemplateInfo *tmplinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_tmplname; |
| int i_tmplnamespace; |
| int i_tmplinit; |
| int i_tmpllexize; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, tmplname, " |
| "tmplnamespace, tmplinit::oid, tmpllexize::oid " |
| "FROM pg_ts_template"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numTSTemplates = ntups; |
| |
| tmplinfo = (TSTemplateInfo *) pg_malloc(ntups * sizeof(TSTemplateInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_tmplname = PQfnumber(res, "tmplname"); |
| i_tmplnamespace = PQfnumber(res, "tmplnamespace"); |
| i_tmplinit = PQfnumber(res, "tmplinit"); |
| i_tmpllexize = PQfnumber(res, "tmpllexize"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| tmplinfo[i].dobj.objType = DO_TSTEMPLATE; |
| tmplinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| tmplinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&tmplinfo[i].dobj); |
| tmplinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_tmplname)); |
| tmplinfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_tmplnamespace))); |
| tmplinfo[i].tmplinit = atooid(PQgetvalue(res, i, i_tmplinit)); |
| tmplinfo[i].tmpllexize = atooid(PQgetvalue(res, i, i_tmpllexize)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(tmplinfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return tmplinfo; |
| } |
| |
| /* |
| * getTSConfigurations: |
| * read all text search configurations in the system catalogs and return |
| * them in the TSConfigInfo* structure |
| * |
| * numTSConfigs is set to the number of configurations read in |
| */ |
| TSConfigInfo * |
| getTSConfigurations(Archive *fout, int *numTSConfigs) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| TSConfigInfo *cfginfo; |
| int i_tableoid; |
| int i_oid; |
| int i_cfgname; |
| int i_cfgnamespace; |
| int i_cfgowner; |
| int i_cfgparser; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, cfgname, " |
| "cfgnamespace, cfgowner, cfgparser " |
| "FROM pg_ts_config"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numTSConfigs = ntups; |
| |
| cfginfo = (TSConfigInfo *) pg_malloc(ntups * sizeof(TSConfigInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_cfgname = PQfnumber(res, "cfgname"); |
| i_cfgnamespace = PQfnumber(res, "cfgnamespace"); |
| i_cfgowner = PQfnumber(res, "cfgowner"); |
| i_cfgparser = PQfnumber(res, "cfgparser"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| cfginfo[i].dobj.objType = DO_TSCONFIG; |
| cfginfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| cfginfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&cfginfo[i].dobj); |
| cfginfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_cfgname)); |
| cfginfo[i].dobj.namespace = |
| findNamespace(atooid(PQgetvalue(res, i, i_cfgnamespace))); |
| cfginfo[i].rolname = getRoleName(PQgetvalue(res, i, i_cfgowner)); |
| cfginfo[i].cfgparser = atooid(PQgetvalue(res, i, i_cfgparser)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(cfginfo[i].dobj), fout); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return cfginfo; |
| } |
| |
| /* |
| * getForeignDataWrappers: |
| * read all foreign-data wrappers in the system catalogs and return |
| * them in the FdwInfo* structure |
| * |
| * numForeignDataWrappers is set to the number of fdws read in |
| */ |
| FdwInfo * |
| getForeignDataWrappers(Archive *fout, int *numForeignDataWrappers) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| FdwInfo *fdwinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_fdwname; |
| int i_fdwowner; |
| int i_fdwhandler; |
| int i_fdwvalidator; |
| int i_fdwacl; |
| int i_acldefault; |
| int i_fdwoptions; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, fdwname, " |
| "fdwowner, " |
| "fdwhandler::pg_catalog.regproc, " |
| "fdwvalidator::pg_catalog.regproc, " |
| "fdwacl, " |
| "acldefault('F', fdwowner) AS acldefault, " |
| "array_to_string(ARRAY(" |
| "SELECT quote_ident(option_name) || ' ' || " |
| "quote_literal(option_value) " |
| "FROM pg_options_to_table(fdwoptions) " |
| "ORDER BY option_name" |
| "), E',\n ') AS fdwoptions " |
| "FROM pg_foreign_data_wrapper"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numForeignDataWrappers = ntups; |
| |
| fdwinfo = (FdwInfo *) pg_malloc(ntups * sizeof(FdwInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_fdwname = PQfnumber(res, "fdwname"); |
| i_fdwowner = PQfnumber(res, "fdwowner"); |
| i_fdwhandler = PQfnumber(res, "fdwhandler"); |
| i_fdwvalidator = PQfnumber(res, "fdwvalidator"); |
| i_fdwacl = PQfnumber(res, "fdwacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| i_fdwoptions = PQfnumber(res, "fdwoptions"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| fdwinfo[i].dobj.objType = DO_FDW; |
| fdwinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| fdwinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&fdwinfo[i].dobj); |
| fdwinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_fdwname)); |
| fdwinfo[i].dobj.namespace = NULL; |
| fdwinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_fdwacl)); |
| fdwinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| fdwinfo[i].dacl.privtype = 0; |
| fdwinfo[i].dacl.initprivs = NULL; |
| fdwinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_fdwowner)); |
| fdwinfo[i].fdwhandler = pg_strdup(PQgetvalue(res, i, i_fdwhandler)); |
| fdwinfo[i].fdwvalidator = pg_strdup(PQgetvalue(res, i, i_fdwvalidator)); |
| fdwinfo[i].fdwoptions = pg_strdup(PQgetvalue(res, i, i_fdwoptions)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(fdwinfo[i].dobj), fout); |
| |
| <<<<<<< HEAD |
| /* Leave out the built-in gp_exttable_fdw */ |
| if (fdwinfo[i].dobj.catId.oid < (Oid) FirstNormalObjectId) |
| fdwinfo[i].dobj.dump = DUMP_COMPONENT_NONE; |
| |
| /* Do not try to dump ACL if no ACL exists. */ |
| if (PQgetisnull(res, i, i_fdwacl) && PQgetisnull(res, i, i_rfdwacl) && |
| PQgetisnull(res, i, i_initfdwacl) && |
| PQgetisnull(res, i, i_initrfdwacl)) |
| fdwinfo[i].dobj.dump &= ~DUMP_COMPONENT_ACL; |
| ======= |
| /* Mark whether FDW has an ACL */ |
| if (!PQgetisnull(res, i, i_fdwacl)) |
| fdwinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| >>>>>>> REL_16_9 |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return fdwinfo; |
| } |
| |
| /* |
| * getForeignServers: |
| * read all foreign servers in the system catalogs and return |
| * them in the ForeignServerInfo * structure |
| * |
| * numForeignServers is set to the number of servers read in |
| */ |
| ForeignServerInfo * |
| getForeignServers(Archive *fout, int *numForeignServers) |
| { |
| PGresult *res; |
| int ntups; |
| int i; |
| PQExpBuffer query; |
| ForeignServerInfo *srvinfo; |
| int i_tableoid; |
| int i_oid; |
| int i_srvname; |
| int i_srvowner; |
| int i_srvfdw; |
| int i_srvtype; |
| int i_srvversion; |
| int i_srvacl; |
| int i_acldefault; |
| int i_srvoptions; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT tableoid, oid, srvname, " |
| "srvowner, " |
| "srvfdw, srvtype, srvversion, srvacl, " |
| "acldefault('S', srvowner) AS acldefault, " |
| "array_to_string(ARRAY(" |
| "SELECT quote_ident(option_name) || ' ' || " |
| "quote_literal(option_value) " |
| "FROM pg_options_to_table(srvoptions) " |
| "ORDER BY option_name" |
| "), E',\n ') AS srvoptions " |
| "FROM pg_foreign_server"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numForeignServers = ntups; |
| |
| srvinfo = (ForeignServerInfo *) pg_malloc(ntups * sizeof(ForeignServerInfo)); |
| |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_oid = PQfnumber(res, "oid"); |
| i_srvname = PQfnumber(res, "srvname"); |
| i_srvowner = PQfnumber(res, "srvowner"); |
| i_srvfdw = PQfnumber(res, "srvfdw"); |
| i_srvtype = PQfnumber(res, "srvtype"); |
| i_srvversion = PQfnumber(res, "srvversion"); |
| i_srvacl = PQfnumber(res, "srvacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| i_srvoptions = PQfnumber(res, "srvoptions"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| srvinfo[i].dobj.objType = DO_FOREIGN_SERVER; |
| srvinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| srvinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&srvinfo[i].dobj); |
| srvinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_srvname)); |
| srvinfo[i].dobj.namespace = NULL; |
| srvinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_srvacl)); |
| srvinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| srvinfo[i].dacl.privtype = 0; |
| srvinfo[i].dacl.initprivs = NULL; |
| srvinfo[i].rolname = getRoleName(PQgetvalue(res, i, i_srvowner)); |
| srvinfo[i].srvfdw = atooid(PQgetvalue(res, i, i_srvfdw)); |
| srvinfo[i].srvtype = pg_strdup(PQgetvalue(res, i, i_srvtype)); |
| srvinfo[i].srvversion = pg_strdup(PQgetvalue(res, i, i_srvversion)); |
| srvinfo[i].srvoptions = pg_strdup(PQgetvalue(res, i, i_srvoptions)); |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableObject(&(srvinfo[i].dobj), fout); |
| |
| <<<<<<< HEAD |
| /* Leave out the built-in gp_exttable_server */ |
| if (srvinfo[i].dobj.catId.oid < (Oid) FirstNormalObjectId) |
| srvinfo[i].dobj.dump = DUMP_COMPONENT_NONE; |
| |
| /* Do not try to dump ACL if no ACL exists. */ |
| if (PQgetisnull(res, i, i_srvacl) && PQgetisnull(res, i, i_rsrvacl) && |
| PQgetisnull(res, i, i_initsrvacl) && |
| PQgetisnull(res, i, i_initrsrvacl)) |
| srvinfo[i].dobj.dump &= ~DUMP_COMPONENT_ACL; |
| ======= |
| /* Servers have user mappings */ |
| srvinfo[i].dobj.components |= DUMP_COMPONENT_USERMAP; |
| |
| /* Mark whether server has an ACL */ |
| if (!PQgetisnull(res, i, i_srvacl)) |
| srvinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| >>>>>>> REL_16_9 |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return srvinfo; |
| } |
| |
| /* |
| * getDefaultACLs: |
| * read all default ACL information in the system catalogs and return |
| * them in the DefaultACLInfo structure |
| * |
| * numDefaultACLs is set to the number of ACLs read in |
| */ |
| DefaultACLInfo * |
| getDefaultACLs(Archive *fout, int *numDefaultACLs) |
| { |
| DumpOptions *dopt = fout->dopt; |
| DefaultACLInfo *daclinfo; |
| PQExpBuffer query; |
| PGresult *res; |
| int i_oid; |
| int i_tableoid; |
| int i_defaclrole; |
| int i_defaclnamespace; |
| int i_defaclobjtype; |
| int i_defaclacl; |
| int i_acldefault; |
| int i, |
| ntups; |
| |
| query = createPQExpBuffer(); |
| |
| <<<<<<< HEAD |
| if (fout->remoteVersion >= 90600) |
| { |
| PQExpBuffer acl_subquery = createPQExpBuffer(); |
| PQExpBuffer racl_subquery = createPQExpBuffer(); |
| PQExpBuffer initacl_subquery = createPQExpBuffer(); |
| PQExpBuffer initracl_subquery = createPQExpBuffer(); |
| |
| /* |
| * Global entries (with defaclnamespace=0) replace the hard-wired |
| * default ACL for their object type. We should dump them as deltas |
| * from the default ACL, since that will be used as a starting point |
| * for interpreting the ALTER DEFAULT PRIVILEGES commands. On the |
| * other hand, non-global entries can only add privileges not revoke |
| * them. We must dump those as-is (i.e., as deltas from an empty |
| * ACL). We implement that by passing NULL as the object type for |
| * acldefault(), which works because acldefault() is STRICT. |
| * |
| * We can use defaclobjtype as the object type for acldefault(), |
| * except for the case of 'S' (DEFACLOBJ_SEQUENCE) which must be |
| * converted to 's'. |
| */ |
| buildACLQueries(acl_subquery, racl_subquery, initacl_subquery, |
| initracl_subquery, "defaclacl", "defaclrole", |
| "CASE WHEN defaclnamespace = 0 THEN" |
| " CASE WHEN defaclobjtype = 'S' THEN 's'::\"char\"" |
| " ELSE defaclobjtype END " |
| "ELSE NULL END", |
| dopt->binary_upgrade); |
| |
| appendPQExpBuffer(query, "SELECT d.oid, d.tableoid, " |
| "(%s d.defaclrole) AS defaclrole, " |
| "d.defaclnamespace, " |
| "d.defaclobjtype, " |
| "%s AS defaclacl, " |
| "%s AS rdefaclacl, " |
| "%s AS initdefaclacl, " |
| "%s AS initrdefaclacl " |
| "FROM pg_default_acl d " |
| "LEFT JOIN pg_init_privs pip ON " |
| "(d.oid = pip.objoid " |
| "AND pip.classoid = 'pg_default_acl'::regclass " |
| "AND pip.objsubid = 0) ", |
| username_subquery, |
| acl_subquery->data, |
| racl_subquery->data, |
| initacl_subquery->data, |
| initracl_subquery->data); |
| |
| destroyPQExpBuffer(acl_subquery); |
| destroyPQExpBuffer(racl_subquery); |
| destroyPQExpBuffer(initacl_subquery); |
| destroyPQExpBuffer(initracl_subquery); |
| } |
| else |
| { |
| appendPQExpBuffer(query, "SELECT oid, tableoid, " |
| "(%s defaclrole) AS defaclrole, " |
| "defaclnamespace, " |
| "defaclobjtype, " |
| "defaclacl, " |
| "NULL AS rdefaclacl, " |
| "NULL AS initdefaclacl, " |
| "NULL AS initrdefaclacl " |
| "FROM pg_default_acl", |
| username_subquery); |
| } |
| ======= |
| /* |
| * Global entries (with defaclnamespace=0) replace the hard-wired default |
| * ACL for their object type. We should dump them as deltas from the |
| * default ACL, since that will be used as a starting point for |
| * interpreting the ALTER DEFAULT PRIVILEGES commands. On the other hand, |
| * non-global entries can only add privileges not revoke them. We must |
| * dump those as-is (i.e., as deltas from an empty ACL). |
| * |
| * We can use defaclobjtype as the object type for acldefault(), except |
| * for the case of 'S' (DEFACLOBJ_SEQUENCE) which must be converted to |
| * 's'. |
| */ |
| appendPQExpBufferStr(query, |
| "SELECT oid, tableoid, " |
| "defaclrole, " |
| "defaclnamespace, " |
| "defaclobjtype, " |
| "defaclacl, " |
| "CASE WHEN defaclnamespace = 0 THEN " |
| "acldefault(CASE WHEN defaclobjtype = 'S' " |
| "THEN 's'::\"char\" ELSE defaclobjtype END, " |
| "defaclrole) ELSE '{}' END AS acldefault " |
| "FROM pg_default_acl"); |
| >>>>>>> REL_16_9 |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| *numDefaultACLs = ntups; |
| |
| daclinfo = (DefaultACLInfo *) pg_malloc(ntups * sizeof(DefaultACLInfo)); |
| |
| i_oid = PQfnumber(res, "oid"); |
| i_tableoid = PQfnumber(res, "tableoid"); |
| i_defaclrole = PQfnumber(res, "defaclrole"); |
| i_defaclnamespace = PQfnumber(res, "defaclnamespace"); |
| i_defaclobjtype = PQfnumber(res, "defaclobjtype"); |
| i_defaclacl = PQfnumber(res, "defaclacl"); |
| i_acldefault = PQfnumber(res, "acldefault"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| Oid nspid = atooid(PQgetvalue(res, i, i_defaclnamespace)); |
| |
| daclinfo[i].dobj.objType = DO_DEFAULT_ACL; |
| daclinfo[i].dobj.catId.tableoid = atooid(PQgetvalue(res, i, i_tableoid)); |
| daclinfo[i].dobj.catId.oid = atooid(PQgetvalue(res, i, i_oid)); |
| AssignDumpId(&daclinfo[i].dobj); |
| /* cheesy ... is it worth coming up with a better object name? */ |
| daclinfo[i].dobj.name = pg_strdup(PQgetvalue(res, i, i_defaclobjtype)); |
| |
| if (nspid != InvalidOid) |
| daclinfo[i].dobj.namespace = findNamespace(nspid); |
| else |
| daclinfo[i].dobj.namespace = NULL; |
| |
| daclinfo[i].dacl.acl = pg_strdup(PQgetvalue(res, i, i_defaclacl)); |
| daclinfo[i].dacl.acldefault = pg_strdup(PQgetvalue(res, i, i_acldefault)); |
| daclinfo[i].dacl.privtype = 0; |
| daclinfo[i].dacl.initprivs = NULL; |
| daclinfo[i].defaclrole = getRoleName(PQgetvalue(res, i, i_defaclrole)); |
| daclinfo[i].defaclobjtype = *(PQgetvalue(res, i, i_defaclobjtype)); |
| |
| /* Default ACLs are ACLs, of course */ |
| daclinfo[i].dobj.components |= DUMP_COMPONENT_ACL; |
| |
| /* Decide whether we want to dump it */ |
| selectDumpableDefaultACL(&(daclinfo[i]), dopt); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| |
| return daclinfo; |
| } |
| |
| /* |
| * getRoleName -- look up the name of a role, given its OID |
| * |
| * In current usage, we don't expect failures, so error out for a bad OID. |
| */ |
| static const char * |
| getRoleName(const char *roleoid_str) |
| { |
| Oid roleoid = atooid(roleoid_str); |
| |
| /* |
| * Do binary search to find the appropriate item. |
| */ |
| if (nrolenames > 0) |
| { |
| RoleNameItem *low = &rolenames[0]; |
| RoleNameItem *high = &rolenames[nrolenames - 1]; |
| |
| while (low <= high) |
| { |
| RoleNameItem *middle = low + (high - low) / 2; |
| |
| if (roleoid < middle->roleoid) |
| high = middle - 1; |
| else if (roleoid > middle->roleoid) |
| low = middle + 1; |
| else |
| return middle->rolename; /* found a match */ |
| } |
| } |
| |
| pg_fatal("role with OID %u does not exist", roleoid); |
| return NULL; /* keep compiler quiet */ |
| } |
| |
| /* |
| * collectRoleNames -- |
| * |
| * Construct a table of all known roles. |
| * The table is sorted by OID for speed in lookup. |
| */ |
| static void |
| collectRoleNames(Archive *fout) |
| { |
| PGresult *res; |
| const char *query; |
| int i; |
| |
| query = "SELECT oid, rolname FROM pg_catalog.pg_roles ORDER BY 1"; |
| |
| res = ExecuteSqlQuery(fout, query, PGRES_TUPLES_OK); |
| |
| nrolenames = PQntuples(res); |
| |
| rolenames = (RoleNameItem *) pg_malloc(nrolenames * sizeof(RoleNameItem)); |
| |
| for (i = 0; i < nrolenames; i++) |
| { |
| rolenames[i].roleoid = atooid(PQgetvalue(res, i, 0)); |
| rolenames[i].rolename = pg_strdup(PQgetvalue(res, i, 1)); |
| } |
| |
| PQclear(res); |
| } |
| |
| /* |
| * getAdditionalACLs |
| * |
| * We have now created all the DumpableObjects, and collected the ACL data |
| * that appears in the directly-associated catalog entries. However, there's |
| * more ACL-related info to collect. If any of a table's columns have ACLs, |
| * we must set the TableInfo's DUMP_COMPONENT_ACL components flag, as well as |
| * its hascolumnACLs flag (we won't store the ACLs themselves here, though). |
| * Also, in versions having the pg_init_privs catalog, read that and load the |
| * information into the relevant DumpableObjects. |
| */ |
| static void |
| getAdditionalACLs(Archive *fout) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| int ntups, |
| i; |
| |
| /* Check for per-column ACLs */ |
| appendPQExpBufferStr(query, |
| "SELECT DISTINCT attrelid FROM pg_attribute " |
| "WHERE attacl IS NOT NULL"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| for (i = 0; i < ntups; i++) |
| { |
| Oid relid = atooid(PQgetvalue(res, i, 0)); |
| TableInfo *tblinfo; |
| |
| tblinfo = findTableByOid(relid); |
| /* OK to ignore tables we haven't got a DumpableObject for */ |
| if (tblinfo) |
| { |
| tblinfo->dobj.components |= DUMP_COMPONENT_ACL; |
| tblinfo->hascolumnACLs = true; |
| } |
| } |
| PQclear(res); |
| |
| /* Fetch initial-privileges data */ |
| if (fout->remoteVersion >= 90600) |
| { |
| printfPQExpBuffer(query, |
| "SELECT objoid, classoid, objsubid, privtype, initprivs " |
| "FROM pg_init_privs"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| for (i = 0; i < ntups; i++) |
| { |
| Oid objoid = atooid(PQgetvalue(res, i, 0)); |
| Oid classoid = atooid(PQgetvalue(res, i, 1)); |
| int objsubid = atoi(PQgetvalue(res, i, 2)); |
| char privtype = *(PQgetvalue(res, i, 3)); |
| char *initprivs = PQgetvalue(res, i, 4); |
| CatalogId objId; |
| DumpableObject *dobj; |
| |
| objId.tableoid = classoid; |
| objId.oid = objoid; |
| dobj = findObjectByCatalogId(objId); |
| /* OK to ignore entries we haven't got a DumpableObject for */ |
| if (dobj) |
| { |
| /* Cope with sub-object initprivs */ |
| if (objsubid != 0) |
| { |
| if (dobj->objType == DO_TABLE) |
| { |
| /* For a column initprivs, set the table's ACL flags */ |
| dobj->components |= DUMP_COMPONENT_ACL; |
| ((TableInfo *) dobj)->hascolumnACLs = true; |
| } |
| else |
| pg_log_warning("unsupported pg_init_privs entry: %u %u %d", |
| classoid, objoid, objsubid); |
| continue; |
| } |
| |
| /* |
| * We ignore any pg_init_privs.initprivs entry for the public |
| * schema, as explained in getNamespaces(). |
| */ |
| if (dobj->objType == DO_NAMESPACE && |
| strcmp(dobj->name, "public") == 0) |
| continue; |
| |
| /* Else it had better be of a type we think has ACLs */ |
| if (dobj->objType == DO_NAMESPACE || |
| dobj->objType == DO_TYPE || |
| dobj->objType == DO_FUNC || |
| dobj->objType == DO_AGG || |
| dobj->objType == DO_TABLE || |
| dobj->objType == DO_PROCLANG || |
| dobj->objType == DO_FDW || |
| dobj->objType == DO_FOREIGN_SERVER) |
| { |
| DumpableObjectWithAcl *daobj = (DumpableObjectWithAcl *) dobj; |
| |
| daobj->dacl.privtype = privtype; |
| daobj->dacl.initprivs = pstrdup(initprivs); |
| } |
| else |
| pg_log_warning("unsupported pg_init_privs entry: %u %u %d", |
| classoid, objoid, objsubid); |
| } |
| } |
| PQclear(res); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpCommentExtended -- |
| * |
| * This routine is used to dump any comments associated with the |
| * object handed to this routine. The routine takes the object type |
| * and object name (ready to print, except for schema decoration), plus |
| * the namespace and owner of the object (for labeling the ArchiveEntry), |
| * plus catalog ID and subid which are the lookup key for pg_description, |
| * plus the dump ID for the object (for setting a dependency). |
| * If a matching pg_description entry is found, it is dumped. |
| * |
| * Note: in some cases, such as comments for triggers and rules, the "type" |
| * string really looks like, e.g., "TRIGGER name ON". This is a bit of a hack |
| * but it doesn't seem worth complicating the API for all callers to make |
| * it cleaner. |
| * |
| * Note: although this routine takes a dumpId for dependency purposes, |
| * that purpose is just to mark the dependency in the emitted dump file |
| * for possible future use by pg_restore. We do NOT use it for determining |
| * ordering of the comment in the dump file, because this routine is called |
| * after dependency sorting occurs. This routine should be called just after |
| * calling ArchiveEntry() for the specified object. |
| */ |
| static void |
| dumpCommentExtended(Archive *fout, const char *type, |
| const char *name, const char *namespace, |
| const char *owner, CatalogId catalogId, |
| int subid, DumpId dumpId, |
| const char *initdb_comment) |
| { |
| DumpOptions *dopt = fout->dopt; |
| CommentItem *comments; |
| int ncomments; |
| |
| /* do nothing, if --no-comments is supplied */ |
| if (dopt->no_comments) |
| return; |
| |
| /* Comments are schema not data ... except LO comments are data */ |
| if (strcmp(type, "LARGE OBJECT") != 0) |
| { |
| if (dopt->dataOnly) |
| return; |
| } |
| else |
| { |
| /* We do dump LO comments in binary-upgrade mode */ |
| if (dopt->schemaOnly && !dopt->binary_upgrade) |
| return; |
| } |
| |
| /* Search for comments associated with catalogId, using table */ |
| ncomments = findComments(catalogId.tableoid, catalogId.oid, |
| &comments); |
| |
| /* Is there one matching the subid? */ |
| while (ncomments > 0) |
| { |
| if (comments->objsubid == subid) |
| break; |
| comments++; |
| ncomments--; |
| } |
| |
| if (initdb_comment != NULL) |
| { |
| static CommentItem empty_comment = {.descr = ""}; |
| |
| /* |
| * initdb creates this object with a comment. Skip dumping the |
| * initdb-provided comment, which would complicate matters for |
| * non-superuser use of pg_dump. When the DBA has removed initdb's |
| * comment, replicate that. |
| */ |
| if (ncomments == 0) |
| { |
| comments = &empty_comment; |
| ncomments = 1; |
| } |
| else if (strcmp(comments->descr, initdb_comment) == 0) |
| ncomments = 0; |
| } |
| |
| /* If a comment exists, build COMMENT ON statement */ |
| if (ncomments > 0) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PQExpBuffer tag = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "COMMENT ON %s ", type); |
| if (namespace && *namespace) |
| appendPQExpBuffer(query, "%s.", fmtId(namespace)); |
| appendPQExpBuffer(query, "%s IS ", name); |
| appendStringLiteralAH(query, comments->descr, fout); |
| appendPQExpBufferStr(query, ";\n"); |
| |
| appendPQExpBuffer(tag, "%s %s", type, name); |
| |
| /* |
| * We mark comments as SECTION_NONE because they really belong in the |
| * same section as their parent, whether that is pre-data or |
| * post-data. |
| */ |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = namespace, |
| .owner = owner, |
| .description = "COMMENT", |
| .section = SECTION_NONE, |
| .createStmt = query->data, |
| .deps = &dumpId, |
| .nDeps = 1)); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(tag); |
| } |
| } |
| |
| /* |
| * dumpComment -- |
| * |
| * Typical simplification of the above function. |
| */ |
| static inline void |
| dumpComment(Archive *fout, const char *type, |
| const char *name, const char *namespace, |
| const char *owner, CatalogId catalogId, |
| int subid, DumpId dumpId) |
| { |
| dumpCommentExtended(fout, type, name, namespace, owner, |
| catalogId, subid, dumpId, NULL); |
| } |
| |
| /* |
| * dumpTableComment -- |
| * |
| * As above, but dump comments for both the specified table (or view) |
| * and its columns. |
| */ |
| static void |
| dumpTableComment(Archive *fout, const TableInfo *tbinfo, |
| const char *reltypename) |
| { |
| DumpOptions *dopt = fout->dopt; |
| CommentItem *comments; |
| int ncomments; |
| PQExpBuffer query; |
| PQExpBuffer tag; |
| |
| /* do nothing, if --no-comments is supplied */ |
| if (dopt->no_comments) |
| return; |
| |
| /* Comments are SCHEMA not data */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* Search for comments associated with relation, using table */ |
| ncomments = findComments(tbinfo->dobj.catId.tableoid, |
| tbinfo->dobj.catId.oid, |
| &comments); |
| |
| /* If comments exist, build COMMENT ON statements */ |
| if (ncomments <= 0) |
| return; |
| |
| query = createPQExpBuffer(); |
| tag = createPQExpBuffer(); |
| |
| while (ncomments > 0) |
| { |
| const char *descr = comments->descr; |
| int objsubid = comments->objsubid; |
| |
| if (objsubid == 0) |
| { |
| resetPQExpBuffer(tag); |
| if (strcmp(reltypename, "EXTERNAL TABLE") == 0) |
| reltypename = "TABLE"; |
| appendPQExpBuffer(tag, "%s %s.", reltypename, |
| fmtId(tbinfo->dobj.namespace->dobj.name)); |
| appendPQExpBuffer(tag, "%s ", fmtId(tbinfo->dobj.name)); |
| |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, "COMMENT ON %s %s IS ", reltypename, |
| fmtQualifiedDumpable(tbinfo)); |
| appendStringLiteralAH(query, descr, fout); |
| appendPQExpBufferStr(query, ";\n"); |
| |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "COMMENT", |
| .section = SECTION_NONE, |
| .createStmt = query->data, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1)); |
| } |
| else if (objsubid > 0 && objsubid <= tbinfo->numatts) |
| { |
| resetPQExpBuffer(tag); |
| appendPQExpBuffer(tag, "COLUMN %s.", |
| fmtId(tbinfo->dobj.name)); |
| appendPQExpBufferStr(tag, fmtId(tbinfo->attnames[objsubid - 1])); |
| |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, "COMMENT ON COLUMN %s.", |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(query, "%s IS ", |
| fmtId(tbinfo->attnames[objsubid - 1])); |
| appendStringLiteralAH(query, descr, fout); |
| appendPQExpBufferStr(query, ";\n"); |
| |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "COMMENT", |
| .section = SECTION_NONE, |
| .createStmt = query->data, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1)); |
| } |
| |
| comments++; |
| ncomments--; |
| } |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(tag); |
| } |
| |
| /* |
| * findComments -- |
| * |
| * Find the comment(s), if any, associated with the given object. All the |
| * objsubid values associated with the given classoid/objoid are found with |
| * one search. |
| */ |
| static int |
| findComments(Oid classoid, Oid objoid, CommentItem **items) |
| { |
| CommentItem *middle = NULL; |
| CommentItem *low; |
| CommentItem *high; |
| int nmatch; |
| |
| /* |
| * Do binary search to find some item matching the object. |
| */ |
| low = &comments[0]; |
| high = &comments[ncomments - 1]; |
| while (low <= high) |
| { |
| middle = low + (high - low) / 2; |
| |
| if (classoid < middle->classoid) |
| high = middle - 1; |
| else if (classoid > middle->classoid) |
| low = middle + 1; |
| else if (objoid < middle->objoid) |
| high = middle - 1; |
| else if (objoid > middle->objoid) |
| low = middle + 1; |
| else |
| break; /* found a match */ |
| } |
| |
| if (low > high) /* no matches */ |
| { |
| *items = NULL; |
| return 0; |
| } |
| |
| /* |
| * Now determine how many items match the object. The search loop |
| * invariant still holds: only items between low and high inclusive could |
| * match. |
| */ |
| nmatch = 1; |
| while (middle > low) |
| { |
| if (classoid != middle[-1].classoid || |
| objoid != middle[-1].objoid) |
| break; |
| middle--; |
| nmatch++; |
| } |
| |
| *items = middle; |
| |
| middle += nmatch; |
| while (middle <= high) |
| { |
| if (classoid != middle->classoid || |
| objoid != middle->objoid) |
| break; |
| middle++; |
| nmatch++; |
| } |
| |
| return nmatch; |
| } |
| |
| /* |
| * collectComments -- |
| * |
| * Construct a table of all comments available for database objects; |
| * also set the has-comment component flag for each relevant object. |
| * |
| * We used to do per-object queries for the comments, but it's much faster |
| * to pull them all over at once, and on most databases the memory cost |
| * isn't high. |
| * |
| * The table is sorted by classoid/objid/objsubid for speed in lookup. |
| */ |
| static void |
| collectComments(Archive *fout) |
| { |
| PGresult *res; |
| PQExpBuffer query; |
| int i_description; |
| int i_classoid; |
| int i_objoid; |
| int i_objsubid; |
| int ntups; |
| int i; |
| DumpableObject *dobj; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, "SELECT description, classoid, objoid, objsubid " |
| "FROM pg_catalog.pg_description " |
| "ORDER BY classoid, objoid, objsubid"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| /* Construct lookup table containing OIDs in numeric form */ |
| |
| i_description = PQfnumber(res, "description"); |
| i_classoid = PQfnumber(res, "classoid"); |
| i_objoid = PQfnumber(res, "objoid"); |
| i_objsubid = PQfnumber(res, "objsubid"); |
| |
| ntups = PQntuples(res); |
| |
| comments = (CommentItem *) pg_malloc(ntups * sizeof(CommentItem)); |
| ncomments = 0; |
| dobj = NULL; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| CatalogId objId; |
| int subid; |
| |
| objId.tableoid = atooid(PQgetvalue(res, i, i_classoid)); |
| objId.oid = atooid(PQgetvalue(res, i, i_objoid)); |
| subid = atoi(PQgetvalue(res, i, i_objsubid)); |
| |
| /* We needn't remember comments that don't match any dumpable object */ |
| if (dobj == NULL || |
| dobj->catId.tableoid != objId.tableoid || |
| dobj->catId.oid != objId.oid) |
| dobj = findObjectByCatalogId(objId); |
| if (dobj == NULL) |
| continue; |
| |
| /* |
| * Comments on columns of composite types are linked to the type's |
| * pg_class entry, but we need to set the DUMP_COMPONENT_COMMENT flag |
| * in the type's own DumpableObject. |
| */ |
| if (subid != 0 && dobj->objType == DO_TABLE && |
| ((TableInfo *) dobj)->relkind == RELKIND_COMPOSITE_TYPE) |
| { |
| TypeInfo *cTypeInfo; |
| |
| cTypeInfo = findTypeByOid(((TableInfo *) dobj)->reltype); |
| if (cTypeInfo) |
| cTypeInfo->dobj.components |= DUMP_COMPONENT_COMMENT; |
| } |
| else |
| dobj->components |= DUMP_COMPONENT_COMMENT; |
| |
| comments[ncomments].descr = pg_strdup(PQgetvalue(res, i, i_description)); |
| comments[ncomments].classoid = objId.tableoid; |
| comments[ncomments].objoid = objId.oid; |
| comments[ncomments].objsubid = subid; |
| ncomments++; |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpDumpableObject |
| * |
| * This routine and its subsidiaries are responsible for creating |
| * ArchiveEntries (TOC objects) for each object to be dumped. |
| */ |
| static void |
| dumpDumpableObject(Archive *fout, DumpableObject *dobj) |
| { |
| /* |
| * Clear any dump-request bits for components that don't exist for this |
| * object. (This makes it safe to initially use DUMP_COMPONENT_ALL as the |
| * request for every kind of object.) |
| */ |
| dobj->dump &= dobj->components; |
| |
| /* Now, short-circuit if there's nothing to be done here. */ |
| if (dobj->dump == 0) |
| return; |
| |
| switch (dobj->objType) |
| { |
| case DO_NAMESPACE: |
| dumpNamespace(fout, (const NamespaceInfo *) dobj); |
| break; |
| case DO_EXTENSION: |
| dumpExtension(fout, (const ExtensionInfo *) dobj); |
| break; |
| case DO_TYPE: |
| dumpType(fout, (const TypeInfo *) dobj); |
| break; |
| case DO_TYPE_STORAGE_OPTIONS: |
| dumpTypeStorageOptions(fout, (const TypeStorageOptions *) dobj); |
| break; |
| case DO_SHELL_TYPE: |
| dumpShellType(fout, (const ShellTypeInfo *) dobj); |
| break; |
| case DO_FUNC: |
| dumpFunc(fout, (const FuncInfo *) dobj); |
| break; |
| case DO_AGG: |
| dumpAgg(fout, (const AggInfo *) dobj); |
| break; |
| case DO_EXTPROTOCOL: |
| dumpExtProtocol(fout, (const ExtProtInfo *) dobj); |
| break; |
| case DO_OPERATOR: |
| dumpOpr(fout, (const OprInfo *) dobj); |
| break; |
| case DO_ACCESS_METHOD: |
| dumpAccessMethod(fout, (const AccessMethodInfo *) dobj); |
| break; |
| case DO_OPCLASS: |
| dumpOpclass(fout, (const OpclassInfo *) dobj); |
| break; |
| case DO_OPFAMILY: |
| dumpOpfamily(fout, (const OpfamilyInfo *) dobj); |
| break; |
| case DO_COLLATION: |
| dumpCollation(fout, (const CollInfo *) dobj); |
| break; |
| case DO_CONVERSION: |
| dumpConversion(fout, (const ConvInfo *) dobj); |
| break; |
| case DO_TABLE: |
| dumpTable(fout, (const TableInfo *) dobj); |
| break; |
| case DO_TABLE_ATTACH: |
| /* GPDB_14_MERGE_FIXME: why some cases don't check postDataSchemaOnly ? */ |
| dumpTableAttach(fout, (const TableAttachInfo *) dobj); |
| break; |
| case DO_ATTRDEF: |
| dumpAttrDef(fout, (const AttrDefInfo *) dobj); |
| break; |
| case DO_INDEX: |
| dumpIndex(fout, (const IndxInfo *) dobj); |
| break; |
| case DO_INDEX_ATTACH: |
| dumpIndexAttach(fout, (const IndexAttachInfo *) dobj); |
| break; |
| case DO_STATSEXT: |
| dumpStatisticsExt(fout, (const StatsExtInfo *) dobj); |
| break; |
| case DO_REFRESH_MATVIEW: |
| refreshMatViewData(fout, (const TableDataInfo *) dobj); |
| break; |
| case DO_RULE: |
| dumpRule(fout, (const RuleInfo *) dobj); |
| break; |
| case DO_TRIGGER: |
| dumpTrigger(fout, (const TriggerInfo *) dobj); |
| break; |
| case DO_EVENT_TRIGGER: |
| dumpEventTrigger(fout, (const EventTriggerInfo *) dobj); |
| break; |
| case DO_CONSTRAINT: |
| dumpConstraint(fout, (const ConstraintInfo *) dobj); |
| break; |
| case DO_FK_CONSTRAINT: |
| dumpConstraint(fout, (const ConstraintInfo *) dobj); |
| break; |
| case DO_PROCLANG: |
| dumpProcLang(fout, (const ProcLangInfo *) dobj); |
| break; |
| case DO_CAST: |
| dumpCast(fout, (const CastInfo *) dobj); |
| break; |
| case DO_TRANSFORM: |
| dumpTransform(fout, (const TransformInfo *) dobj); |
| break; |
| case DO_SEQUENCE_SET: |
| dumpSequenceData(fout, (const TableDataInfo *) dobj); |
| break; |
| case DO_TABLE_DATA: |
| dumpTableData(fout, (const TableDataInfo *) dobj); |
| break; |
| case DO_DUMMY_TYPE: |
| /* table rowtypes and array types are never dumped separately */ |
| break; |
| case DO_TSPARSER: |
| dumpTSParser(fout, (const TSParserInfo *) dobj); |
| break; |
| case DO_TSDICT: |
| dumpTSDictionary(fout, (const TSDictInfo *) dobj); |
| break; |
| case DO_TSTEMPLATE: |
| dumpTSTemplate(fout, (const TSTemplateInfo *) dobj); |
| break; |
| case DO_TSCONFIG: |
| dumpTSConfig(fout, (const TSConfigInfo *) dobj); |
| break; |
| case DO_FDW: |
| dumpForeignDataWrapper(fout, (const FdwInfo *) dobj); |
| break; |
| case DO_FOREIGN_SERVER: |
| dumpForeignServer(fout, (const ForeignServerInfo *) dobj); |
| break; |
| case DO_DEFAULT_ACL: |
| dumpDefaultACL(fout, (const DefaultACLInfo *) dobj); |
| break; |
| case DO_LARGE_OBJECT: |
| dumpLO(fout, (const LoInfo *) dobj); |
| break; |
| case DO_LARGE_OBJECT_DATA: |
| if (dobj->dump & DUMP_COMPONENT_DATA) |
| { |
| TocEntry *te; |
| |
| te = ArchiveEntry(fout, dobj->catId, dobj->dumpId, |
| ARCHIVE_OPTS(.tag = dobj->name, |
| .description = "BLOBS", |
| .section = SECTION_DATA, |
| .dumpFn = dumpLOs)); |
| |
| /* |
| * Set the TocEntry's dataLength in case we are doing a |
| * parallel dump and want to order dump jobs by table size. |
| * (We need some size estimate for every TocEntry with a |
| * DataDumper function.) We don't currently have any cheap |
| * way to estimate the size of LOs, but it doesn't matter; |
| * let's just set the size to a large value so parallel dumps |
| * will launch this job first. If there's lots of LOs, we |
| * win, and if there aren't, we don't lose much. (If you want |
| * to improve on this, really what you should be thinking |
| * about is allowing LO dumping to be parallelized, not just |
| * getting a smarter estimate for the single TOC entry.) |
| */ |
| te->dataLength = INT_MAX; |
| } |
| break; |
| case DO_POLICY: |
| dumpPolicy(fout, (const PolicyInfo *) dobj); |
| break; |
| case DO_PUBLICATION: |
| dumpPublication(fout, (const PublicationInfo *) dobj); |
| break; |
| case DO_PUBLICATION_REL: |
| dumpPublicationTable(fout, (const PublicationRelInfo *) dobj); |
| break; |
| case DO_PUBLICATION_TABLE_IN_SCHEMA: |
| dumpPublicationNamespace(fout, |
| (const PublicationSchemaInfo *) dobj); |
| break; |
| case DO_SUBSCRIPTION: |
| dumpSubscription(fout, (const SubscriptionInfo *) dobj); |
| break; |
| case DO_PRE_DATA_BOUNDARY: |
| case DO_POST_DATA_BOUNDARY: |
| /* never dumped, nothing to do */ |
| break; |
| |
| case DO_BINARY_UPGRADE: |
| dumpPreassignedOidArchiveEntry(fout, (BinaryUpgradeInfo *) dobj); |
| break; |
| } |
| } |
| |
| /* |
| * dumpNamespace |
| * writes out to fout the queries to recreate a user-defined namespace |
| */ |
| static void |
| dumpNamespace(Archive *fout, const NamespaceInfo *nspinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qnspname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qnspname = pg_strdup(fmtId(nspinfo->dobj.name)); |
| |
| <<<<<<< HEAD |
| appendPQExpBuffer(delq, "DROP SCHEMA %s;\n", qnspname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_namespace_oid(fout, q, nspinfo->dobj.catId.oid); |
| |
| appendPQExpBuffer(q, "CREATE SCHEMA %s;\n", qnspname); |
| ======= |
| if (nspinfo->create) |
| { |
| appendPQExpBuffer(delq, "DROP SCHEMA %s;\n", qnspname); |
| appendPQExpBuffer(q, "CREATE SCHEMA %s;\n", qnspname); |
| } |
| else |
| { |
| /* see selectDumpableNamespace() */ |
| appendPQExpBufferStr(delq, |
| "-- *not* dropping schema, since initdb creates it\n"); |
| appendPQExpBufferStr(q, |
| "-- *not* creating schema, since initdb creates it\n"); |
| } |
| >>>>>>> REL_16_9 |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &nspinfo->dobj, |
| "SCHEMA", qnspname, NULL); |
| |
| if (nspinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, nspinfo->dobj.catId, nspinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = nspinfo->dobj.name, |
| .owner = nspinfo->rolname, |
| .description = "SCHEMA", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Schema Comments and Security Labels */ |
| if (nspinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| { |
| const char *initdb_comment = NULL; |
| |
| if (!nspinfo->create && strcmp(qnspname, "public") == 0) |
| initdb_comment = "standard public schema"; |
| dumpCommentExtended(fout, "SCHEMA", qnspname, |
| NULL, nspinfo->rolname, |
| nspinfo->dobj.catId, 0, nspinfo->dobj.dumpId, |
| initdb_comment); |
| } |
| |
| if (nspinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "SCHEMA", qnspname, |
| NULL, nspinfo->rolname, |
| nspinfo->dobj.catId, 0, nspinfo->dobj.dumpId); |
| |
| if (nspinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, nspinfo->dobj.dumpId, InvalidDumpId, "SCHEMA", |
| qnspname, NULL, NULL, |
| nspinfo->rolname, &nspinfo->dacl); |
| |
| free(qnspname); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| } |
| |
| /* |
| * dumpExtension |
| * writes out to fout the queries to recreate an extension |
| */ |
| static void |
| dumpExtension(Archive *fout, const ExtensionInfo *extinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qextname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qextname = pg_strdup(fmtId(extinfo->dobj.name)); |
| |
| appendPQExpBuffer(delq, "DROP EXTENSION %s;\n", qextname); |
| |
| if (!dopt->binary_upgrade) |
| { |
| /* |
| * In a regular dump, we simply create the extension, intentionally |
| * not specifying a version, so that the destination installation's |
| * default version is used. |
| * |
| * Use of IF NOT EXISTS here is unlike our behavior for other object |
| * types; but there are various scenarios in which it's convenient to |
| * manually create the desired extension before restoring, so we |
| * prefer to allow it to exist already. |
| */ |
| appendPQExpBuffer(q, "CREATE EXTENSION IF NOT EXISTS %s WITH SCHEMA %s;\n", |
| qextname, fmtId(extinfo->namespace)); |
| } |
| else |
| { |
| /* |
| * In binary-upgrade mode, it's critical to reproduce the state of the |
| * database exactly, so our procedure is to create an empty extension, |
| * restore all the contained objects normally, and add them to the |
| * extension one by one. This function performs just the first of |
| * those steps. binary_upgrade_extension_member() takes care of |
| * adding member objects as they're created. |
| */ |
| int i; |
| int n; |
| |
| appendPQExpBufferStr(q, "-- For binary upgrade, create an empty extension and insert objects into it\n"); |
| |
| /* |
| * We unconditionally create the extension, so we must drop it if it |
| * exists. This could happen if the user deleted 'plpgsql' and then |
| * readded it, causing its oid to be greater than g_last_builtin_oid. |
| */ |
| appendPQExpBuffer(q, "DROP EXTENSION IF EXISTS %s;\n", qextname); |
| |
| appendPQExpBufferStr(q, |
| "SELECT pg_catalog.binary_upgrade_create_empty_extension("); |
| appendStringLiteralAH(q, extinfo->dobj.name, fout); |
| appendPQExpBufferStr(q, ", "); |
| appendStringLiteralAH(q, extinfo->namespace, fout); |
| appendPQExpBufferStr(q, ", "); |
| appendPQExpBuffer(q, "%s, ", extinfo->relocatable ? "true" : "false"); |
| appendStringLiteralAH(q, extinfo->extversion, fout); |
| appendPQExpBufferStr(q, ", "); |
| |
| /* |
| * Note that we're pushing extconfig (an OID array) back into |
| * pg_extension exactly as-is. This is OK because pg_class OIDs are |
| * preserved in binary upgrade. |
| */ |
| if (strlen(extinfo->extconfig) > 2) |
| appendStringLiteralAH(q, extinfo->extconfig, fout); |
| else |
| appendPQExpBufferStr(q, "NULL"); |
| appendPQExpBufferStr(q, ", "); |
| if (strlen(extinfo->extcondition) > 2) |
| appendStringLiteralAH(q, extinfo->extcondition, fout); |
| else |
| appendPQExpBufferStr(q, "NULL"); |
| appendPQExpBufferStr(q, ", "); |
| appendPQExpBufferStr(q, "ARRAY["); |
| n = 0; |
| for (i = 0; i < extinfo->dobj.nDeps; i++) |
| { |
| DumpableObject *extobj; |
| |
| extobj = findObjectByDumpId(extinfo->dobj.dependencies[i]); |
| if (extobj && extobj->objType == DO_EXTENSION) |
| { |
| if (n++ > 0) |
| appendPQExpBufferChar(q, ','); |
| appendStringLiteralAH(q, extobj->name, fout); |
| } |
| } |
| appendPQExpBufferStr(q, "]::pg_catalog.text[]"); |
| appendPQExpBufferStr(q, ");\n"); |
| } |
| |
| if (extinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, extinfo->dobj.catId, extinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = extinfo->dobj.name, |
| .description = "EXTENSION", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Extension Comments and Security Labels */ |
| if (extinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "EXTENSION", qextname, |
| NULL, "", |
| extinfo->dobj.catId, 0, extinfo->dobj.dumpId); |
| |
| if (extinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "EXTENSION", qextname, |
| NULL, "", |
| extinfo->dobj.catId, 0, extinfo->dobj.dumpId); |
| |
| free(qextname); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| } |
| |
| /* |
| * dumpType |
| * writes out to fout the queries to recreate a user-defined type |
| */ |
| static void |
| dumpType(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* Dump out in proper style */ |
| if (tyinfo->typtype == TYPTYPE_BASE) |
| dumpBaseType(fout, tyinfo); |
| else if (tyinfo->typtype == TYPTYPE_DOMAIN) |
| dumpDomain(fout, tyinfo); |
| else if (tyinfo->typtype == TYPTYPE_COMPOSITE) |
| dumpCompositeType(fout, tyinfo); |
| else if (tyinfo->typtype == TYPTYPE_ENUM) |
| dumpEnumType(fout, tyinfo); |
| else if (tyinfo->typtype == TYPTYPE_RANGE) |
| dumpRangeType(fout, tyinfo); |
| else if (tyinfo->typtype == TYPTYPE_PSEUDO && !tyinfo->isDefined) |
| dumpUndefinedType(fout, tyinfo); |
| else |
| pg_log_warning("typtype of data type \"%s\" appears to be invalid", |
| tyinfo->dobj.name); |
| } |
| |
| /* |
| * dumpEnumType |
| * writes out to fout the queries to recreate a user-defined enum type |
| */ |
| static void |
| dumpEnumType(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| int num, |
| i; |
| Oid enum_oid; |
| char *qtypname; |
| char *qualtypname; |
| char *label; |
| int i_enumlabel; |
| int i_oid; |
| |
| if (!fout->is_prepared[PREPQUERY_DUMPENUMTYPE]) |
| { |
| /* Set up query for enum-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpEnumType(pg_catalog.oid) AS\n" |
| "SELECT oid, enumlabel " |
| "FROM pg_catalog.pg_enum " |
| "WHERE enumtypid = $1 " |
| "ORDER BY enumsortorder"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPENUMTYPE] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpEnumType('%u')", |
| tyinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| num = PQntuples(res); |
| |
| qtypname = pg_strdup(fmtId(tyinfo->dobj.name)); |
| qualtypname = pg_strdup(fmtQualifiedDumpable(tyinfo)); |
| |
| /* |
| * CASCADE shouldn't be required here as for normal types since the I/O |
| * functions are generic and do not get dropped. |
| */ |
| appendPQExpBuffer(delq, "DROP TYPE %s;\n", qualtypname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| tyinfo->dobj.catId.oid, |
| tyinfo->dobj.namespace->dobj.catId.oid, |
| tyinfo->dobj.name, |
| false, false); |
| |
| appendPQExpBuffer(q, "CREATE TYPE %s AS ENUM (", |
| qualtypname); |
| |
| if (!dopt->binary_upgrade) |
| { |
| i_enumlabel = PQfnumber(res, "enumlabel"); |
| |
| /* Labels with server-assigned oids */ |
| for (i = 0; i < num; i++) |
| { |
| label = PQgetvalue(res, i, i_enumlabel); |
| if (i > 0) |
| appendPQExpBufferChar(q, ','); |
| appendPQExpBufferStr(q, "\n "); |
| appendStringLiteralAH(q, label, fout); |
| } |
| } |
| |
| appendPQExpBufferStr(q, "\n);\n"); |
| |
| if (dopt->binary_upgrade) |
| { |
| i_oid = PQfnumber(res, "oid"); |
| i_enumlabel = PQfnumber(res, "enumlabel"); |
| |
| /* Labels with dump-assigned (preserved) oids */ |
| for (i = 0; i < num; i++) |
| { |
| enum_oid = atooid(PQgetvalue(res, i, i_oid)); |
| label = PQgetvalue(res, i, i_enumlabel); |
| |
| if (i == 0) |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, must preserve pg_enum oids\n"); |
| appendPQExpBuffer(q, |
| "SELECT pg_catalog.binary_upgrade_set_next_pg_enum_oid('%u'::pg_catalog.oid, '%u'::pg_catalog.oid, $$%s$$::text);\n", |
| enum_oid, tyinfo->dobj.catId.oid, label); |
| appendPQExpBuffer(q, "ALTER TYPE %s ADD VALUE ", qualtypname); |
| appendStringLiteralAH(q, label, fout); |
| appendPQExpBufferStr(q, ";\n\n"); |
| } |
| } |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tyinfo->dobj, |
| "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tyinfo->dobj.catId, tyinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tyinfo->dobj.name, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "TYPE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Type Comments and Security Labels */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, tyinfo->dobj.dumpId, InvalidDumpId, "TYPE", |
| qtypname, NULL, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, &tyinfo->dacl); |
| |
| PQclear(res); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qtypname); |
| free(qualtypname); |
| } |
| |
| /* |
| * dumpRangeType |
| * writes out to fout the queries to recreate a user-defined range type |
| */ |
| static void |
| dumpRangeType(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| Oid collationOid; |
| char *qtypname; |
| char *qualtypname; |
| char *procname; |
| |
| if (!fout->is_prepared[PREPQUERY_DUMPRANGETYPE]) |
| { |
| /* Set up query for range-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpRangeType(pg_catalog.oid) AS\n"); |
| |
| appendPQExpBufferStr(query, |
| "SELECT "); |
| |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(query, |
| "pg_catalog.format_type(rngmultitypid, NULL) AS rngmultitype, "); |
| else |
| appendPQExpBufferStr(query, |
| "NULL AS rngmultitype, "); |
| |
| appendPQExpBufferStr(query, |
| "pg_catalog.format_type(rngsubtype, NULL) AS rngsubtype, " |
| "opc.opcname AS opcname, " |
| "(SELECT nspname FROM pg_catalog.pg_namespace nsp " |
| " WHERE nsp.oid = opc.opcnamespace) AS opcnsp, " |
| "opc.opcdefault, " |
| "CASE WHEN rngcollation = st.typcollation THEN 0 " |
| " ELSE rngcollation END AS collation, " |
| "rngcanonical, rngsubdiff " |
| "FROM pg_catalog.pg_range r, pg_catalog.pg_type st, " |
| " pg_catalog.pg_opclass opc " |
| "WHERE st.oid = rngsubtype AND opc.oid = rngsubopc AND " |
| "rngtypid = $1"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPRANGETYPE] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpRangeType('%u')", |
| tyinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| qtypname = pg_strdup(fmtId(tyinfo->dobj.name)); |
| qualtypname = pg_strdup(fmtQualifiedDumpable(tyinfo)); |
| |
| /* |
| * CASCADE shouldn't be required here as for normal types since the I/O |
| * functions are generic and do not get dropped. |
| */ |
| appendPQExpBuffer(delq, "DROP TYPE %s;\n", qualtypname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| tyinfo->dobj.catId.oid, |
| tyinfo->dobj.namespace->dobj.catId.oid, |
| tyinfo->dobj.name, |
| false, true); |
| |
| appendPQExpBuffer(q, "CREATE TYPE %s AS RANGE (", |
| qualtypname); |
| |
| appendPQExpBuffer(q, "\n subtype = %s", |
| PQgetvalue(res, 0, PQfnumber(res, "rngsubtype"))); |
| |
| if (!PQgetisnull(res, 0, PQfnumber(res, "rngmultitype"))) |
| appendPQExpBuffer(q, ",\n multirange_type_name = %s", |
| PQgetvalue(res, 0, PQfnumber(res, "rngmultitype"))); |
| |
| /* print subtype_opclass only if not default for subtype */ |
| if (PQgetvalue(res, 0, PQfnumber(res, "opcdefault"))[0] != 't') |
| { |
| char *opcname = PQgetvalue(res, 0, PQfnumber(res, "opcname")); |
| char *nspname = PQgetvalue(res, 0, PQfnumber(res, "opcnsp")); |
| |
| appendPQExpBuffer(q, ",\n subtype_opclass = %s.", |
| fmtId(nspname)); |
| appendPQExpBufferStr(q, fmtId(opcname)); |
| } |
| |
| collationOid = atooid(PQgetvalue(res, 0, PQfnumber(res, "collation"))); |
| if (OidIsValid(collationOid)) |
| { |
| CollInfo *coll = findCollationByOid(collationOid); |
| |
| if (coll) |
| appendPQExpBuffer(q, ",\n collation = %s", |
| fmtQualifiedDumpable(coll)); |
| } |
| |
| procname = PQgetvalue(res, 0, PQfnumber(res, "rngcanonical")); |
| if (strcmp(procname, "-") != 0) |
| appendPQExpBuffer(q, ",\n canonical = %s", procname); |
| |
| procname = PQgetvalue(res, 0, PQfnumber(res, "rngsubdiff")); |
| if (strcmp(procname, "-") != 0) |
| appendPQExpBuffer(q, ",\n subtype_diff = %s", procname); |
| |
| appendPQExpBufferStr(q, "\n);\n"); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tyinfo->dobj, |
| "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tyinfo->dobj.catId, tyinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tyinfo->dobj.name, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "TYPE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Type Comments and Security Labels */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, tyinfo->dobj.dumpId, InvalidDumpId, "TYPE", |
| qtypname, NULL, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, &tyinfo->dacl); |
| |
| PQclear(res); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qtypname); |
| free(qualtypname); |
| } |
| |
| /* |
| * dumpUndefinedType |
| * writes out to fout the queries to recreate a !typisdefined type |
| * |
| * This is a shell type, but we use different terminology to distinguish |
| * this case from where we have to emit a shell type definition to break |
| * circular dependencies. An undefined type shouldn't ever have anything |
| * depending on it. |
| */ |
| static void |
| dumpUndefinedType(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| char *qtypname; |
| char *qualtypname; |
| |
| qtypname = pg_strdup(fmtId(tyinfo->dobj.name)); |
| qualtypname = pg_strdup(fmtQualifiedDumpable(tyinfo)); |
| |
| appendPQExpBuffer(delq, "DROP TYPE %s;\n", qualtypname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| tyinfo->dobj.catId.oid, |
| tyinfo->dobj.namespace->dobj.catId.oid, |
| tyinfo->dobj.name, |
| false, false); |
| |
| appendPQExpBuffer(q, "CREATE TYPE %s;\n", |
| qualtypname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tyinfo->dobj, |
| "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tyinfo->dobj.catId, tyinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tyinfo->dobj.name, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "TYPE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Type Comments and Security Labels */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, tyinfo->dobj.dumpId, InvalidDumpId, "TYPE", |
| qtypname, NULL, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, &tyinfo->dacl); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qtypname); |
| free(qualtypname); |
| } |
| |
| /* |
| * dumpBaseType |
| * writes out to fout the queries to recreate a user-defined base type |
| */ |
| static void |
| dumpBaseType(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| char *qtypname; |
| char *qualtypname; |
| char *typlen; |
| char *typinput; |
| char *typoutput; |
| char *typreceive; |
| char *typsend; |
| char *typmodin; |
| char *typmodout; |
| char *typanalyze; |
| char *typsubscript; |
| Oid typreceiveoid; |
| Oid typsendoid; |
| Oid typmodinoid; |
| Oid typmodoutoid; |
| Oid typanalyzeoid; |
| Oid typsubscriptoid; |
| char *typcategory; |
| char *typispreferred; |
| char *typdelim; |
| char *typbyval; |
| char *typalign; |
| char *typstorage; |
| char *typcollatable; |
| char *typdefault; |
| bool typdefault_is_literal = false; |
| |
| if (!fout->is_prepared[PREPQUERY_DUMPBASETYPE]) |
| { |
| /* Set up query for type-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpBaseType(pg_catalog.oid) AS\n" |
| "SELECT typlen, " |
| "typinput, typoutput, typreceive, typsend, " |
| "typreceive::pg_catalog.oid AS typreceiveoid, " |
| "typsend::pg_catalog.oid AS typsendoid, " |
| "typanalyze, " |
| "typanalyze::pg_catalog.oid AS typanalyzeoid, " |
| "typdelim, typbyval, typalign, typstorage, " |
| "typmodin, typmodout, " |
| "typmodin::pg_catalog.oid AS typmodinoid, " |
| "typmodout::pg_catalog.oid AS typmodoutoid, " |
| "typcategory, typispreferred, " |
| "(typcollation <> 0) AS typcollatable, " |
| "pg_catalog.pg_get_expr(typdefaultbin, 0) AS typdefaultbin, typdefault, "); |
| |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(query, |
| "typsubscript, " |
| "typsubscript::pg_catalog.oid AS typsubscriptoid "); |
| else |
| appendPQExpBufferStr(query, |
| "'-' AS typsubscript, 0 AS typsubscriptoid "); |
| |
| appendPQExpBufferStr(query, "FROM pg_catalog.pg_type " |
| "WHERE oid = $1"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPBASETYPE] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpBaseType('%u')", |
| tyinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| typlen = PQgetvalue(res, 0, PQfnumber(res, "typlen")); |
| typinput = PQgetvalue(res, 0, PQfnumber(res, "typinput")); |
| typoutput = PQgetvalue(res, 0, PQfnumber(res, "typoutput")); |
| typreceive = PQgetvalue(res, 0, PQfnumber(res, "typreceive")); |
| typsend = PQgetvalue(res, 0, PQfnumber(res, "typsend")); |
| typmodin = PQgetvalue(res, 0, PQfnumber(res, "typmodin")); |
| typmodout = PQgetvalue(res, 0, PQfnumber(res, "typmodout")); |
| typanalyze = PQgetvalue(res, 0, PQfnumber(res, "typanalyze")); |
| typsubscript = PQgetvalue(res, 0, PQfnumber(res, "typsubscript")); |
| typreceiveoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typreceiveoid"))); |
| typsendoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typsendoid"))); |
| typmodinoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typmodinoid"))); |
| typmodoutoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typmodoutoid"))); |
| typanalyzeoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typanalyzeoid"))); |
| typsubscriptoid = atooid(PQgetvalue(res, 0, PQfnumber(res, "typsubscriptoid"))); |
| typcategory = PQgetvalue(res, 0, PQfnumber(res, "typcategory")); |
| typispreferred = PQgetvalue(res, 0, PQfnumber(res, "typispreferred")); |
| typdelim = PQgetvalue(res, 0, PQfnumber(res, "typdelim")); |
| typbyval = PQgetvalue(res, 0, PQfnumber(res, "typbyval")); |
| typalign = PQgetvalue(res, 0, PQfnumber(res, "typalign")); |
| typstorage = PQgetvalue(res, 0, PQfnumber(res, "typstorage")); |
| typcollatable = PQgetvalue(res, 0, PQfnumber(res, "typcollatable")); |
| if (!PQgetisnull(res, 0, PQfnumber(res, "typdefaultbin"))) |
| typdefault = PQgetvalue(res, 0, PQfnumber(res, "typdefaultbin")); |
| else if (!PQgetisnull(res, 0, PQfnumber(res, "typdefault"))) |
| { |
| typdefault = PQgetvalue(res, 0, PQfnumber(res, "typdefault")); |
| typdefault_is_literal = true; /* it needs quotes */ |
| } |
| else |
| typdefault = NULL; |
| |
| qtypname = pg_strdup(fmtId(tyinfo->dobj.name)); |
| qualtypname = pg_strdup(fmtQualifiedDumpable(tyinfo)); |
| |
| /* |
| * The reason we include CASCADE is that the circular dependency between |
| * the type and its I/O functions makes it impossible to drop the type any |
| * other way. |
| */ |
| appendPQExpBuffer(delq, "DROP TYPE %s CASCADE;\n", qualtypname); |
| |
| /* |
| * We might already have a shell type, but setting pg_type_oid is |
| * harmless, and in any case we'd better set the array type OID. |
| */ |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| tyinfo->dobj.catId.oid, |
| tyinfo->dobj.namespace->dobj.catId.oid, |
| tyinfo->dobj.name, |
| false, false); |
| |
| appendPQExpBuffer(q, |
| "CREATE TYPE %s (\n" |
| " INTERNALLENGTH = %s", |
| qualtypname, |
| (strcmp(typlen, "-1") == 0) ? "variable" : typlen); |
| |
| /* regproc result is sufficiently quoted already */ |
| appendPQExpBuffer(q, ",\n INPUT = %s", typinput); |
| appendPQExpBuffer(q, ",\n OUTPUT = %s", typoutput); |
| if (OidIsValid(typreceiveoid)) |
| appendPQExpBuffer(q, ",\n RECEIVE = %s", typreceive); |
| if (OidIsValid(typsendoid)) |
| appendPQExpBuffer(q, ",\n SEND = %s", typsend); |
| if (OidIsValid(typmodinoid)) |
| appendPQExpBuffer(q, ",\n TYPMOD_IN = %s", typmodin); |
| if (OidIsValid(typmodoutoid)) |
| appendPQExpBuffer(q, ",\n TYPMOD_OUT = %s", typmodout); |
| if (OidIsValid(typanalyzeoid)) |
| appendPQExpBuffer(q, ",\n ANALYZE = %s", typanalyze); |
| |
| if (strcmp(typcollatable, "t") == 0) |
| appendPQExpBufferStr(q, ",\n COLLATABLE = true"); |
| |
| if (typdefault != NULL) |
| { |
| appendPQExpBufferStr(q, ",\n DEFAULT = "); |
| if (typdefault_is_literal) |
| appendStringLiteralAH(q, typdefault, fout); |
| else |
| appendPQExpBufferStr(q, typdefault); |
| } |
| |
| if (OidIsValid(typsubscriptoid)) |
| appendPQExpBuffer(q, ",\n SUBSCRIPT = %s", typsubscript); |
| |
| if (OidIsValid(tyinfo->typelem)) |
| appendPQExpBuffer(q, ",\n ELEMENT = %s", |
| getFormattedTypeName(fout, tyinfo->typelem, |
| zeroIsError)); |
| |
| if (strcmp(typcategory, "U") != 0) |
| { |
| appendPQExpBufferStr(q, ",\n CATEGORY = "); |
| appendStringLiteralAH(q, typcategory, fout); |
| } |
| |
| if (strcmp(typispreferred, "t") == 0) |
| appendPQExpBufferStr(q, ",\n PREFERRED = true"); |
| |
| if (typdelim && strcmp(typdelim, ",") != 0) |
| { |
| appendPQExpBufferStr(q, ",\n DELIMITER = "); |
| appendStringLiteralAH(q, typdelim, fout); |
| } |
| |
| if (*typalign == TYPALIGN_CHAR) |
| appendPQExpBufferStr(q, ",\n ALIGNMENT = char"); |
| else if (*typalign == TYPALIGN_SHORT) |
| appendPQExpBufferStr(q, ",\n ALIGNMENT = int2"); |
| else if (*typalign == TYPALIGN_INT) |
| appendPQExpBufferStr(q, ",\n ALIGNMENT = int4"); |
| else if (*typalign == TYPALIGN_DOUBLE) |
| appendPQExpBufferStr(q, ",\n ALIGNMENT = double"); |
| |
| if (*typstorage == TYPSTORAGE_PLAIN) |
| appendPQExpBufferStr(q, ",\n STORAGE = plain"); |
| else if (*typstorage == TYPSTORAGE_EXTERNAL) |
| appendPQExpBufferStr(q, ",\n STORAGE = external"); |
| else if (*typstorage == TYPSTORAGE_EXTENDED) |
| appendPQExpBufferStr(q, ",\n STORAGE = extended"); |
| else if (*typstorage == TYPSTORAGE_MAIN) |
| appendPQExpBufferStr(q, ",\n STORAGE = main"); |
| |
| if (strcmp(typbyval, "t") == 0) |
| appendPQExpBufferStr(q, ",\n PASSEDBYVALUE"); |
| |
| appendPQExpBufferStr(q, "\n);\n"); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tyinfo->dobj, |
| "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tyinfo->dobj.catId, tyinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tyinfo->dobj.name, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "TYPE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Type Comments and Security Labels */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, tyinfo->dobj.dumpId, InvalidDumpId, "TYPE", |
| qtypname, NULL, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, &tyinfo->dacl); |
| |
| PQclear(res); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qtypname); |
| free(qualtypname); |
| } |
| |
| /* |
| * dumpTypeStorageOptions |
| * writes out to fout the ALTER TYPE queries to set default storage options for type |
| */ |
| static void |
| dumpTypeStorageOptions(Archive *fout, const TypeStorageOptions *tstorageoptions) |
| { |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| /* |
| * Type name is already quoted by caller using quote_ident, hence used |
| * directly here. |
| */ |
| appendPQExpBuffer(q, "ALTER TYPE %s.", |
| fmtId(tstorageoptions->dobj.namespace->dobj.name)); |
| appendPQExpBuffer(q, "%s SET DEFAULT ENCODING (%s);\n", |
| tstorageoptions->dobj.name, |
| tstorageoptions->typoptions); |
| |
| ArchiveEntry(fout, |
| tstorageoptions->dobj.catId, |
| tstorageoptions->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tstorageoptions->dobj.name, |
| .namespace = tstorageoptions->dobj.namespace->dobj.name, |
| .owner = tstorageoptions->rolname, |
| .description = "TYPE STORAGE OPTIONS", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data)); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| } |
| |
| /* |
| * dumpDomain |
| * writes out to fout the queries to recreate a user-defined domain |
| */ |
| static void |
| dumpDomain(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| int i; |
| char *qtypname; |
| char *qualtypname; |
| char *typnotnull; |
| char *typdefn; |
| char *typdefault; |
| Oid typcollation; |
| bool typdefault_is_literal = false; |
| |
| if (!fout->is_prepared[PREPQUERY_DUMPDOMAIN]) |
| { |
| /* Set up query for domain-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpDomain(pg_catalog.oid) AS\n"); |
| |
| appendPQExpBufferStr(query, "SELECT t.typnotnull, " |
| "pg_catalog.format_type(t.typbasetype, t.typtypmod) AS typdefn, " |
| "pg_catalog.pg_get_expr(t.typdefaultbin, 'pg_catalog.pg_type'::pg_catalog.regclass) AS typdefaultbin, " |
| "t.typdefault, " |
| "CASE WHEN t.typcollation <> u.typcollation " |
| "THEN t.typcollation ELSE 0 END AS typcollation " |
| "FROM pg_catalog.pg_type t " |
| "LEFT JOIN pg_catalog.pg_type u ON (t.typbasetype = u.oid) " |
| "WHERE t.oid = $1"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPDOMAIN] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpDomain('%u')", |
| tyinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| typnotnull = PQgetvalue(res, 0, PQfnumber(res, "typnotnull")); |
| typdefn = PQgetvalue(res, 0, PQfnumber(res, "typdefn")); |
| if (!PQgetisnull(res, 0, PQfnumber(res, "typdefaultbin"))) |
| typdefault = PQgetvalue(res, 0, PQfnumber(res, "typdefaultbin")); |
| else if (!PQgetisnull(res, 0, PQfnumber(res, "typdefault"))) |
| { |
| typdefault = PQgetvalue(res, 0, PQfnumber(res, "typdefault")); |
| typdefault_is_literal = true; /* it needs quotes */ |
| } |
| else |
| typdefault = NULL; |
| typcollation = atooid(PQgetvalue(res, 0, PQfnumber(res, "typcollation"))); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| tyinfo->dobj.catId.oid, |
| tyinfo->dobj.namespace->dobj.catId.oid, |
| tyinfo->dobj.name, |
| true, /* force array type */ |
| false); /* force multirange type */ |
| |
| qtypname = pg_strdup(fmtId(tyinfo->dobj.name)); |
| qualtypname = pg_strdup(fmtQualifiedDumpable(tyinfo)); |
| |
| appendPQExpBuffer(q, |
| "CREATE DOMAIN %s AS %s", |
| qualtypname, |
| typdefn); |
| |
| /* Print collation only if different from base type's collation */ |
| if (OidIsValid(typcollation)) |
| { |
| CollInfo *coll; |
| |
| coll = findCollationByOid(typcollation); |
| if (coll) |
| appendPQExpBuffer(q, " COLLATE %s", fmtQualifiedDumpable(coll)); |
| } |
| |
| if (typnotnull[0] == 't') |
| appendPQExpBufferStr(q, " NOT NULL"); |
| |
| if (typdefault != NULL) |
| { |
| appendPQExpBufferStr(q, " DEFAULT "); |
| if (typdefault_is_literal) |
| appendStringLiteralAH(q, typdefault, fout); |
| else |
| appendPQExpBufferStr(q, typdefault); |
| } |
| |
| PQclear(res); |
| |
| /* |
| * Add any CHECK constraints for the domain |
| */ |
| for (i = 0; i < tyinfo->nDomChecks; i++) |
| { |
| ConstraintInfo *domcheck = &(tyinfo->domChecks[i]); |
| |
| if (!domcheck->separate) |
| appendPQExpBuffer(q, "\n\tCONSTRAINT %s %s", |
| fmtId(domcheck->dobj.name), domcheck->condef); |
| } |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| appendPQExpBuffer(delq, "DROP DOMAIN %s;\n", qualtypname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tyinfo->dobj, |
| "DOMAIN", qtypname, |
| tyinfo->dobj.namespace->dobj.name); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tyinfo->dobj.catId, tyinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tyinfo->dobj.name, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "DOMAIN", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Domain Comments and Security Labels */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "DOMAIN", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "DOMAIN", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, tyinfo->dobj.dumpId, InvalidDumpId, "TYPE", |
| qtypname, NULL, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, &tyinfo->dacl); |
| |
| /* Dump any per-constraint comments */ |
| for (i = 0; i < tyinfo->nDomChecks; i++) |
| { |
| ConstraintInfo *domcheck = &(tyinfo->domChecks[i]); |
| PQExpBuffer conprefix = createPQExpBuffer(); |
| |
| appendPQExpBuffer(conprefix, "CONSTRAINT %s ON DOMAIN", |
| fmtId(domcheck->dobj.name)); |
| |
| if (domcheck->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, conprefix->data, qtypname, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, |
| domcheck->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(conprefix); |
| } |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qtypname); |
| free(qualtypname); |
| } |
| |
| /* |
| * dumpCompositeType |
| * writes out to fout the queries to recreate a user-defined stand-alone |
| * composite type |
| */ |
| static void |
| dumpCompositeType(Archive *fout, const TypeInfo *tyinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer dropped = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| char *qtypname; |
| char *qualtypname; |
| int ntups; |
| int i_attname; |
| int i_atttypdefn; |
| int i_attlen; |
| int i_attalign; |
| int i_attisdropped; |
| int i_attcollation; |
| int i; |
| int actual_atts; |
| |
| if (!fout->is_prepared[PREPQUERY_DUMPCOMPOSITETYPE]) |
| { |
| /* |
| * Set up query for type-specific details. |
| * |
| * Since we only want to dump COLLATE clauses for attributes whose |
| * collation is different from their type's default, we use a CASE |
| * here to suppress uninteresting attcollations cheaply. atttypid |
| * will be 0 for dropped columns; collation does not matter for those. |
| */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpCompositeType(pg_catalog.oid) AS\n" |
| "SELECT a.attname, a.attnum, " |
| "pg_catalog.format_type(a.atttypid, a.atttypmod) AS atttypdefn, " |
| "a.attlen, a.attalign, a.attisdropped, " |
| "CASE WHEN a.attcollation <> at.typcollation " |
| "THEN a.attcollation ELSE 0 END AS attcollation " |
| "FROM pg_catalog.pg_type ct " |
| "JOIN pg_catalog.pg_attribute a ON a.attrelid = ct.typrelid " |
| "LEFT JOIN pg_catalog.pg_type at ON at.oid = a.atttypid " |
| "WHERE ct.oid = $1 " |
| "ORDER BY a.attnum"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPCOMPOSITETYPE] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpCompositeType('%u')", |
| tyinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_attname = PQfnumber(res, "attname"); |
| i_atttypdefn = PQfnumber(res, "atttypdefn"); |
| i_attlen = PQfnumber(res, "attlen"); |
| i_attalign = PQfnumber(res, "attalign"); |
| i_attisdropped = PQfnumber(res, "attisdropped"); |
| i_attcollation = PQfnumber(res, "attcollation"); |
| |
| if (dopt->binary_upgrade) |
| { |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| tyinfo->dobj.catId.oid, |
| tyinfo->dobj.namespace->dobj.catId.oid, |
| tyinfo->dobj.name, |
| false, false); |
| binary_upgrade_set_pg_class_oids(fout, q, tyinfo->typrelid, false); |
| } |
| |
| qtypname = pg_strdup(fmtId(tyinfo->dobj.name)); |
| qualtypname = pg_strdup(fmtQualifiedDumpable(tyinfo)); |
| |
| appendPQExpBuffer(q, "CREATE TYPE %s AS (", |
| qualtypname); |
| |
| actual_atts = 0; |
| for (i = 0; i < ntups; i++) |
| { |
| char *attname; |
| char *atttypdefn; |
| char *attlen; |
| char *attalign; |
| bool attisdropped; |
| Oid attcollation; |
| |
| attname = PQgetvalue(res, i, i_attname); |
| atttypdefn = PQgetvalue(res, i, i_atttypdefn); |
| attlen = PQgetvalue(res, i, i_attlen); |
| attalign = PQgetvalue(res, i, i_attalign); |
| attisdropped = (PQgetvalue(res, i, i_attisdropped)[0] == 't'); |
| attcollation = atooid(PQgetvalue(res, i, i_attcollation)); |
| |
| if (attisdropped && !dopt->binary_upgrade) |
| continue; |
| |
| /* Format properly if not first attr */ |
| if (actual_atts++ > 0) |
| appendPQExpBufferChar(q, ','); |
| appendPQExpBufferStr(q, "\n\t"); |
| |
| if (!attisdropped) |
| { |
| appendPQExpBuffer(q, "%s %s", fmtId(attname), atttypdefn); |
| |
| /* Add collation if not default for the column type */ |
| if (OidIsValid(attcollation)) |
| { |
| CollInfo *coll; |
| |
| coll = findCollationByOid(attcollation); |
| if (coll) |
| appendPQExpBuffer(q, " COLLATE %s", |
| fmtQualifiedDumpable(coll)); |
| } |
| } |
| else |
| { |
| /* |
| * This is a dropped attribute and we're in binary_upgrade mode. |
| * Insert a placeholder for it in the CREATE TYPE command, and set |
| * length and alignment with direct UPDATE to the catalogs |
| * afterwards. See similar code in dumpTableSchema(). |
| */ |
| appendPQExpBuffer(q, "%s INTEGER /* dummy */", fmtId(attname)); |
| |
| /* stash separately for insertion after the CREATE TYPE */ |
| appendPQExpBufferStr(dropped, |
| "\n-- For binary upgrade, recreate dropped column.\n"); |
| appendPQExpBuffer(dropped, "SET allow_system_table_mods = true;\n"); |
| appendPQExpBuffer(dropped, "UPDATE pg_catalog.pg_attribute\n" |
| "SET attlen = %s, " |
| "attalign = '%s', attbyval = false\n" |
| "WHERE attname = ", attlen, attalign); |
| appendStringLiteralAH(dropped, attname, fout); |
| appendPQExpBufferStr(dropped, "\n AND attrelid = "); |
| appendStringLiteralAH(dropped, qualtypname, fout); |
| appendPQExpBufferStr(dropped, "::pg_catalog.regclass;\n"); |
| |
| appendPQExpBuffer(dropped, "ALTER TYPE %s ", |
| qualtypname); |
| appendPQExpBuffer(dropped, "DROP ATTRIBUTE %s;\n", |
| fmtId(attname)); |
| appendPQExpBuffer(dropped, "RESET allow_system_table_mods;\n"); |
| } |
| } |
| appendPQExpBufferStr(q, "\n);\n"); |
| appendPQExpBufferStr(q, dropped->data); |
| |
| appendPQExpBuffer(delq, "DROP TYPE %s;\n", qualtypname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tyinfo->dobj, |
| "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tyinfo->dobj.catId, tyinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tyinfo->dobj.name, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "TYPE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| |
| /* Dump Type Comments and Security Labels */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "TYPE", qtypname, |
| tyinfo->dobj.namespace->dobj.name, tyinfo->rolname, |
| tyinfo->dobj.catId, 0, tyinfo->dobj.dumpId); |
| |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, tyinfo->dobj.dumpId, InvalidDumpId, "TYPE", |
| qtypname, NULL, |
| tyinfo->dobj.namespace->dobj.name, |
| tyinfo->rolname, &tyinfo->dacl); |
| |
| /* Dump any per-column comments */ |
| if (tyinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpCompositeTypeColComments(fout, tyinfo, res); |
| |
| PQclear(res); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(dropped); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qtypname); |
| free(qualtypname); |
| } |
| |
| /* |
| * dumpCompositeTypeColComments |
| * writes out to fout the queries to recreate comments on the columns of |
| * a user-defined stand-alone composite type. |
| * |
| * The caller has already made a query to collect the names and attnums |
| * of the type's columns, so we just pass that result into here rather |
| * than reading them again. |
| */ |
| static void |
| dumpCompositeTypeColComments(Archive *fout, const TypeInfo *tyinfo, |
| PGresult *res) |
| { |
| CommentItem *comments; |
| int ncomments; |
| PQExpBuffer query; |
| PQExpBuffer target; |
| int i; |
| int ntups; |
| int i_attname; |
| int i_attnum; |
| int i_attisdropped; |
| |
| /* do nothing, if --no-comments is supplied */ |
| if (fout->dopt->no_comments) |
| return; |
| |
| /* Search for comments associated with type's pg_class OID */ |
| ncomments = findComments(RelationRelationId, tyinfo->typrelid, |
| &comments); |
| |
| /* If no comments exist, we're done */ |
| if (ncomments <= 0) |
| return; |
| |
| /* Build COMMENT ON statements */ |
| query = createPQExpBuffer(); |
| target = createPQExpBuffer(); |
| |
| ntups = PQntuples(res); |
| i_attnum = PQfnumber(res, "attnum"); |
| i_attname = PQfnumber(res, "attname"); |
| i_attisdropped = PQfnumber(res, "attisdropped"); |
| while (ncomments > 0) |
| { |
| const char *attname; |
| |
| attname = NULL; |
| for (i = 0; i < ntups; i++) |
| { |
| if (atoi(PQgetvalue(res, i, i_attnum)) == comments->objsubid && |
| PQgetvalue(res, i, i_attisdropped)[0] != 't') |
| { |
| attname = PQgetvalue(res, i, i_attname); |
| break; |
| } |
| } |
| if (attname) /* just in case we don't find it */ |
| { |
| const char *descr = comments->descr; |
| |
| resetPQExpBuffer(target); |
| appendPQExpBuffer(target, "COLUMN %s.", |
| fmtId(tyinfo->dobj.name)); |
| appendPQExpBufferStr(target, fmtId(attname)); |
| |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, "COMMENT ON COLUMN %s.", |
| fmtQualifiedDumpable(tyinfo)); |
| appendPQExpBuffer(query, "%s IS ", fmtId(attname)); |
| appendStringLiteralAH(query, descr, fout); |
| appendPQExpBufferStr(query, ";\n"); |
| |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = target->data, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "COMMENT", |
| .section = SECTION_NONE, |
| .createStmt = query->data, |
| .deps = &(tyinfo->dobj.dumpId), |
| .nDeps = 1)); |
| } |
| |
| comments++; |
| ncomments--; |
| } |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(target); |
| } |
| |
| /* |
| * dumpShellType |
| * writes out to fout the queries to create a shell type |
| * |
| * We dump a shell definition in advance of the I/O functions for the type. |
| */ |
| static void |
| dumpShellType(Archive *fout, const ShellTypeInfo *stinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| |
| /* |
| * Note the lack of a DROP command for the shell type; any required DROP |
| * is driven off the base type entry, instead. This interacts with |
| * _printTocEntry()'s use of the presence of a DROP command to decide |
| * whether an entry needs an ALTER OWNER command. We don't want to alter |
| * the shell type's owner immediately on creation; that should happen only |
| * after it's filled in, otherwise the backend complains. |
| */ |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_type_oid(fout, q, |
| stinfo->baseType->dobj.catId.oid, |
| stinfo->baseType->dobj.namespace->dobj.catId.oid, |
| stinfo->baseType->dobj.name, |
| false, false); |
| |
| appendPQExpBuffer(q, "CREATE TYPE %s;\n", |
| fmtQualifiedDumpable(stinfo)); |
| |
| if (stinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, stinfo->dobj.catId, stinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = stinfo->dobj.name, |
| .namespace = stinfo->dobj.namespace->dobj.name, |
| .owner = stinfo->baseType->rolname, |
| .description = "SHELL TYPE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data)); |
| |
| destroyPQExpBuffer(q); |
| } |
| |
| /* |
| * dumpProcLang |
| * writes out to fout the queries to recreate a user-defined |
| * procedural language |
| */ |
| static void |
| dumpProcLang(Archive *fout, const ProcLangInfo *plang) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer defqry; |
| PQExpBuffer delqry; |
| bool useParams; |
| char *qlanname; |
| FuncInfo *funcInfo; |
| FuncInfo *inlineInfo = NULL; |
| FuncInfo *validatorInfo = NULL; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* |
| * Try to find the support function(s). It is not an error if we don't |
| * find them --- if the functions are in the pg_catalog schema, as is |
| * standard in 8.1 and up, then we won't have loaded them. (In this case |
| * we will emit a parameterless CREATE LANGUAGE command, which will |
| * require PL template knowledge in the backend to reload.) |
| */ |
| |
| funcInfo = findFuncByOid(plang->lanplcallfoid); |
| if (funcInfo != NULL && !funcInfo->dobj.dump) |
| funcInfo = NULL; /* treat not-dumped same as not-found */ |
| |
| if (OidIsValid(plang->laninline)) |
| { |
| inlineInfo = findFuncByOid(plang->laninline); |
| if (inlineInfo != NULL && !inlineInfo->dobj.dump) |
| inlineInfo = NULL; |
| } |
| |
| if (OidIsValid(plang->lanvalidator)) |
| { |
| validatorInfo = findFuncByOid(plang->lanvalidator); |
| if (validatorInfo != NULL && !validatorInfo->dobj.dump) |
| validatorInfo = NULL; |
| } |
| |
| /* |
| * If the functions are dumpable then emit a complete CREATE LANGUAGE with |
| * parameters. Otherwise, we'll write a parameterless command, which will |
| * be interpreted as CREATE EXTENSION. |
| */ |
| useParams = (funcInfo != NULL && |
| (inlineInfo != NULL || !OidIsValid(plang->laninline)) && |
| (validatorInfo != NULL || !OidIsValid(plang->lanvalidator))); |
| |
| defqry = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| |
| qlanname = pg_strdup(fmtId(plang->dobj.name)); |
| |
| appendPQExpBuffer(delqry, "DROP PROCEDURAL LANGUAGE %s;\n", |
| qlanname); |
| |
| if (useParams) |
| { |
| appendPQExpBuffer(defqry, "CREATE %sPROCEDURAL LANGUAGE %s", |
| plang->lanpltrusted ? "TRUSTED " : "", |
| qlanname); |
| appendPQExpBuffer(defqry, " HANDLER %s", |
| fmtQualifiedDumpable(funcInfo)); |
| if (OidIsValid(plang->laninline)) |
| appendPQExpBuffer(defqry, " INLINE %s", |
| fmtQualifiedDumpable(inlineInfo)); |
| if (OidIsValid(plang->lanvalidator)) |
| appendPQExpBuffer(defqry, " VALIDATOR %s", |
| fmtQualifiedDumpable(validatorInfo)); |
| } |
| else |
| { |
| /* |
| * If not dumping parameters, then use CREATE OR REPLACE so that the |
| * command will not fail if the language is preinstalled in the target |
| * database. |
| * |
| * Modern servers will interpret this as CREATE EXTENSION IF NOT |
| * EXISTS; perhaps we should emit that instead? But it might just add |
| * confusion. |
| */ |
| appendPQExpBuffer(defqry, "CREATE OR REPLACE PROCEDURAL LANGUAGE %s", |
| qlanname); |
| } |
| appendPQExpBufferStr(defqry, ";\n"); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(defqry, &plang->dobj, |
| "LANGUAGE", qlanname, NULL); |
| |
| if (plang->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, plang->dobj.catId, plang->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = plang->dobj.name, |
| .owner = plang->lanowner, |
| .description = "PROCEDURAL LANGUAGE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = defqry->data, |
| .dropStmt = delqry->data, |
| )); |
| |
| /* Dump Proc Lang Comments and Security Labels */ |
| if (plang->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "LANGUAGE", qlanname, |
| NULL, plang->lanowner, |
| plang->dobj.catId, 0, plang->dobj.dumpId); |
| |
| if (plang->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "LANGUAGE", qlanname, |
| NULL, plang->lanowner, |
| plang->dobj.catId, 0, plang->dobj.dumpId); |
| |
| if (plang->lanpltrusted && plang->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, plang->dobj.dumpId, InvalidDumpId, "LANGUAGE", |
| qlanname, NULL, NULL, |
| plang->lanowner, &plang->dacl); |
| |
| free(qlanname); |
| |
| destroyPQExpBuffer(defqry); |
| destroyPQExpBuffer(delqry); |
| } |
| |
| /* |
| * format_function_arguments: generate function name and argument list |
| * |
| * This is used when we can rely on pg_get_function_arguments to format |
| * the argument list. Note, however, that pg_get_function_arguments |
| * does not special-case zero-argument aggregates. |
| */ |
| static char * |
| format_function_arguments(const FuncInfo *finfo, const char *funcargs, bool is_agg) |
| { |
| PQExpBufferData fn; |
| |
| initPQExpBuffer(&fn); |
| appendPQExpBufferStr(&fn, fmtId(finfo->dobj.name)); |
| if (is_agg && finfo->nargs == 0) |
| appendPQExpBufferStr(&fn, "(*)"); |
| else |
| appendPQExpBuffer(&fn, "(%s)", funcargs); |
| return fn.data; |
| } |
| |
| /* |
| <<<<<<< HEAD |
| * format_function_arguments_old: generate function name and argument list |
| * |
| * The argument type names are qualified if needed. The function name |
| * is never qualified. |
| * |
| * This is used only with pre-GPDB 5.0 servers, so we aren't expecting to see |
| * DEFAULT arguments. |
| * |
| * Any or all of allargtypes, argmodes, argnames may be NULL. |
| */ |
| static char * |
| format_function_arguments_old(Archive *fout, |
| const FuncInfo *finfo, int nallargs, |
| char **allargtypes, |
| char **argmodes, |
| char **argnames) |
| { |
| PQExpBufferData fn; |
| int j; |
| |
| initPQExpBuffer(&fn); |
| appendPQExpBuffer(&fn, "%s(", fmtId(finfo->dobj.name)); |
| for (j = 0; j < nallargs; j++) |
| { |
| Oid typid; |
| const char *typname; |
| const char *argmode; |
| const char *argname; |
| |
| typid = allargtypes ? atooid(allargtypes[j]) : finfo->argtypes[j]; |
| typname = getFormattedTypeName(fout, typid, zeroIsError); |
| |
| if (argmodes) |
| { |
| switch (argmodes[j][0]) |
| { |
| case PROARGMODE_IN: |
| argmode = ""; |
| break; |
| case PROARGMODE_OUT: |
| argmode = "OUT "; |
| break; |
| case PROARGMODE_INOUT: |
| argmode = "INOUT "; |
| break; |
| case PROARGMODE_VARIADIC: |
| argmode = "VARIADIC "; |
| break; |
| case PROARGMODE_TABLE: |
| /* skip table column's names */ |
| free((void *)typname); |
| continue; |
| default: |
| pg_log_warning("bogus value in proargmodes array"); |
| argmode = ""; |
| break; |
| } |
| } |
| else |
| argmode = ""; |
| |
| argname = argnames ? argnames[j] : (char *) NULL; |
| if (argname && argname[0] == '\0') |
| argname = NULL; |
| |
| appendPQExpBuffer(&fn, "%s%s%s%s%s", |
| (j > 0) ? ", " : "", |
| argmode, |
| argname ? fmtId(argname) : "", |
| argname ? " " : "", |
| typname); |
| } |
| appendPQExpBufferChar(&fn, ')'); |
| return fn.data; |
| } |
| |
| /* |
| * is_returns_table_function: returns true if function id declared as |
| * RETURNS TABLE, i.e. at least one argument is PROARGMODE_TABLE |
| */ |
| static bool |
| is_returns_table_function(int nallargs, char **argmodes) |
| { |
| int j; |
| |
| if (argmodes) |
| for (j = 0; j < nallargs; j++) |
| if (argmodes[j][0] == PROARGMODE_TABLE) |
| return true; |
| |
| return false; |
| } |
| |
| |
| /* |
| * format_table_function_columns: generate column list for |
| * table functions. |
| */ |
| static char * |
| format_table_function_columns(Archive *fout, const FuncInfo *finfo, int nallargs, |
| char **allargtypes, |
| char **argmodes, |
| char **argnames) |
| { |
| PQExpBufferData fn; |
| int j; |
| bool first_column = true; |
| |
| initPQExpBuffer(&fn); |
| appendPQExpBuffer(&fn, "("); |
| |
| for (j = 0; j < nallargs; j++) |
| { |
| Oid typid; |
| const char *typname; |
| |
| /* |
| * argmodes are checked in format_function_arguments, it isn't necessary |
| * to check argmodes here again |
| */ |
| if (argmodes[j][0] == PROARGMODE_TABLE) |
| { |
| typid = allargtypes ? atooid(allargtypes[j]) : finfo->argtypes[j]; |
| typname = getFormattedTypeName(fout, typid, zeroIsError); |
| |
| /* column's name is always NOT NULL (checked in gram.y) */ |
| appendPQExpBuffer(&fn, "%s%s %s", |
| first_column ? "" : ", ", |
| fmtId(argnames[j]), |
| typname); |
| free((void *)typname); |
| first_column = false; |
| } |
| } |
| |
| appendPQExpBuffer(&fn, ")"); |
| return fn.data; |
| } |
| |
| |
| /* |
| ======= |
| >>>>>>> REL_16_9 |
| * format_function_signature: generate function name and argument list |
| * |
| * Only a minimal list of input argument types is generated; this is |
| * sufficient to reference the function, but not to define it. |
| * |
| * If honor_quotes is false then the function name is never quoted. |
| * This is appropriate for use in TOC tags, but not in SQL commands. |
| */ |
| static char * |
| format_function_signature(Archive *fout, const FuncInfo *finfo, bool honor_quotes) |
| { |
| PQExpBufferData fn; |
| int j; |
| |
| initPQExpBuffer(&fn); |
| if (honor_quotes) |
| appendPQExpBuffer(&fn, "%s(", fmtId(finfo->dobj.name)); |
| else |
| appendPQExpBuffer(&fn, "%s(", finfo->dobj.name); |
| for (j = 0; j < finfo->nargs; j++) |
| { |
| if (j > 0) |
| appendPQExpBufferStr(&fn, ", "); |
| |
| appendPQExpBufferStr(&fn, |
| getFormattedTypeName(fout, finfo->argtypes[j], |
| zeroIsError)); |
| } |
| appendPQExpBufferChar(&fn, ')'); |
| return fn.data; |
| } |
| |
| |
| /* |
| * dumpFunc: |
| * dump out one function |
| */ |
| static void |
| dumpFunc(Archive *fout, const FuncInfo *finfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delqry; |
| PQExpBuffer asPart; |
| PGresult *res; |
| char *funcsig; /* identity signature */ |
| char *funcfullsig = NULL; /* full signature */ |
| char *funcsig_tag; |
| char *qual_funcsig; |
| char *proretset; |
| char *prosrc; |
| char *probin; |
| char *prosqlbody; |
| char *funcargs; |
| char *funciargs; |
| char *funcresult; |
| char *protrftypes; |
| char *prokind; |
| char *provolatile; |
| char *proisstrict; |
| char *prosecdef; |
| char *proleakproof; |
| char *proconfig; |
| char *procost; |
| char *prorows; |
| char *prosupport; |
| char *proparallel; |
| char *lanname; |
| <<<<<<< HEAD |
| int nallargs; |
| char **allargtypes = NULL; |
| char **argmodes = NULL; |
| char **argnames = NULL; |
| bool isGE50 = isGPDB5000OrLater(fout); |
| ======= |
| >>>>>>> REL_16_9 |
| char **configitems = NULL; |
| int nconfigitems = 0; |
| const char *keyword; |
| |
| <<<<<<< HEAD |
| /* |
| * GPDB_14_MERGE_FIXME: additional columns in GPDB |
| * these are added except Version >= 90500 which may |
| * be a bug. |
| * The new sqls are added in decent version order. |
| * So we can't handle addition values like before. |
| * These variables are handled differently base on ranges: |
| * Version: |
| * < 80300, >= 80300 < 90200, >= 90200 |
| */ |
| |
| char *callbackfunc; |
| char *prodataaccess; |
| char *proexeclocation; |
| |
| /* Skip if not to be dumped */ |
| if (!finfo->dobj.dump || dopt->dataOnly) |
| ======= |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| >>>>>>> REL_16_9 |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| asPart = createPQExpBuffer(); |
| |
| <<<<<<< HEAD |
| /* Fetch function-specific details */ |
| appendPQExpBufferStr(query, |
| "SELECT\n" |
| "proretset,\n" |
| "prosrc,\n" |
| "probin,\n" |
| "provolatile,\n" |
| "proisstrict,\n" |
| "prosecdef,\n" |
| "lanname,\n"); |
| |
| if (fout->remoteVersion >= 80300) |
| { |
| ======= |
| if (!fout->is_prepared[PREPQUERY_DUMPFUNC]) |
| { |
| /* Set up query for function-specific details */ |
| >>>>>>> REL_16_9 |
| appendPQExpBufferStr(query, |
| "PREPARE dumpFunc(pg_catalog.oid) AS\n"); |
| |
| appendPQExpBufferStr(query, |
| "SELECT\n" |
| "proretset,\n" |
| "prosrc,\n" |
| "probin,\n" |
| "provolatile,\n" |
| "proisstrict,\n" |
| "prosecdef,\n" |
| "lanname,\n" |
| "proconfig,\n" |
| "procost,\n" |
| <<<<<<< HEAD |
| "prorows,\n"); |
| |
| if (isGE50 && fout->remoteVersion < 90200) |
| { |
| appendPQExpBufferStr(query, |
| "prodataaccess,\n" |
| "'a' as proexeclocation,\n" |
| "(SELECT procallback FROM pg_catalog.pg_proc_callback WHERE profnoid::pg_catalog.oid = p.oid) as callbackfunc,\n"); |
| } |
| } |
| else |
| { |
| appendPQExpBufferStr(query, |
| "null AS proconfig,\n" |
| "0 AS procost,\n" |
| "0 AS prorows,\n"); |
| |
| appendPQExpBuffer(query, |
| "%s\n" |
| "'a' as proexeclocation,\n" |
| "(SELECT procallback FROM pg_catalog.pg_proc_callback WHERE profnoid::pg_catalog.oid = p.oid) as callbackfunc,\n", |
| "prodataaccess,"); |
| } |
| |
| if (fout->remoteVersion >= 80400) |
| { |
| /* |
| * In GPDB 5.0 and up we rely on pg_get_function_arguments and |
| * pg_get_function_result instead of examining proallargtypes etc. |
| */ |
| appendPQExpBufferStr(query, |
| "pg_catalog.pg_get_function_arguments(p.oid) AS funcargs,\n" |
| "pg_catalog.pg_get_function_identity_arguments(p.oid) AS funciargs,\n" |
| "pg_catalog.pg_get_function_result(p.oid) AS funcresult,\n"); |
| } |
| else if (fout->remoteVersion >= 80100) |
| appendPQExpBufferStr(query, |
| "proallargtypes,\n" |
| "proargmodes,\n" |
| "proargnames,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "null AS proallargtypes,\n" |
| "null AS proargmodes,\n" |
| "proargnames,\n"); |
| |
| if (fout->remoteVersion >= 90200) |
| appendPQExpBufferStr(query, |
| "prodataaccess,\n" |
| "proexeclocation,\n" |
| "(SELECT procallback FROM pg_catalog.pg_proc_callback WHERE profnoid::pg_catalog.oid = p.oid) as callbackfunc,\n" |
| ======= |
| "prorows,\n" |
| "pg_catalog.pg_get_function_arguments(p.oid) AS funcargs,\n" |
| "pg_catalog.pg_get_function_identity_arguments(p.oid) AS funciargs,\n" |
| "pg_catalog.pg_get_function_result(p.oid) AS funcresult,\n" |
| >>>>>>> REL_16_9 |
| "proleakproof,\n"); |
| |
| if (fout->remoteVersion >= 90500) |
| appendPQExpBufferStr(query, |
| "array_to_string(protrftypes, ' ') AS protrftypes,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "NULL AS protrftypes,\n"); |
| |
| if (fout->remoteVersion >= 90600) |
| appendPQExpBufferStr(query, |
| "proparallel,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "'u' AS proparallel,\n"); |
| |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(query, |
| "prokind,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "CASE WHEN proiswindow THEN 'w' ELSE 'f' END AS prokind,\n"); |
| |
| if (fout->remoteVersion >= 120000) |
| appendPQExpBufferStr(query, |
| "prosupport,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "'-' AS prosupport,\n"); |
| |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(query, |
| "pg_get_function_sqlbody(p.oid) AS prosqlbody\n"); |
| else |
| appendPQExpBufferStr(query, |
| "NULL AS prosqlbody\n"); |
| |
| appendPQExpBufferStr(query, |
| "FROM pg_catalog.pg_proc p, pg_catalog.pg_language l\n" |
| "WHERE p.oid = $1 " |
| "AND l.oid = p.prolang"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPFUNC] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpFunc('%u')", |
| finfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| proretset = PQgetvalue(res, 0, PQfnumber(res, "proretset")); |
| if (PQgetisnull(res, 0, PQfnumber(res, "prosqlbody"))) |
| { |
| prosrc = PQgetvalue(res, 0, PQfnumber(res, "prosrc")); |
| probin = PQgetvalue(res, 0, PQfnumber(res, "probin")); |
| prosqlbody = NULL; |
| } |
| else |
| { |
| prosrc = NULL; |
| probin = NULL; |
| prosqlbody = PQgetvalue(res, 0, PQfnumber(res, "prosqlbody")); |
| } |
| <<<<<<< HEAD |
| /* |
| * GPDB_14_MERGE_FIXME: |
| * isGE50 is >= 80300 but upstream is (fout->remoteVersion >= 80400) |
| * Need to check. |
| */ |
| if (isGE50) |
| { |
| funcargs = PQgetvalue(res, 0, PQfnumber(res, "funcargs")); |
| funciargs = PQgetvalue(res, 0, PQfnumber(res, "funciargs")); |
| funcresult = PQgetvalue(res, 0, PQfnumber(res, "funcresult")); |
| proallargtypes = proargmodes = proargnames = NULL; |
| } |
| else |
| { |
| proallargtypes = PQgetvalue(res, 0, PQfnumber(res, "proallargtypes")); |
| proargmodes = PQgetvalue(res, 0, PQfnumber(res, "proargmodes")); |
| proargnames = PQgetvalue(res, 0, PQfnumber(res, "proargnames")); |
| funcargs = funciargs = funcresult = NULL; |
| } |
| if (PQfnumber(res, "protrftypes") != -1) |
| protrftypes = PQgetvalue(res, 0, PQfnumber(res, "protrftypes")); |
| else |
| protrftypes = NULL; |
| ======= |
| funcargs = PQgetvalue(res, 0, PQfnumber(res, "funcargs")); |
| funciargs = PQgetvalue(res, 0, PQfnumber(res, "funciargs")); |
| funcresult = PQgetvalue(res, 0, PQfnumber(res, "funcresult")); |
| protrftypes = PQgetvalue(res, 0, PQfnumber(res, "protrftypes")); |
| >>>>>>> REL_16_9 |
| prokind = PQgetvalue(res, 0, PQfnumber(res, "prokind")); |
| provolatile = PQgetvalue(res, 0, PQfnumber(res, "provolatile")); |
| proisstrict = PQgetvalue(res, 0, PQfnumber(res, "proisstrict")); |
| prosecdef = PQgetvalue(res, 0, PQfnumber(res, "prosecdef")); |
| proleakproof = PQgetvalue(res, 0, PQfnumber(res, "proleakproof")); |
| proconfig = PQgetvalue(res, 0, PQfnumber(res, "proconfig")); |
| procost = PQgetvalue(res, 0, PQfnumber(res, "procost")); |
| prorows = PQgetvalue(res, 0, PQfnumber(res, "prorows")); |
| prosupport = PQgetvalue(res, 0, PQfnumber(res, "prosupport")); |
| proparallel = PQgetvalue(res, 0, PQfnumber(res, "proparallel")); |
| lanname = PQgetvalue(res, 0, PQfnumber(res, "lanname")); |
| callbackfunc = PQgetvalue(res, 0, PQfnumber(res, "callbackfunc")); |
| prodataaccess = PQgetvalue(res, 0, PQfnumber(res, "prodataaccess")); |
| proexeclocation = PQgetvalue(res, 0, PQfnumber(res, "proexeclocation")); |
| |
| /* |
| * See backend/commands/functioncmds.c for details of how the 'AS' clause |
| * is used. |
| */ |
| if (prosqlbody) |
| { |
| appendPQExpBufferStr(asPart, prosqlbody); |
| } |
| else if (probin[0] != '\0') |
| { |
| appendPQExpBufferStr(asPart, "AS "); |
| appendStringLiteralAH(asPart, probin, fout); |
| if (prosrc[0] != '\0') |
| { |
| appendPQExpBufferStr(asPart, ", "); |
| |
| /* |
| * where we have bin, use dollar quoting if allowed and src |
| * contains quote or backslash; else use regular quoting. |
| */ |
| if (dopt->disable_dollar_quoting || |
| (strchr(prosrc, '\'') == NULL && strchr(prosrc, '\\') == NULL)) |
| appendStringLiteralAH(asPart, prosrc, fout); |
| else |
| appendStringLiteralDQ(asPart, prosrc, NULL); |
| } |
| } |
| else |
| { |
| appendPQExpBufferStr(asPart, "AS "); |
| /* with no bin, dollar quote src unconditionally if allowed */ |
| if (dopt->disable_dollar_quoting) |
| appendStringLiteralAH(asPart, prosrc, fout); |
| else |
| appendStringLiteralDQ(asPart, prosrc, NULL); |
| } |
| |
| <<<<<<< HEAD |
| if (proargmodes && *proargmodes) |
| { |
| int nitems = 0; |
| |
| if (!parsePGArray(proargmodes, &argmodes, &nitems) || |
| nitems != nallargs) |
| { |
| pg_log_warning("could not parse proargmodes array"); |
| if (argmodes) |
| free(argmodes); |
| argmodes = NULL; |
| } |
| } |
| |
| if (proargnames && *proargnames) |
| { |
| int nitems = 0; |
| |
| if (!parsePGArray(proargnames, &argnames, &nitems) || |
| nitems != nallargs) |
| { |
| pg_log_warning("could not parse proargnames array"); |
| if (argnames) |
| free(argnames); |
| argnames = NULL; |
| } |
| } |
| |
| if (funcargs) |
| { |
| /* GPDB 5.0 or later; we rely on server-side code for most of the work */ |
| funcfullsig = format_function_arguments(finfo, funcargs, false); |
| funcsig = format_function_arguments(finfo, funciargs, false); |
| } |
| else |
| { |
| /* pre-GPDB 5.0, do it ourselves */ |
| funcsig = format_function_arguments_old(fout, finfo, nallargs, allargtypes, |
| argmodes, argnames); |
| funcfullsig = funcsig; |
| } |
| funcsig_tag = format_function_signature(fout, finfo, false); |
| |
| if (proconfig && *proconfig) |
| ======= |
| if (*proconfig) |
| >>>>>>> REL_16_9 |
| { |
| if (!parsePGArray(proconfig, &configitems, &nconfigitems)) |
| pg_fatal("could not parse %s array", "proconfig"); |
| } |
| else |
| { |
| configitems = NULL; |
| nconfigitems = 0; |
| } |
| |
| <<<<<<< HEAD |
| funcsig_tag = format_function_signature(fout, finfo, false); |
| |
| if (prokind[0] == PROKIND_PROCEDURE) |
| keyword = "PROCEDURE"; |
| else |
| { |
| configitems = NULL; |
| nconfigitems = 0; |
| } |
| |
| if (funcargs) |
| { |
| /* 8.4 or later; we rely on server-side code for most of the work */ |
| funcfullsig = format_function_arguments(finfo, funcargs, false); |
| funcsig = format_function_arguments(finfo, funciargs, false); |
| } |
| else |
| /* pre-8.4, do it ourselves */ |
| funcsig = format_function_arguments_old(fout, |
| finfo, nallargs, allargtypes, |
| argmodes, argnames); |
| ======= |
| funcfullsig = format_function_arguments(finfo, funcargs, false); |
| funcsig = format_function_arguments(finfo, funciargs, false); |
| >>>>>>> REL_16_9 |
| |
| funcsig_tag = format_function_signature(fout, finfo, false); |
| |
| qual_funcsig = psprintf("%s.%s", |
| fmtId(finfo->dobj.namespace->dobj.name), |
| funcsig); |
| |
| if (prokind[0] == PROKIND_PROCEDURE) |
| keyword = "PROCEDURE"; |
| else |
| keyword = "FUNCTION"; /* works for window functions too */ |
| |
| appendPQExpBuffer(delqry, "DROP %s %s;\n", |
| keyword, qual_funcsig); |
| |
| appendPQExpBuffer(q, "CREATE %s %s.%s", |
| keyword, |
| fmtId(finfo->dobj.namespace->dobj.name), |
| funcfullsig ? funcfullsig : |
| funcsig); |
| |
| if (prokind[0] == PROKIND_PROCEDURE) |
| /* no result type to output */ ; |
| else if (funcresult) |
| appendPQExpBuffer(q, " RETURNS %s", funcresult); |
| <<<<<<< HEAD |
| else if (!is_returns_table_function(nallargs, argmodes)) |
| { |
| ======= |
| else |
| >>>>>>> REL_16_9 |
| appendPQExpBuffer(q, " RETURNS %s%s", |
| (proretset[0] == 't') ? "SETOF " : "", |
| getFormattedTypeName(fout, finfo->prorettype, |
| zeroIsError)); |
| <<<<<<< HEAD |
| } |
| else |
| { |
| /* RETURNS TABLE functions */ |
| char *func_cols; |
| func_cols = format_table_function_columns(fout, finfo, nallargs, allargtypes, |
| argmodes, argnames); |
| appendPQExpBuffer(q, "RETURNS TABLE %s", func_cols); |
| free(func_cols); |
| } |
| ======= |
| >>>>>>> REL_16_9 |
| |
| appendPQExpBuffer(q, "\n LANGUAGE %s", fmtId(lanname)); |
| |
| if (*protrftypes) |
| { |
| Oid *typeids = palloc(FUNC_MAX_ARGS * sizeof(Oid)); |
| int i; |
| |
| appendPQExpBufferStr(q, " TRANSFORM "); |
| parseOidArray(protrftypes, typeids, FUNC_MAX_ARGS); |
| for (i = 0; typeids[i]; i++) |
| { |
| if (i != 0) |
| appendPQExpBufferStr(q, ", "); |
| appendPQExpBuffer(q, "FOR TYPE %s", |
| getFormattedTypeName(fout, typeids[i], zeroAsNone)); |
| } |
| } |
| |
| if (prokind[0] == PROKIND_WINDOW) |
| appendPQExpBufferStr(q, " WINDOW"); |
| |
| if (provolatile[0] != PROVOLATILE_VOLATILE) |
| { |
| if (provolatile[0] == PROVOLATILE_IMMUTABLE) |
| appendPQExpBufferStr(q, " IMMUTABLE"); |
| else if (provolatile[0] == PROVOLATILE_STABLE) |
| appendPQExpBufferStr(q, " STABLE"); |
| else if (provolatile[0] != PROVOLATILE_VOLATILE) |
| pg_fatal("unrecognized provolatile value for function \"%s\"", |
| finfo->dobj.name); |
| } |
| |
| if (proisstrict[0] == 't') |
| appendPQExpBufferStr(q, " STRICT"); |
| |
| if (prosecdef[0] == 't') |
| appendPQExpBufferStr(q, " SECURITY DEFINER"); |
| |
| if (proleakproof[0] == 't') |
| appendPQExpBufferStr(q, " LEAKPROOF"); |
| |
| /* |
| * COST and ROWS are emitted only if present and not default, so as not to |
| * break backwards-compatibility of the dump without need. Keep this code |
| * in sync with the defaults in functioncmds.c. |
| */ |
| if (strcmp(procost, "0") != 0) |
| { |
| if (strcmp(lanname, "internal") == 0 || strcmp(lanname, "c") == 0) |
| { |
| /* default cost is 1 */ |
| if (strcmp(procost, "1") != 0) |
| appendPQExpBuffer(q, " COST %s", procost); |
| } |
| else |
| { |
| /* default cost is 100 */ |
| if (strcmp(procost, "100") != 0) |
| appendPQExpBuffer(q, " COST %s", procost); |
| } |
| } |
| if (proretset[0] == 't' && |
| strcmp(prorows, "0") != 0 && strcmp(prorows, "1000") != 0) |
| appendPQExpBuffer(q, " ROWS %s", prorows); |
| |
| if (strcmp(prosupport, "-") != 0) |
| { |
| /* We rely on regprocout to provide quoting and qualification */ |
| appendPQExpBuffer(q, " SUPPORT %s", prosupport); |
| } |
| |
| if (proparallel[0] != PROPARALLEL_UNSAFE) |
| { |
| if (proparallel[0] == PROPARALLEL_SAFE) |
| appendPQExpBufferStr(q, " PARALLEL SAFE"); |
| else if (proparallel[0] == PROPARALLEL_RESTRICTED) |
| appendPQExpBufferStr(q, " PARALLEL RESTRICTED"); |
| else if (proparallel[0] != PROPARALLEL_UNSAFE) |
| pg_fatal("unrecognized proparallel value for function \"%s\"", |
| finfo->dobj.name); |
| } |
| |
| <<<<<<< HEAD |
| if (prodataaccess[0] == PRODATAACCESS_NONE) |
| appendPQExpBuffer(q, " NO SQL"); |
| else if (prodataaccess[0] == PRODATAACCESS_CONTAINS) |
| appendPQExpBuffer(q, " CONTAINS SQL"); |
| else if (prodataaccess[0] == PRODATAACCESS_READS) |
| appendPQExpBuffer(q, " READS SQL DATA"); |
| else if (prodataaccess[0] == PRODATAACCESS_MODIFIES) |
| appendPQExpBuffer(q, " MODIFIES SQL DATA"); |
| |
| if (proexeclocation[0] == PROEXECLOCATION_ANY) |
| { |
| /* the default, omit */ |
| } |
| else if (proexeclocation[0] == 'm') |
| appendPQExpBuffer(q, " EXECUTE ON COORDINATOR"); |
| else if (proexeclocation[0] == PROEXECLOCATION_COORDINATOR) |
| appendPQExpBuffer(q, " EXECUTE ON COORDINATOR"); |
| else if (proexeclocation[0] == PROEXECLOCATION_ALL_SEGMENTS) |
| appendPQExpBuffer(q, " EXECUTE ON ALL SEGMENTS"); |
| else if (proexeclocation[0] == PROEXECLOCATION_INITPLAN) |
| appendPQExpBuffer(q, " EXECUTE ON INITPLAN"); |
| else |
| { |
| pg_log_warning("unrecognized proexeclocation value: %c", |
| proexeclocation[0]); |
| exit_nicely(1); |
| } |
| |
| for (i = 0; i < nconfigitems; i++) |
| ======= |
| for (int i = 0; i < nconfigitems; i++) |
| >>>>>>> REL_16_9 |
| { |
| /* we feel free to scribble on configitems[] here */ |
| char *configitem = configitems[i]; |
| char *pos; |
| |
| pos = strchr(configitem, '='); |
| if (pos == NULL) |
| continue; |
| *pos++ = '\0'; |
| appendPQExpBuffer(q, "\n SET %s TO ", fmtId(configitem)); |
| |
| /* |
| * Variables that are marked GUC_LIST_QUOTE were already fully quoted |
| * by flatten_set_variable_args() before they were put into the |
| * proconfig array. However, because the quoting rules used there |
| * aren't exactly like SQL's, we have to break the list value apart |
| * and then quote the elements as string literals. (The elements may |
| * be double-quoted as-is, but we can't just feed them to the SQL |
| * parser; it would do the wrong thing with elements that are |
| * zero-length or longer than NAMEDATALEN.) |
| * |
| * Variables that are not so marked should just be emitted as simple |
| * string literals. If the variable is not known to |
| * variable_is_guc_list_quote(), we'll do that; this makes it unsafe |
| * to use GUC_LIST_QUOTE for extension variables. |
| */ |
| if (variable_is_guc_list_quote(configitem)) |
| { |
| char **namelist; |
| char **nameptr; |
| |
| /* Parse string into list of identifiers */ |
| /* this shouldn't fail really */ |
| if (SplitGUCList(pos, ',', &namelist)) |
| { |
| for (nameptr = namelist; *nameptr; nameptr++) |
| { |
| if (nameptr != namelist) |
| appendPQExpBufferStr(q, ", "); |
| appendStringLiteralAH(q, *nameptr, fout); |
| } |
| } |
| pg_free(namelist); |
| } |
| else |
| appendStringLiteralAH(q, pos, fout); |
| } |
| |
| appendPQExpBuffer(q, "\n %s", asPart->data); |
| |
| /* Append callback function */ |
| if (callbackfunc && callbackfunc[0] != '\0') |
| appendPQExpBuffer(q, "\n WITH (describe = %s)", callbackfunc); |
| |
| appendPQExpBuffer(q, ";\n"); |
| |
| append_depends_on_extension(fout, q, &finfo->dobj, |
| "pg_catalog.pg_proc", keyword, |
| qual_funcsig); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &finfo->dobj, |
| keyword, funcsig, |
| finfo->dobj.namespace->dobj.name); |
| |
| if (finfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, finfo->dobj.catId, finfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = funcsig_tag, |
| .namespace = finfo->dobj.namespace->dobj.name, |
| .owner = finfo->rolname, |
| .description = keyword, |
| .section = finfo->postponed_def ? |
| SECTION_POST_DATA : SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delqry->data)); |
| |
| /* Dump Function Comments and Security Labels */ |
| if (finfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, keyword, funcsig, |
| finfo->dobj.namespace->dobj.name, finfo->rolname, |
| finfo->dobj.catId, 0, finfo->dobj.dumpId); |
| |
| if (finfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, keyword, funcsig, |
| finfo->dobj.namespace->dobj.name, finfo->rolname, |
| finfo->dobj.catId, 0, finfo->dobj.dumpId); |
| |
| if (finfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, finfo->dobj.dumpId, InvalidDumpId, keyword, |
| funcsig, NULL, |
| finfo->dobj.namespace->dobj.name, |
| finfo->rolname, &finfo->dacl); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delqry); |
| destroyPQExpBuffer(asPart); |
| free(funcsig); |
| free(funcfullsig); |
| free(funcsig_tag); |
| free(qual_funcsig); |
| <<<<<<< HEAD |
| if (allargtypes) |
| free(allargtypes); |
| if (argmodes) |
| free(argmodes); |
| if (argnames) |
| free(argnames); |
| if (configitems) |
| free(configitems); |
| ======= |
| free(configitems); |
| >>>>>>> REL_16_9 |
| } |
| |
| |
| /* |
| * Dump a user-defined cast |
| */ |
| static void |
| dumpCast(Archive *fout, const CastInfo *cast) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer defqry; |
| PQExpBuffer delqry; |
| PQExpBuffer labelq; |
| PQExpBuffer castargs; |
| FuncInfo *funcInfo = NULL; |
| const char *sourceType; |
| const char *targetType; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* Cannot dump if we don't have the cast function's info */ |
| if (OidIsValid(cast->castfunc)) |
| { |
| funcInfo = findFuncByOid(cast->castfunc); |
| if (funcInfo == NULL) |
| pg_fatal("could not find function definition for function with OID %u", |
| cast->castfunc); |
| } |
| |
| defqry = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| labelq = createPQExpBuffer(); |
| castargs = createPQExpBuffer(); |
| |
| sourceType = getFormattedTypeName(fout, cast->castsource, zeroAsNone); |
| targetType = getFormattedTypeName(fout, cast->casttarget, zeroAsNone); |
| appendPQExpBuffer(delqry, "DROP CAST (%s AS %s);\n", |
| sourceType, targetType); |
| |
| appendPQExpBuffer(defqry, "CREATE CAST (%s AS %s) ", |
| sourceType, targetType); |
| |
| switch (cast->castmethod) |
| { |
| case COERCION_METHOD_BINARY: |
| appendPQExpBufferStr(defqry, "WITHOUT FUNCTION"); |
| break; |
| case COERCION_METHOD_INOUT: |
| appendPQExpBufferStr(defqry, "WITH INOUT"); |
| break; |
| case COERCION_METHOD_FUNCTION: |
| if (funcInfo) |
| { |
| char *fsig = format_function_signature(fout, funcInfo, true); |
| |
| /* |
| * Always qualify the function name (format_function_signature |
| * won't qualify it). |
| */ |
| appendPQExpBuffer(defqry, "WITH FUNCTION %s.%s", |
| fmtId(funcInfo->dobj.namespace->dobj.name), fsig); |
| free(fsig); |
| } |
| else |
| pg_log_warning("bogus value in pg_cast.castfunc or pg_cast.castmethod field"); |
| break; |
| default: |
| pg_log_warning("bogus value in pg_cast.castmethod field"); |
| } |
| |
| if (cast->castcontext == 'a') |
| appendPQExpBufferStr(defqry, " AS ASSIGNMENT"); |
| else if (cast->castcontext == 'i') |
| appendPQExpBufferStr(defqry, " AS IMPLICIT"); |
| appendPQExpBufferStr(defqry, ";\n"); |
| |
| appendPQExpBuffer(labelq, "CAST (%s AS %s)", |
| sourceType, targetType); |
| |
| appendPQExpBuffer(castargs, "(%s AS %s)", |
| sourceType, targetType); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(defqry, &cast->dobj, |
| "CAST", castargs->data, NULL); |
| |
| if (cast->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, cast->dobj.catId, cast->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = labelq->data, |
| .description = "CAST", |
| .section = SECTION_PRE_DATA, |
| .createStmt = defqry->data, |
| .dropStmt = delqry->data)); |
| |
| /* Dump Cast Comments */ |
| if (cast->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "CAST", castargs->data, |
| NULL, "", |
| cast->dobj.catId, 0, cast->dobj.dumpId); |
| |
| destroyPQExpBuffer(defqry); |
| destroyPQExpBuffer(delqry); |
| destroyPQExpBuffer(labelq); |
| destroyPQExpBuffer(castargs); |
| } |
| |
| /* |
| * Dump a transform |
| */ |
| static void |
| dumpTransform(Archive *fout, const TransformInfo *transform) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer defqry; |
| PQExpBuffer delqry; |
| PQExpBuffer labelq; |
| PQExpBuffer transformargs; |
| FuncInfo *fromsqlFuncInfo = NULL; |
| FuncInfo *tosqlFuncInfo = NULL; |
| char *lanname; |
| const char *transformType; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* Cannot dump if we don't have the transform functions' info */ |
| if (OidIsValid(transform->trffromsql)) |
| { |
| fromsqlFuncInfo = findFuncByOid(transform->trffromsql); |
| if (fromsqlFuncInfo == NULL) |
| pg_fatal("could not find function definition for function with OID %u", |
| transform->trffromsql); |
| } |
| if (OidIsValid(transform->trftosql)) |
| { |
| tosqlFuncInfo = findFuncByOid(transform->trftosql); |
| if (tosqlFuncInfo == NULL) |
| pg_fatal("could not find function definition for function with OID %u", |
| transform->trftosql); |
| } |
| |
| defqry = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| labelq = createPQExpBuffer(); |
| transformargs = createPQExpBuffer(); |
| |
| lanname = get_language_name(fout, transform->trflang); |
| transformType = getFormattedTypeName(fout, transform->trftype, zeroAsNone); |
| |
| appendPQExpBuffer(delqry, "DROP TRANSFORM FOR %s LANGUAGE %s;\n", |
| transformType, lanname); |
| |
| appendPQExpBuffer(defqry, "CREATE TRANSFORM FOR %s LANGUAGE %s (", |
| transformType, lanname); |
| |
| if (!transform->trffromsql && !transform->trftosql) |
| pg_log_warning("bogus transform definition, at least one of trffromsql and trftosql should be nonzero"); |
| |
| if (transform->trffromsql) |
| { |
| if (fromsqlFuncInfo) |
| { |
| char *fsig = format_function_signature(fout, fromsqlFuncInfo, true); |
| |
| /* |
| * Always qualify the function name (format_function_signature |
| * won't qualify it). |
| */ |
| appendPQExpBuffer(defqry, "FROM SQL WITH FUNCTION %s.%s", |
| fmtId(fromsqlFuncInfo->dobj.namespace->dobj.name), fsig); |
| free(fsig); |
| } |
| else |
| pg_log_warning("bogus value in pg_transform.trffromsql field"); |
| } |
| |
| if (transform->trftosql) |
| { |
| if (transform->trffromsql) |
| appendPQExpBufferStr(defqry, ", "); |
| |
| if (tosqlFuncInfo) |
| { |
| char *fsig = format_function_signature(fout, tosqlFuncInfo, true); |
| |
| /* |
| * Always qualify the function name (format_function_signature |
| * won't qualify it). |
| */ |
| appendPQExpBuffer(defqry, "TO SQL WITH FUNCTION %s.%s", |
| fmtId(tosqlFuncInfo->dobj.namespace->dobj.name), fsig); |
| free(fsig); |
| } |
| else |
| pg_log_warning("bogus value in pg_transform.trftosql field"); |
| } |
| |
| appendPQExpBufferStr(defqry, ");\n"); |
| |
| appendPQExpBuffer(labelq, "TRANSFORM FOR %s LANGUAGE %s", |
| transformType, lanname); |
| |
| appendPQExpBuffer(transformargs, "FOR %s LANGUAGE %s", |
| transformType, lanname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(defqry, &transform->dobj, |
| "TRANSFORM", transformargs->data, NULL); |
| |
| if (transform->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, transform->dobj.catId, transform->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = labelq->data, |
| .description = "TRANSFORM", |
| .section = SECTION_PRE_DATA, |
| .createStmt = defqry->data, |
| .dropStmt = delqry->data, |
| .deps = transform->dobj.dependencies, |
| .nDeps = transform->dobj.nDeps)); |
| |
| /* Dump Transform Comments */ |
| if (transform->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TRANSFORM", transformargs->data, |
| NULL, "", |
| transform->dobj.catId, 0, transform->dobj.dumpId); |
| |
| free(lanname); |
| destroyPQExpBuffer(defqry); |
| destroyPQExpBuffer(delqry); |
| destroyPQExpBuffer(labelq); |
| destroyPQExpBuffer(transformargs); |
| } |
| |
| |
| /* |
| * dumpOpr |
| * write out a single operator definition |
| */ |
| static void |
| dumpOpr(Archive *fout, const OprInfo *oprinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer oprid; |
| PQExpBuffer details; |
| PGresult *res; |
| int i_oprkind; |
| int i_oprcode; |
| int i_oprleft; |
| int i_oprright; |
| int i_oprcom; |
| int i_oprnegate; |
| int i_oprrest; |
| int i_oprjoin; |
| int i_oprcanmerge; |
| int i_oprcanhash; |
| char *oprkind; |
| char *oprcode; |
| char *oprleft; |
| char *oprright; |
| char *oprcom; |
| char *oprnegate; |
| char *oprrest; |
| char *oprjoin; |
| char *oprcanmerge; |
| char *oprcanhash; |
| char *oprregproc; |
| char *oprref; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* |
| * some operators are invalid because they were the result of user |
| * defining operators before commutators exist |
| */ |
| if (!OidIsValid(oprinfo->oprcode)) |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| oprid = createPQExpBuffer(); |
| details = createPQExpBuffer(); |
| |
| <<<<<<< HEAD |
| |
| appendPQExpBuffer(query, "SELECT oprkind, " |
| "oprcode::pg_catalog.regprocedure, " |
| "oprleft::pg_catalog.regtype, " |
| "oprright::pg_catalog.regtype, " |
| "oprcom, " |
| "oprnegate, " |
| "oprrest::pg_catalog.regprocedure, " |
| "oprjoin::pg_catalog.regprocedure, " |
| "oprcanmerge, oprcanhash " |
| "FROM pg_catalog.pg_operator " |
| "WHERE oid = '%u'::pg_catalog.oid", |
| oprinfo->dobj.catId.oid); |
| ======= |
| if (!fout->is_prepared[PREPQUERY_DUMPOPR]) |
| { |
| /* Set up query for operator-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpOpr(pg_catalog.oid) AS\n" |
| "SELECT oprkind, " |
| "oprcode::pg_catalog.regprocedure, " |
| "oprleft::pg_catalog.regtype, " |
| "oprright::pg_catalog.regtype, " |
| "oprcom, " |
| "oprnegate, " |
| "oprrest::pg_catalog.regprocedure, " |
| "oprjoin::pg_catalog.regprocedure, " |
| "oprcanmerge, oprcanhash " |
| "FROM pg_catalog.pg_operator " |
| "WHERE oid = $1"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPOPR] = true; |
| } |
| >>>>>>> REL_16_9 |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpOpr('%u')", |
| oprinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| i_oprkind = PQfnumber(res, "oprkind"); |
| i_oprcode = PQfnumber(res, "oprcode"); |
| i_oprleft = PQfnumber(res, "oprleft"); |
| i_oprright = PQfnumber(res, "oprright"); |
| i_oprcom = PQfnumber(res, "oprcom"); |
| i_oprnegate = PQfnumber(res, "oprnegate"); |
| i_oprrest = PQfnumber(res, "oprrest"); |
| i_oprjoin = PQfnumber(res, "oprjoin"); |
| i_oprcanmerge = PQfnumber(res, "oprcanmerge"); |
| i_oprcanhash = PQfnumber(res, "oprcanhash"); |
| |
| oprkind = PQgetvalue(res, 0, i_oprkind); |
| oprcode = PQgetvalue(res, 0, i_oprcode); |
| oprleft = PQgetvalue(res, 0, i_oprleft); |
| oprright = PQgetvalue(res, 0, i_oprright); |
| oprcom = PQgetvalue(res, 0, i_oprcom); |
| oprnegate = PQgetvalue(res, 0, i_oprnegate); |
| oprrest = PQgetvalue(res, 0, i_oprrest); |
| oprjoin = PQgetvalue(res, 0, i_oprjoin); |
| oprcanmerge = PQgetvalue(res, 0, i_oprcanmerge); |
| oprcanhash = PQgetvalue(res, 0, i_oprcanhash); |
| |
| /* In PG14 upwards postfix operator support does not exist anymore. */ |
| if (strcmp(oprkind, "r") == 0) |
| pg_log_warning("postfix operators are not supported anymore (operator \"%s\")", |
| oprcode); |
| |
| oprregproc = convertRegProcReference(oprcode); |
| if (oprregproc) |
| { |
| appendPQExpBuffer(details, " FUNCTION = %s", oprregproc); |
| free(oprregproc); |
| } |
| |
| appendPQExpBuffer(oprid, "%s (", |
| oprinfo->dobj.name); |
| |
| /* |
| * right unary means there's a left arg and left unary means there's a |
| * right arg. (Although the "r" case is dead code for PG14 and later, |
| * continue to support it in case we're dumping from an old server.) |
| */ |
| if (strcmp(oprkind, "r") == 0 || |
| strcmp(oprkind, "b") == 0) |
| { |
| appendPQExpBuffer(details, ",\n LEFTARG = %s", oprleft); |
| appendPQExpBufferStr(oprid, oprleft); |
| } |
| else |
| appendPQExpBufferStr(oprid, "NONE"); |
| |
| if (strcmp(oprkind, "l") == 0 || |
| strcmp(oprkind, "b") == 0) |
| { |
| appendPQExpBuffer(details, ",\n RIGHTARG = %s", oprright); |
| appendPQExpBuffer(oprid, ", %s)", oprright); |
| } |
| else |
| appendPQExpBufferStr(oprid, ", NONE)"); |
| |
| oprref = getFormattedOperatorName(oprcom); |
| if (oprref) |
| { |
| appendPQExpBuffer(details, ",\n COMMUTATOR = %s", oprref); |
| free(oprref); |
| } |
| |
| oprref = getFormattedOperatorName(oprnegate); |
| if (oprref) |
| { |
| appendPQExpBuffer(details, ",\n NEGATOR = %s", oprref); |
| free(oprref); |
| } |
| |
| if (strcmp(oprcanmerge, "t") == 0) |
| appendPQExpBufferStr(details, ",\n MERGES"); |
| |
| if (strcmp(oprcanhash, "t") == 0) |
| appendPQExpBufferStr(details, ",\n HASHES"); |
| |
| oprregproc = convertRegProcReference(oprrest); |
| if (oprregproc) |
| { |
| appendPQExpBuffer(details, ",\n RESTRICT = %s", oprregproc); |
| free(oprregproc); |
| } |
| |
| oprregproc = convertRegProcReference(oprjoin); |
| if (oprregproc) |
| { |
| appendPQExpBuffer(details, ",\n JOIN = %s", oprregproc); |
| free(oprregproc); |
| } |
| |
| appendPQExpBuffer(delq, "DROP OPERATOR %s.%s;\n", |
| fmtId(oprinfo->dobj.namespace->dobj.name), |
| oprid->data); |
| |
| appendPQExpBuffer(q, "CREATE OPERATOR %s.%s (\n%s\n);\n", |
| fmtId(oprinfo->dobj.namespace->dobj.name), |
| oprinfo->dobj.name, details->data); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &oprinfo->dobj, |
| "OPERATOR", oprid->data, |
| oprinfo->dobj.namespace->dobj.name); |
| |
| if (oprinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, oprinfo->dobj.catId, oprinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = oprinfo->dobj.name, |
| .namespace = oprinfo->dobj.namespace->dobj.name, |
| .owner = oprinfo->rolname, |
| .description = "OPERATOR", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Operator Comments */ |
| if (oprinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "OPERATOR", oprid->data, |
| oprinfo->dobj.namespace->dobj.name, oprinfo->rolname, |
| oprinfo->dobj.catId, 0, oprinfo->dobj.dumpId); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(oprid); |
| destroyPQExpBuffer(details); |
| } |
| |
| /* |
| * Convert a function reference obtained from pg_operator |
| * |
| * Returns allocated string of what to print, or NULL if function references |
| * is InvalidOid. Returned string is expected to be free'd by the caller. |
| * |
| * The input is a REGPROCEDURE display; we have to strip the argument-types |
| * part. |
| */ |
| static char * |
| convertRegProcReference(const char *proc) |
| { |
| char *name; |
| char *paren; |
| bool inquote; |
| |
| /* In all cases "-" means a null reference */ |
| if (strcmp(proc, "-") == 0) |
| return NULL; |
| |
| name = pg_strdup(proc); |
| /* find non-double-quoted left paren */ |
| inquote = false; |
| for (paren = name; *paren; paren++) |
| { |
| if (*paren == '(' && !inquote) |
| { |
| *paren = '\0'; |
| break; |
| } |
| if (*paren == '"') |
| inquote = !inquote; |
| } |
| return name; |
| } |
| |
| /* |
| * getFormattedOperatorName - retrieve the operator name for the |
| * given operator OID (presented in string form). |
| * |
| * Returns an allocated string, or NULL if the given OID is invalid. |
| * Caller is responsible for free'ing result string. |
| * |
| * What we produce has the format "OPERATOR(schema.oprname)". This is only |
| * useful in commands where the operator's argument types can be inferred from |
| * context. We always schema-qualify the name, though. The predecessor to |
| * this code tried to skip the schema qualification if possible, but that led |
| * to wrong results in corner cases, such as if an operator and its negator |
| * are in different schemas. |
| */ |
| static char * |
| getFormattedOperatorName(const char *oproid) |
| { |
| OprInfo *oprInfo; |
| |
| /* In all cases "0" means a null reference */ |
| if (strcmp(oproid, "0") == 0) |
| return NULL; |
| |
| oprInfo = findOprByOid(atooid(oproid)); |
| if (oprInfo == NULL) |
| { |
| pg_log_warning("could not find operator with OID %s", |
| oproid); |
| return NULL; |
| } |
| |
| return psprintf("OPERATOR(%s.%s)", |
| fmtId(oprInfo->dobj.namespace->dobj.name), |
| oprInfo->dobj.name); |
| } |
| |
| /* |
| * Convert a function OID obtained from pg_ts_parser or pg_ts_template |
| * |
| * It is sufficient to use REGPROC rather than REGPROCEDURE, since the |
| * argument lists of these functions are predetermined. Note that the |
| * caller should ensure we are in the proper schema, because the results |
| * are search path dependent! |
| */ |
| static char * |
| convertTSFunction(Archive *fout, Oid funcOid) |
| { |
| char *result; |
| char query[128]; |
| PGresult *res; |
| |
| snprintf(query, sizeof(query), |
| "SELECT '%u'::pg_catalog.regproc", funcOid); |
| res = ExecuteSqlQueryForSingleRow(fout, query); |
| |
| result = pg_strdup(PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| |
| return result; |
| } |
| |
| /* |
| * dumpAccessMethod |
| * write out a single access method definition |
| */ |
| static void |
| dumpAccessMethod(Archive *fout, const AccessMethodInfo *aminfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qamname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qamname = pg_strdup(fmtId(aminfo->dobj.name)); |
| |
| appendPQExpBuffer(q, "CREATE ACCESS METHOD %s ", qamname); |
| |
| switch (aminfo->amtype) |
| { |
| case AMTYPE_INDEX: |
| appendPQExpBufferStr(q, "TYPE INDEX "); |
| break; |
| case AMTYPE_TABLE: |
| appendPQExpBufferStr(q, "TYPE TABLE "); |
| break; |
| default: |
| pg_log_warning("invalid type \"%c\" of access method \"%s\"", |
| aminfo->amtype, qamname); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qamname); |
| return; |
| } |
| |
| appendPQExpBuffer(q, "HANDLER %s;\n", aminfo->amhandler); |
| |
| appendPQExpBuffer(delq, "DROP ACCESS METHOD %s;\n", |
| qamname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &aminfo->dobj, |
| "ACCESS METHOD", qamname, NULL); |
| |
| if (aminfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, aminfo->dobj.catId, aminfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = aminfo->dobj.name, |
| .description = "ACCESS METHOD", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Access Method Comments */ |
| if (aminfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "ACCESS METHOD", qamname, |
| NULL, "", |
| aminfo->dobj.catId, 0, aminfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qamname); |
| } |
| |
| /* |
| * dumpOpclass |
| * write out a single operator class definition |
| */ |
| static void |
| dumpOpclass(Archive *fout, const OpclassInfo *opcinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer nameusing; |
| PGresult *res; |
| int ntups; |
| int i_opcintype; |
| int i_opckeytype; |
| int i_opcdefault; |
| int i_opcfamily; |
| int i_opcfamilyname; |
| int i_opcfamilynsp; |
| int i_amname; |
| int i_amopstrategy; |
| int i_amopopr; |
| int i_sortfamily; |
| int i_sortfamilynsp; |
| int i_amprocnum; |
| int i_amproc; |
| int i_amproclefttype; |
| int i_amprocrighttype; |
| char *opcintype; |
| char *opckeytype; |
| char *opcdefault; |
| char *opcfamily; |
| char *opcfamilyname; |
| char *opcfamilynsp; |
| char *amname; |
| char *amopstrategy; |
| char *amopopr; |
| char *sortfamily; |
| char *sortfamilynsp; |
| char *amprocnum; |
| char *amproc; |
| char *amproclefttype; |
| char *amprocrighttype; |
| bool needComma; |
| int i; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| nameusing = createPQExpBuffer(); |
| |
| /* Get additional fields from the pg_opclass row */ |
| appendPQExpBuffer(query, "SELECT opcintype::pg_catalog.regtype, " |
| "opckeytype::pg_catalog.regtype, " |
| "opcdefault, opcfamily, " |
| "opfname AS opcfamilyname, " |
| "nspname AS opcfamilynsp, " |
| "(SELECT amname FROM pg_catalog.pg_am WHERE oid = opcmethod) AS amname " |
| "FROM pg_catalog.pg_opclass c " |
| "LEFT JOIN pg_catalog.pg_opfamily f ON f.oid = opcfamily " |
| "LEFT JOIN pg_catalog.pg_namespace n ON n.oid = opfnamespace " |
| "WHERE c.oid = '%u'::pg_catalog.oid", |
| opcinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| i_opcintype = PQfnumber(res, "opcintype"); |
| i_opckeytype = PQfnumber(res, "opckeytype"); |
| i_opcdefault = PQfnumber(res, "opcdefault"); |
| i_opcfamily = PQfnumber(res, "opcfamily"); |
| i_opcfamilyname = PQfnumber(res, "opcfamilyname"); |
| i_opcfamilynsp = PQfnumber(res, "opcfamilynsp"); |
| i_amname = PQfnumber(res, "amname"); |
| |
| /* opcintype may still be needed after we PQclear res */ |
| opcintype = pg_strdup(PQgetvalue(res, 0, i_opcintype)); |
| opckeytype = PQgetvalue(res, 0, i_opckeytype); |
| opcdefault = PQgetvalue(res, 0, i_opcdefault); |
| /* opcfamily will still be needed after we PQclear res */ |
| opcfamily = pg_strdup(PQgetvalue(res, 0, i_opcfamily)); |
| opcfamilyname = PQgetvalue(res, 0, i_opcfamilyname); |
| opcfamilynsp = PQgetvalue(res, 0, i_opcfamilynsp); |
| /* amname will still be needed after we PQclear res */ |
| amname = pg_strdup(PQgetvalue(res, 0, i_amname)); |
| |
| appendPQExpBuffer(delq, "DROP OPERATOR CLASS %s", |
| fmtQualifiedDumpable(opcinfo)); |
| appendPQExpBuffer(delq, " USING %s;\n", |
| fmtId(amname)); |
| |
| /* Build the fixed portion of the CREATE command */ |
| appendPQExpBuffer(q, "CREATE OPERATOR CLASS %s\n ", |
| fmtQualifiedDumpable(opcinfo)); |
| if (strcmp(opcdefault, "t") == 0) |
| appendPQExpBufferStr(q, "DEFAULT "); |
| appendPQExpBuffer(q, "FOR TYPE %s USING %s", |
| opcintype, |
| fmtId(amname)); |
| if (strlen(opcfamilyname) > 0) |
| { |
| appendPQExpBufferStr(q, " FAMILY "); |
| appendPQExpBuffer(q, "%s.", fmtId(opcfamilynsp)); |
| appendPQExpBufferStr(q, fmtId(opcfamilyname)); |
| } |
| appendPQExpBufferStr(q, " AS\n "); |
| |
| needComma = false; |
| |
| if (strcmp(opckeytype, "-") != 0) |
| { |
| appendPQExpBuffer(q, "STORAGE %s", |
| opckeytype); |
| needComma = true; |
| } |
| |
| PQclear(res); |
| |
| /* |
| * Now fetch and print the OPERATOR entries (pg_amop rows). |
| * |
| * Print only those opfamily members that are tied to the opclass by |
| * pg_depend entries. |
| */ |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, "SELECT amopstrategy, " |
| "amopopr::pg_catalog.regoperator, " |
| "opfname AS sortfamily, " |
| "nspname AS sortfamilynsp " |
| "FROM pg_catalog.pg_amop ao JOIN pg_catalog.pg_depend ON " |
| "(classid = 'pg_catalog.pg_amop'::pg_catalog.regclass AND objid = ao.oid) " |
| "LEFT JOIN pg_catalog.pg_opfamily f ON f.oid = amopsortfamily " |
| "LEFT JOIN pg_catalog.pg_namespace n ON n.oid = opfnamespace " |
| "WHERE refclassid = 'pg_catalog.pg_opclass'::pg_catalog.regclass " |
| "AND refobjid = '%u'::pg_catalog.oid " |
| "AND amopfamily = '%s'::pg_catalog.oid " |
| "ORDER BY amopstrategy", |
| opcinfo->dobj.catId.oid, |
| opcfamily); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_amopstrategy = PQfnumber(res, "amopstrategy"); |
| i_amopopr = PQfnumber(res, "amopopr"); |
| i_sortfamily = PQfnumber(res, "sortfamily"); |
| i_sortfamilynsp = PQfnumber(res, "sortfamilynsp"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| amopstrategy = PQgetvalue(res, i, i_amopstrategy); |
| amopopr = PQgetvalue(res, i, i_amopopr); |
| sortfamily = PQgetvalue(res, i, i_sortfamily); |
| sortfamilynsp = PQgetvalue(res, i, i_sortfamilynsp); |
| |
| if (needComma) |
| appendPQExpBufferStr(q, " ,\n "); |
| |
| appendPQExpBuffer(q, "OPERATOR %s %s", |
| amopstrategy, amopopr); |
| |
| if (strlen(sortfamily) > 0) |
| { |
| appendPQExpBufferStr(q, " FOR ORDER BY "); |
| appendPQExpBuffer(q, "%s.", fmtId(sortfamilynsp)); |
| appendPQExpBufferStr(q, fmtId(sortfamily)); |
| } |
| |
| needComma = true; |
| } |
| |
| PQclear(res); |
| |
| /* |
| * Now fetch and print the FUNCTION entries (pg_amproc rows). |
| * |
| * Print only those opfamily members that are tied to the opclass by |
| * pg_depend entries. |
| * |
| * We print the amproclefttype/amprocrighttype even though in most cases |
| * the backend could deduce the right values, because of the corner case |
| * of a btree sort support function for a cross-type comparison. |
| */ |
| resetPQExpBuffer(query); |
| |
| appendPQExpBuffer(query, "SELECT amprocnum, " |
| "amproc::pg_catalog.regprocedure, " |
| "amproclefttype::pg_catalog.regtype, " |
| "amprocrighttype::pg_catalog.regtype " |
| "FROM pg_catalog.pg_amproc ap, pg_catalog.pg_depend " |
| "WHERE refclassid = 'pg_catalog.pg_opclass'::pg_catalog.regclass " |
| "AND refobjid = '%u'::pg_catalog.oid " |
| "AND classid = 'pg_catalog.pg_amproc'::pg_catalog.regclass " |
| "AND objid = ap.oid " |
| "ORDER BY amprocnum", |
| opcinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_amprocnum = PQfnumber(res, "amprocnum"); |
| i_amproc = PQfnumber(res, "amproc"); |
| i_amproclefttype = PQfnumber(res, "amproclefttype"); |
| i_amprocrighttype = PQfnumber(res, "amprocrighttype"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| amprocnum = PQgetvalue(res, i, i_amprocnum); |
| amproc = PQgetvalue(res, i, i_amproc); |
| amproclefttype = PQgetvalue(res, i, i_amproclefttype); |
| amprocrighttype = PQgetvalue(res, i, i_amprocrighttype); |
| |
| if (needComma) |
| appendPQExpBufferStr(q, " ,\n "); |
| |
| appendPQExpBuffer(q, "FUNCTION %s", amprocnum); |
| |
| if (*amproclefttype && *amprocrighttype) |
| appendPQExpBuffer(q, " (%s, %s)", amproclefttype, amprocrighttype); |
| |
| appendPQExpBuffer(q, " %s", amproc); |
| |
| needComma = true; |
| } |
| |
| PQclear(res); |
| |
| /* |
| * If needComma is still false it means we haven't added anything after |
| * the AS keyword. To avoid printing broken SQL, append a dummy STORAGE |
| * clause with the same datatype. This isn't sanctioned by the |
| * documentation, but actually DefineOpClass will treat it as a no-op. |
| */ |
| if (!needComma) |
| appendPQExpBuffer(q, "STORAGE %s", opcintype); |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| appendPQExpBufferStr(nameusing, fmtId(opcinfo->dobj.name)); |
| appendPQExpBuffer(nameusing, " USING %s", |
| fmtId(amname)); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &opcinfo->dobj, |
| "OPERATOR CLASS", nameusing->data, |
| opcinfo->dobj.namespace->dobj.name); |
| |
| if (opcinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, opcinfo->dobj.catId, opcinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = opcinfo->dobj.name, |
| .namespace = opcinfo->dobj.namespace->dobj.name, |
| .owner = opcinfo->rolname, |
| .description = "OPERATOR CLASS", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Operator Class Comments */ |
| if (opcinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "OPERATOR CLASS", nameusing->data, |
| opcinfo->dobj.namespace->dobj.name, opcinfo->rolname, |
| opcinfo->dobj.catId, 0, opcinfo->dobj.dumpId); |
| |
| free(opcintype); |
| free(opcfamily); |
| free(amname); |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(nameusing); |
| } |
| |
| /* |
| * dumpOpfamily |
| * write out a single operator family definition |
| * |
| * Note: this also dumps any "loose" operator members that aren't bound to a |
| * specific opclass within the opfamily. |
| */ |
| static void |
| dumpOpfamily(Archive *fout, const OpfamilyInfo *opfinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer nameusing; |
| PGresult *res; |
| PGresult *res_ops; |
| PGresult *res_procs; |
| int ntups; |
| int i_amname; |
| int i_amopstrategy; |
| int i_amopopr; |
| int i_sortfamily; |
| int i_sortfamilynsp; |
| int i_amprocnum; |
| int i_amproc; |
| int i_amproclefttype; |
| int i_amprocrighttype; |
| char *amname; |
| char *amopstrategy; |
| char *amopopr; |
| char *sortfamily; |
| char *sortfamilynsp; |
| char *amprocnum; |
| char *amproc; |
| char *amproclefttype; |
| char *amprocrighttype; |
| bool needComma; |
| int i; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| nameusing = createPQExpBuffer(); |
| |
| /* |
| * Fetch only those opfamily members that are tied directly to the |
| * opfamily by pg_depend entries. |
| */ |
| appendPQExpBuffer(query, "SELECT amopstrategy, " |
| "amopopr::pg_catalog.regoperator, " |
| "opfname AS sortfamily, " |
| "nspname AS sortfamilynsp " |
| "FROM pg_catalog.pg_amop ao JOIN pg_catalog.pg_depend ON " |
| "(classid = 'pg_catalog.pg_amop'::pg_catalog.regclass AND objid = ao.oid) " |
| "LEFT JOIN pg_catalog.pg_opfamily f ON f.oid = amopsortfamily " |
| "LEFT JOIN pg_catalog.pg_namespace n ON n.oid = opfnamespace " |
| "WHERE refclassid = 'pg_catalog.pg_opfamily'::pg_catalog.regclass " |
| "AND refobjid = '%u'::pg_catalog.oid " |
| "AND amopfamily = '%u'::pg_catalog.oid " |
| "ORDER BY amopstrategy", |
| opfinfo->dobj.catId.oid, |
| opfinfo->dobj.catId.oid); |
| |
| res_ops = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| resetPQExpBuffer(query); |
| |
| appendPQExpBuffer(query, "SELECT amprocnum, " |
| "amproc::pg_catalog.regprocedure, " |
| "amproclefttype::pg_catalog.regtype, " |
| "amprocrighttype::pg_catalog.regtype " |
| "FROM pg_catalog.pg_amproc ap, pg_catalog.pg_depend " |
| "WHERE refclassid = 'pg_catalog.pg_opfamily'::pg_catalog.regclass " |
| "AND refobjid = '%u'::pg_catalog.oid " |
| "AND classid = 'pg_catalog.pg_amproc'::pg_catalog.regclass " |
| "AND objid = ap.oid " |
| "ORDER BY amprocnum", |
| opfinfo->dobj.catId.oid); |
| |
| res_procs = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| /* Get additional fields from the pg_opfamily row */ |
| resetPQExpBuffer(query); |
| |
| appendPQExpBuffer(query, "SELECT " |
| "(SELECT amname FROM pg_catalog.pg_am WHERE oid = opfmethod) AS amname " |
| "FROM pg_catalog.pg_opfamily " |
| "WHERE oid = '%u'::pg_catalog.oid", |
| opfinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| i_amname = PQfnumber(res, "amname"); |
| |
| /* amname will still be needed after we PQclear res */ |
| amname = pg_strdup(PQgetvalue(res, 0, i_amname)); |
| |
| appendPQExpBuffer(delq, "DROP OPERATOR FAMILY %s", |
| fmtQualifiedDumpable(opfinfo)); |
| appendPQExpBuffer(delq, " USING %s;\n", |
| fmtId(amname)); |
| |
| /* Build the fixed portion of the CREATE command */ |
| appendPQExpBuffer(q, "CREATE OPERATOR FAMILY %s", |
| fmtQualifiedDumpable(opfinfo)); |
| appendPQExpBuffer(q, " USING %s;\n", |
| fmtId(amname)); |
| |
| PQclear(res); |
| |
| /* Do we need an ALTER to add loose members? */ |
| if (PQntuples(res_ops) > 0 || PQntuples(res_procs) > 0) |
| { |
| appendPQExpBuffer(q, "ALTER OPERATOR FAMILY %s", |
| fmtQualifiedDumpable(opfinfo)); |
| appendPQExpBuffer(q, " USING %s ADD\n ", |
| fmtId(amname)); |
| |
| needComma = false; |
| |
| /* |
| * Now fetch and print the OPERATOR entries (pg_amop rows). |
| */ |
| ntups = PQntuples(res_ops); |
| |
| i_amopstrategy = PQfnumber(res_ops, "amopstrategy"); |
| i_amopopr = PQfnumber(res_ops, "amopopr"); |
| i_sortfamily = PQfnumber(res_ops, "sortfamily"); |
| i_sortfamilynsp = PQfnumber(res_ops, "sortfamilynsp"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| amopstrategy = PQgetvalue(res_ops, i, i_amopstrategy); |
| amopopr = PQgetvalue(res_ops, i, i_amopopr); |
| sortfamily = PQgetvalue(res_ops, i, i_sortfamily); |
| sortfamilynsp = PQgetvalue(res_ops, i, i_sortfamilynsp); |
| |
| if (needComma) |
| appendPQExpBufferStr(q, " ,\n "); |
| |
| appendPQExpBuffer(q, "OPERATOR %s %s", |
| amopstrategy, amopopr); |
| |
| if (strlen(sortfamily) > 0) |
| { |
| appendPQExpBufferStr(q, " FOR ORDER BY "); |
| appendPQExpBuffer(q, "%s.", fmtId(sortfamilynsp)); |
| appendPQExpBufferStr(q, fmtId(sortfamily)); |
| } |
| |
| needComma = true; |
| } |
| |
| /* |
| * Now fetch and print the FUNCTION entries (pg_amproc rows). |
| */ |
| ntups = PQntuples(res_procs); |
| |
| i_amprocnum = PQfnumber(res_procs, "amprocnum"); |
| i_amproc = PQfnumber(res_procs, "amproc"); |
| i_amproclefttype = PQfnumber(res_procs, "amproclefttype"); |
| i_amprocrighttype = PQfnumber(res_procs, "amprocrighttype"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| amprocnum = PQgetvalue(res_procs, i, i_amprocnum); |
| amproc = PQgetvalue(res_procs, i, i_amproc); |
| amproclefttype = PQgetvalue(res_procs, i, i_amproclefttype); |
| amprocrighttype = PQgetvalue(res_procs, i, i_amprocrighttype); |
| |
| if (needComma) |
| appendPQExpBufferStr(q, " ,\n "); |
| |
| appendPQExpBuffer(q, "FUNCTION %s (%s, %s) %s", |
| amprocnum, amproclefttype, amprocrighttype, |
| amproc); |
| |
| needComma = true; |
| } |
| |
| appendPQExpBufferStr(q, ";\n"); |
| } |
| |
| appendPQExpBufferStr(nameusing, fmtId(opfinfo->dobj.name)); |
| appendPQExpBuffer(nameusing, " USING %s", |
| fmtId(amname)); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &opfinfo->dobj, |
| "OPERATOR FAMILY", nameusing->data, |
| opfinfo->dobj.namespace->dobj.name); |
| |
| if (opfinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, opfinfo->dobj.catId, opfinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = opfinfo->dobj.name, |
| .namespace = opfinfo->dobj.namespace->dobj.name, |
| .owner = opfinfo->rolname, |
| .description = "OPERATOR FAMILY", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Operator Family Comments */ |
| if (opfinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "OPERATOR FAMILY", nameusing->data, |
| opfinfo->dobj.namespace->dobj.name, opfinfo->rolname, |
| opfinfo->dobj.catId, 0, opfinfo->dobj.dumpId); |
| |
| free(amname); |
| PQclear(res_ops); |
| PQclear(res_procs); |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(nameusing); |
| } |
| |
| /* |
| * dumpCollation |
| * write out a single collation definition |
| */ |
| static void |
| dumpCollation(Archive *fout, const CollInfo *collinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qcollname; |
| PGresult *res; |
| int i_collprovider; |
| int i_collisdeterministic; |
| int i_collcollate; |
| int i_collctype; |
| int i_colliculocale; |
| int i_collicurules; |
| const char *collprovider; |
| const char *collcollate; |
| const char *collctype; |
| const char *colliculocale; |
| const char *collicurules; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qcollname = pg_strdup(fmtId(collinfo->dobj.name)); |
| |
| /* Get collation-specific details */ |
| appendPQExpBufferStr(query, "SELECT "); |
| |
| if (fout->remoteVersion >= 100000) |
| appendPQExpBufferStr(query, |
| "collprovider, " |
| "collversion, "); |
| else |
| appendPQExpBufferStr(query, |
| "'c' AS collprovider, " |
| "NULL AS collversion, "); |
| |
| if (fout->remoteVersion >= 120000) |
| appendPQExpBufferStr(query, |
| "collisdeterministic, "); |
| else |
| appendPQExpBufferStr(query, |
| "true AS collisdeterministic, "); |
| |
| if (fout->remoteVersion >= 150000) |
| appendPQExpBufferStr(query, |
| "colliculocale, "); |
| else |
| appendPQExpBufferStr(query, |
| "NULL AS colliculocale, "); |
| |
| if (fout->remoteVersion >= 160000) |
| appendPQExpBufferStr(query, |
| "collicurules, "); |
| else |
| appendPQExpBufferStr(query, |
| "NULL AS collicurules, "); |
| |
| appendPQExpBuffer(query, |
| "collcollate, " |
| "collctype " |
| "FROM pg_catalog.pg_collation c " |
| "WHERE c.oid = '%u'::pg_catalog.oid", |
| collinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| i_collprovider = PQfnumber(res, "collprovider"); |
| i_collisdeterministic = PQfnumber(res, "collisdeterministic"); |
| i_collcollate = PQfnumber(res, "collcollate"); |
| i_collctype = PQfnumber(res, "collctype"); |
| i_colliculocale = PQfnumber(res, "colliculocale"); |
| i_collicurules = PQfnumber(res, "collicurules"); |
| |
| collprovider = PQgetvalue(res, 0, i_collprovider); |
| |
| if (!PQgetisnull(res, 0, i_collcollate)) |
| collcollate = PQgetvalue(res, 0, i_collcollate); |
| else |
| collcollate = NULL; |
| |
| if (!PQgetisnull(res, 0, i_collctype)) |
| collctype = PQgetvalue(res, 0, i_collctype); |
| else |
| collctype = NULL; |
| |
| /* |
| * Before version 15, collcollate and collctype were of type NAME and |
| * non-nullable. Treat empty strings as NULL for consistency. |
| */ |
| if (fout->remoteVersion < 150000) |
| { |
| if (collcollate[0] == '\0') |
| collcollate = NULL; |
| if (collctype[0] == '\0') |
| collctype = NULL; |
| } |
| |
| if (!PQgetisnull(res, 0, i_colliculocale)) |
| colliculocale = PQgetvalue(res, 0, i_colliculocale); |
| else |
| colliculocale = NULL; |
| |
| if (!PQgetisnull(res, 0, i_collicurules)) |
| collicurules = PQgetvalue(res, 0, i_collicurules); |
| else |
| collicurules = NULL; |
| |
| appendPQExpBuffer(delq, "DROP COLLATION %s;\n", |
| fmtQualifiedDumpable(collinfo)); |
| |
| appendPQExpBuffer(q, "CREATE COLLATION %s (", |
| fmtQualifiedDumpable(collinfo)); |
| |
| appendPQExpBufferStr(q, "provider = "); |
| if (collprovider[0] == 'c') |
| appendPQExpBufferStr(q, "libc"); |
| else if (collprovider[0] == 'i') |
| appendPQExpBufferStr(q, "icu"); |
| else if (collprovider[0] == 'd') |
| /* to allow dumping pg_catalog; not accepted on input */ |
| appendPQExpBufferStr(q, "default"); |
| else |
| pg_fatal("unrecognized collation provider: %s", |
| collprovider); |
| |
| if (strcmp(PQgetvalue(res, 0, i_collisdeterministic), "f") == 0) |
| appendPQExpBufferStr(q, ", deterministic = false"); |
| |
| if (collprovider[0] == 'd') |
| { |
| if (collcollate || collctype || colliculocale || collicurules) |
| pg_log_warning("invalid collation \"%s\"", qcollname); |
| |
| /* no locale -- the default collation cannot be reloaded anyway */ |
| } |
| else if (collprovider[0] == 'i') |
| { |
| if (fout->remoteVersion >= 150000) |
| { |
| if (collcollate || collctype || !colliculocale) |
| pg_log_warning("invalid collation \"%s\"", qcollname); |
| |
| appendPQExpBufferStr(q, ", locale = "); |
| appendStringLiteralAH(q, colliculocale ? colliculocale : "", |
| fout); |
| } |
| else |
| { |
| if (!collcollate || !collctype || colliculocale || |
| strcmp(collcollate, collctype) != 0) |
| pg_log_warning("invalid collation \"%s\"", qcollname); |
| |
| appendPQExpBufferStr(q, ", locale = "); |
| appendStringLiteralAH(q, collcollate ? collcollate : "", fout); |
| } |
| |
| if (collicurules) |
| { |
| appendPQExpBufferStr(q, ", rules = "); |
| appendStringLiteralAH(q, collicurules ? collicurules : "", fout); |
| } |
| } |
| else if (collprovider[0] == 'c') |
| { |
| if (colliculocale || collicurules || !collcollate || !collctype) |
| pg_log_warning("invalid collation \"%s\"", qcollname); |
| |
| if (collcollate && collctype && strcmp(collcollate, collctype) == 0) |
| { |
| appendPQExpBufferStr(q, ", locale = "); |
| appendStringLiteralAH(q, collcollate ? collcollate : "", fout); |
| } |
| else |
| { |
| appendPQExpBufferStr(q, ", lc_collate = "); |
| appendStringLiteralAH(q, collcollate ? collcollate : "", fout); |
| appendPQExpBufferStr(q, ", lc_ctype = "); |
| appendStringLiteralAH(q, collctype ? collctype : "", fout); |
| } |
| } |
| else |
| pg_fatal("unrecognized collation provider: %s", collprovider); |
| |
| /* |
| * For binary upgrade, carry over the collation version. For normal |
| * dump/restore, omit the version, so that it is computed upon restore. |
| */ |
| if (dopt->binary_upgrade) |
| { |
| int i_collversion; |
| |
| i_collversion = PQfnumber(res, "collversion"); |
| if (!PQgetisnull(res, 0, i_collversion)) |
| { |
| appendPQExpBufferStr(q, ", version = "); |
| appendStringLiteralAH(q, |
| PQgetvalue(res, 0, i_collversion), |
| fout); |
| } |
| } |
| |
| appendPQExpBufferStr(q, ");\n"); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &collinfo->dobj, |
| "COLLATION", qcollname, |
| collinfo->dobj.namespace->dobj.name); |
| |
| if (collinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, collinfo->dobj.catId, collinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = collinfo->dobj.name, |
| .namespace = collinfo->dobj.namespace->dobj.name, |
| .owner = collinfo->rolname, |
| .description = "COLLATION", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Collation Comments */ |
| if (collinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "COLLATION", qcollname, |
| collinfo->dobj.namespace->dobj.name, collinfo->rolname, |
| collinfo->dobj.catId, 0, collinfo->dobj.dumpId); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qcollname); |
| } |
| |
| /* |
| * dumpConversion |
| * write out a single conversion definition |
| */ |
| static void |
| dumpConversion(Archive *fout, const ConvInfo *convinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qconvname; |
| PGresult *res; |
| int i_conforencoding; |
| int i_contoencoding; |
| int i_conproc; |
| int i_condefault; |
| const char *conforencoding; |
| const char *contoencoding; |
| const char *conproc; |
| bool condefault; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qconvname = pg_strdup(fmtId(convinfo->dobj.name)); |
| |
| /* Get conversion-specific details */ |
| appendPQExpBuffer(query, "SELECT " |
| "pg_catalog.pg_encoding_to_char(conforencoding) AS conforencoding, " |
| "pg_catalog.pg_encoding_to_char(contoencoding) AS contoencoding, " |
| "conproc, condefault " |
| "FROM pg_catalog.pg_conversion c " |
| "WHERE c.oid = '%u'::pg_catalog.oid", |
| convinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| i_conforencoding = PQfnumber(res, "conforencoding"); |
| i_contoencoding = PQfnumber(res, "contoencoding"); |
| i_conproc = PQfnumber(res, "conproc"); |
| i_condefault = PQfnumber(res, "condefault"); |
| |
| conforencoding = PQgetvalue(res, 0, i_conforencoding); |
| contoencoding = PQgetvalue(res, 0, i_contoencoding); |
| conproc = PQgetvalue(res, 0, i_conproc); |
| condefault = (PQgetvalue(res, 0, i_condefault)[0] == 't'); |
| |
| appendPQExpBuffer(delq, "DROP CONVERSION %s;\n", |
| fmtQualifiedDumpable(convinfo)); |
| |
| appendPQExpBuffer(q, "CREATE %sCONVERSION %s FOR ", |
| (condefault) ? "DEFAULT " : "", |
| fmtQualifiedDumpable(convinfo)); |
| appendStringLiteralAH(q, conforencoding, fout); |
| appendPQExpBufferStr(q, " TO "); |
| appendStringLiteralAH(q, contoencoding, fout); |
| /* regproc output is already sufficiently quoted */ |
| appendPQExpBuffer(q, " FROM %s;\n", conproc); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &convinfo->dobj, |
| "CONVERSION", qconvname, |
| convinfo->dobj.namespace->dobj.name); |
| |
| if (convinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, convinfo->dobj.catId, convinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = convinfo->dobj.name, |
| .namespace = convinfo->dobj.namespace->dobj.name, |
| .owner = convinfo->rolname, |
| .description = "CONVERSION", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Conversion Comments */ |
| if (convinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "CONVERSION", qconvname, |
| convinfo->dobj.namespace->dobj.name, convinfo->rolname, |
| convinfo->dobj.catId, 0, convinfo->dobj.dumpId); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qconvname); |
| } |
| |
| /* |
| * format_aggregate_signature: generate aggregate name and argument list |
| * |
| * The argument type names are qualified if needed. The aggregate name |
| * is never qualified. |
| */ |
| static char * |
| format_aggregate_signature(const AggInfo *agginfo, Archive *fout, bool honor_quotes) |
| { |
| PQExpBufferData buf; |
| int j; |
| |
| initPQExpBuffer(&buf); |
| if (honor_quotes) |
| appendPQExpBufferStr(&buf, fmtId(agginfo->aggfn.dobj.name)); |
| else |
| appendPQExpBufferStr(&buf, agginfo->aggfn.dobj.name); |
| |
| if (agginfo->aggfn.nargs == 0) |
| appendPQExpBufferStr(&buf, "(*)"); |
| else |
| { |
| appendPQExpBufferChar(&buf, '('); |
| for (j = 0; j < agginfo->aggfn.nargs; j++) |
| appendPQExpBuffer(&buf, "%s%s", |
| (j > 0) ? ", " : "", |
| getFormattedTypeName(fout, |
| agginfo->aggfn.argtypes[j], |
| zeroIsError)); |
| appendPQExpBufferChar(&buf, ')'); |
| } |
| return buf.data; |
| } |
| |
| /* |
| * dumpAgg |
| * write out a single aggregate definition |
| */ |
| static void |
| dumpAgg(Archive *fout, const AggInfo *agginfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer details; |
| char *aggsig; /* identity signature */ |
| char *aggfullsig = NULL; /* full signature */ |
| char *aggsig_tag; |
| PGresult *res; |
| int i_agginitval; |
| int i_aggminitval; |
| const char *aggtransfn; |
| const char *aggfinalfn; |
| const char *aggcombinefn; |
| const char *aggserialfn; |
| const char *aggdeserialfn; |
| const char *aggmtransfn; |
| const char *aggminvtransfn; |
| const char *aggmfinalfn; |
| bool aggfinalextra; |
| bool aggmfinalextra; |
| char aggfinalmodify; |
| char aggmfinalmodify; |
| const char *aggsortop; |
| char *aggsortconvop; |
| char aggkind; |
| const char *aggtranstype; |
| const char *aggtransspace; |
| const char *aggmtranstype; |
| const char *aggmtransspace; |
| const char *agginitval; |
| const char *aggminitval; |
| bool aggrepsafeexec; |
| const char *proparallel; |
| char defaultfinalmodify; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| details = createPQExpBuffer(); |
| |
| <<<<<<< HEAD |
| |
| /* |
| * GPDB_14_MERGE_FIXME: |
| * Some fields: |
| * aggcombinefn, aggserialfn, |
| * aggdeserialfn, aggmtransfn, |
| * aggminvtransfn, aggmfinalfn, |
| * aggmtranstype::pg_catalog.regtype, aggfinalextra, |
| * aggmfinalextra, |
| * diffs from upsreaam in certain versions between [80100, 90600]. |
| * But others versions are same. |
| * I think we should keep the same with upsream here, special |
| * handles are not need in such versions. |
| * Need to recheck. |
| */ |
| |
| /* Get aggregate-specific details */ |
| appendPQExpBufferStr(query, |
| "SELECT\n" |
| "aggtransfn,\n" |
| "aggfinalfn,\n" |
| "aggtranstype::pg_catalog.regtype,\n" |
| "agginitval,\n"); |
| |
| if (fout->remoteVersion >= 80100) |
| appendPQExpBufferStr(query, |
| "aggsortop,\n"); |
| |
| if (fout->remoteVersion >= 80400) |
| ======= |
| if (!fout->is_prepared[PREPQUERY_DUMPAGG]) |
| { |
| /* Set up query for aggregate-specific details */ |
| appendPQExpBufferStr(query, |
| "PREPARE dumpAgg(pg_catalog.oid) AS\n"); |
| |
| >>>>>>> REL_16_9 |
| appendPQExpBufferStr(query, |
| "SELECT " |
| "aggtransfn,\n" |
| "aggfinalfn,\n" |
| "aggtranstype::pg_catalog.regtype,\n" |
| "agginitval,\n" |
| "aggsortop,\n" |
| "pg_catalog.pg_get_function_arguments(p.oid) AS funcargs,\n" |
| "pg_catalog.pg_get_function_identity_arguments(p.oid) AS funciargs,\n"); |
| |
| if (fout->remoteVersion >= 90400) |
| appendPQExpBufferStr(query, |
| "aggkind,\n" |
| "aggmtransfn,\n" |
| "aggminvtransfn,\n" |
| "aggmfinalfn,\n" |
| "aggmtranstype::pg_catalog.regtype,\n" |
| "aggfinalextra,\n" |
| "aggmfinalextra,\n" |
| "aggtransspace,\n" |
| "aggmtransspace,\n" |
| "aggminitval,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "'n' AS aggkind,\n" |
| "'-' AS aggmtransfn,\n" |
| "'-' AS aggminvtransfn,\n" |
| "'-' AS aggmfinalfn,\n" |
| "0 AS aggmtranstype,\n" |
| "false AS aggfinalextra,\n" |
| "false AS aggmfinalextra,\n" |
| "0 AS aggtransspace,\n" |
| "0 AS aggmtransspace,\n" |
| "NULL AS aggminitval,\n"); |
| |
| if (fout->remoteVersion >= 90600) |
| appendPQExpBufferStr(query, |
| "aggcombinefn,\n" |
| "aggserialfn,\n" |
| "aggdeserialfn,\n" |
| "proparallel,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "'-' AS aggcombinefn,\n" |
| "'-' AS aggserialfn,\n" |
| "'-' AS aggdeserialfn,\n" |
| "'u' AS proparallel,\n"); |
| |
| if (fout->remoteVersion >= 110000) |
| appendPQExpBufferStr(query, |
| "aggfinalmodify,\n" |
| "aggmfinalmodify\n"); |
| else |
| appendPQExpBufferStr(query, |
| "'0' AS aggfinalmodify,\n" |
| "'0' AS aggmfinalmodify\n"); |
| |
| <<<<<<< HEAD |
| if (fout->remoteVersion >= 140000) |
| appendPQExpBufferStr(query, |
| "aggrepsafeexec,\n"); |
| else |
| appendPQExpBufferStr(query, |
| "false AS aggrepsafeexec,\n"); |
| |
| appendPQExpBuffer(query, |
| "FROM pg_catalog.pg_aggregate a, pg_catalog.pg_proc p " |
| "WHERE a.aggfnoid = p.oid " |
| "AND p.oid = '%u'::pg_catalog.oid", |
| ======= |
| appendPQExpBufferStr(query, |
| "FROM pg_catalog.pg_aggregate a, pg_catalog.pg_proc p " |
| "WHERE a.aggfnoid = p.oid " |
| "AND p.oid = $1"); |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPAGG] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE dumpAgg('%u')", |
| >>>>>>> REL_16_9 |
| agginfo->aggfn.dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| i_agginitval = PQfnumber(res, "agginitval"); |
| i_aggminitval = PQfnumber(res, "aggminitval"); |
| |
| aggtransfn = PQgetvalue(res, 0, PQfnumber(res, "aggtransfn")); |
| aggfinalfn = PQgetvalue(res, 0, PQfnumber(res, "aggfinalfn")); |
| aggcombinefn = PQgetvalue(res, 0, PQfnumber(res, "aggcombinefn")); |
| aggserialfn = PQgetvalue(res, 0, PQfnumber(res, "aggserialfn")); |
| aggdeserialfn = PQgetvalue(res, 0, PQfnumber(res, "aggdeserialfn")); |
| aggmtransfn = PQgetvalue(res, 0, PQfnumber(res, "aggmtransfn")); |
| aggminvtransfn = PQgetvalue(res, 0, PQfnumber(res, "aggminvtransfn")); |
| aggmfinalfn = PQgetvalue(res, 0, PQfnumber(res, "aggmfinalfn")); |
| aggfinalextra = (PQgetvalue(res, 0, PQfnumber(res, "aggfinalextra"))[0] == 't'); |
| aggmfinalextra = (PQgetvalue(res, 0, PQfnumber(res, "aggmfinalextra"))[0] == 't'); |
| aggfinalmodify = PQgetvalue(res, 0, PQfnumber(res, "aggfinalmodify"))[0]; |
| aggmfinalmodify = PQgetvalue(res, 0, PQfnumber(res, "aggmfinalmodify"))[0]; |
| aggsortop = PQgetvalue(res, 0, PQfnumber(res, "aggsortop")); |
| aggkind = PQgetvalue(res, 0, PQfnumber(res, "aggkind"))[0]; |
| aggtranstype = PQgetvalue(res, 0, PQfnumber(res, "aggtranstype")); |
| aggtransspace = PQgetvalue(res, 0, PQfnumber(res, "aggtransspace")); |
| aggmtranstype = PQgetvalue(res, 0, PQfnumber(res, "aggmtranstype")); |
| aggmtransspace = PQgetvalue(res, 0, PQfnumber(res, "aggmtransspace")); |
| agginitval = PQgetvalue(res, 0, i_agginitval); |
| aggminitval = PQgetvalue(res, 0, i_aggminitval); |
| aggrepsafeexec = (PQgetvalue(res, 0, PQfnumber(res, "aggrepsafeexec"))[0] == 't'); |
| proparallel = PQgetvalue(res, 0, PQfnumber(res, "proparallel")); |
| |
| { |
| char *funcargs; |
| char *funciargs; |
| |
| funcargs = PQgetvalue(res, 0, PQfnumber(res, "funcargs")); |
| funciargs = PQgetvalue(res, 0, PQfnumber(res, "funciargs")); |
| aggfullsig = format_function_arguments(&agginfo->aggfn, funcargs, true); |
| aggsig = format_function_arguments(&agginfo->aggfn, funciargs, true); |
| } |
| |
| aggsig_tag = format_aggregate_signature(agginfo, fout, false); |
| |
| /* identify default modify flag for aggkind (must match DefineAggregate) */ |
| defaultfinalmodify = (aggkind == AGGKIND_NORMAL) ? AGGMODIFY_READ_ONLY : AGGMODIFY_READ_WRITE; |
| /* replace omitted flags for old versions */ |
| if (aggfinalmodify == '0') |
| aggfinalmodify = defaultfinalmodify; |
| if (aggmfinalmodify == '0') |
| aggmfinalmodify = defaultfinalmodify; |
| |
| /* regproc and regtype output is already sufficiently quoted */ |
| appendPQExpBuffer(details, " SFUNC = %s,\n STYPE = %s", |
| aggtransfn, aggtranstype); |
| |
| if (strcmp(aggtransspace, "0") != 0) |
| { |
| appendPQExpBuffer(details, ",\n SSPACE = %s", |
| aggtransspace); |
| } |
| |
| if (!PQgetisnull(res, 0, i_agginitval)) |
| { |
| appendPQExpBufferStr(details, ",\n INITCOND = "); |
| appendStringLiteralAH(details, agginitval, fout); |
| } |
| |
| if (strcmp(aggfinalfn, "-") != 0) |
| { |
| appendPQExpBuffer(details, ",\n FINALFUNC = %s", |
| aggfinalfn); |
| if (aggfinalextra) |
| appendPQExpBufferStr(details, ",\n FINALFUNC_EXTRA"); |
| if (aggfinalmodify != defaultfinalmodify) |
| { |
| switch (aggfinalmodify) |
| { |
| case AGGMODIFY_READ_ONLY: |
| appendPQExpBufferStr(details, ",\n FINALFUNC_MODIFY = READ_ONLY"); |
| break; |
| case AGGMODIFY_SHAREABLE: |
| appendPQExpBufferStr(details, ",\n FINALFUNC_MODIFY = SHAREABLE"); |
| break; |
| case AGGMODIFY_READ_WRITE: |
| appendPQExpBufferStr(details, ",\n FINALFUNC_MODIFY = READ_WRITE"); |
| break; |
| default: |
| pg_fatal("unrecognized aggfinalmodify value for aggregate \"%s\"", |
| agginfo->aggfn.dobj.name); |
| break; |
| } |
| } |
| } |
| |
| if (strcmp(aggcombinefn, "-") != 0) |
| appendPQExpBuffer(details, ",\n COMBINEFUNC = %s", aggcombinefn); |
| |
| if (strcmp(aggserialfn, "-") != 0) |
| appendPQExpBuffer(details, ",\n SERIALFUNC = %s", aggserialfn); |
| |
| if (strcmp(aggdeserialfn, "-") != 0) |
| appendPQExpBuffer(details, ",\n DESERIALFUNC = %s", aggdeserialfn); |
| |
| if (strcmp(aggmtransfn, "-") != 0) |
| { |
| appendPQExpBuffer(details, ",\n MSFUNC = %s,\n MINVFUNC = %s,\n MSTYPE = %s", |
| aggmtransfn, |
| aggminvtransfn, |
| aggmtranstype); |
| } |
| |
| if (strcmp(aggmtransspace, "0") != 0) |
| { |
| appendPQExpBuffer(details, ",\n MSSPACE = %s", |
| aggmtransspace); |
| } |
| |
| if (!PQgetisnull(res, 0, i_aggminitval)) |
| { |
| appendPQExpBufferStr(details, ",\n MINITCOND = "); |
| appendStringLiteralAH(details, aggminitval, fout); |
| } |
| |
| if (strcmp(aggmfinalfn, "-") != 0) |
| { |
| appendPQExpBuffer(details, ",\n MFINALFUNC = %s", |
| aggmfinalfn); |
| if (aggmfinalextra) |
| appendPQExpBufferStr(details, ",\n MFINALFUNC_EXTRA"); |
| if (aggmfinalmodify != defaultfinalmodify) |
| { |
| switch (aggmfinalmodify) |
| { |
| case AGGMODIFY_READ_ONLY: |
| appendPQExpBufferStr(details, ",\n MFINALFUNC_MODIFY = READ_ONLY"); |
| break; |
| case AGGMODIFY_SHAREABLE: |
| appendPQExpBufferStr(details, ",\n MFINALFUNC_MODIFY = SHAREABLE"); |
| break; |
| case AGGMODIFY_READ_WRITE: |
| appendPQExpBufferStr(details, ",\n MFINALFUNC_MODIFY = READ_WRITE"); |
| break; |
| default: |
| pg_fatal("unrecognized aggmfinalmodify value for aggregate \"%s\"", |
| agginfo->aggfn.dobj.name); |
| break; |
| } |
| } |
| } |
| |
| if (aggrepsafeexec) |
| appendPQExpBuffer(details, ",\n REPSAFE = true"); |
| |
| aggsortconvop = getFormattedOperatorName(aggsortop); |
| if (aggsortconvop) |
| { |
| appendPQExpBuffer(details, ",\n SORTOP = %s", |
| aggsortconvop); |
| free(aggsortconvop); |
| } |
| |
| if (aggkind == AGGKIND_HYPOTHETICAL) |
| appendPQExpBufferStr(details, ",\n HYPOTHETICAL"); |
| |
| if (proparallel[0] != PROPARALLEL_UNSAFE) |
| { |
| if (proparallel[0] == PROPARALLEL_SAFE) |
| appendPQExpBufferStr(details, ",\n PARALLEL = safe"); |
| else if (proparallel[0] == PROPARALLEL_RESTRICTED) |
| appendPQExpBufferStr(details, ",\n PARALLEL = restricted"); |
| else if (proparallel[0] != PROPARALLEL_UNSAFE) |
| pg_fatal("unrecognized proparallel value for function \"%s\"", |
| agginfo->aggfn.dobj.name); |
| } |
| |
| appendPQExpBuffer(delq, "DROP AGGREGATE %s.%s;\n", |
| fmtId(agginfo->aggfn.dobj.namespace->dobj.name), |
| aggsig); |
| |
| appendPQExpBuffer(q, "CREATE AGGREGATE %s.%s (\n%s\n);\n", |
| fmtId(agginfo->aggfn.dobj.namespace->dobj.name), |
| aggfullsig ? aggfullsig : aggsig, details->data); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &agginfo->aggfn.dobj, |
| "AGGREGATE", aggsig, |
| agginfo->aggfn.dobj.namespace->dobj.name); |
| |
| if (agginfo->aggfn.dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, agginfo->aggfn.dobj.catId, |
| agginfo->aggfn.dobj.dumpId, |
| ARCHIVE_OPTS(.tag = aggsig_tag, |
| .namespace = agginfo->aggfn.dobj.namespace->dobj.name, |
| .owner = agginfo->aggfn.rolname, |
| .description = "AGGREGATE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Aggregate Comments */ |
| if (agginfo->aggfn.dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "AGGREGATE", aggsig, |
| agginfo->aggfn.dobj.namespace->dobj.name, |
| agginfo->aggfn.rolname, |
| agginfo->aggfn.dobj.catId, 0, agginfo->aggfn.dobj.dumpId); |
| |
| if (agginfo->aggfn.dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "AGGREGATE", aggsig, |
| agginfo->aggfn.dobj.namespace->dobj.name, |
| agginfo->aggfn.rolname, |
| agginfo->aggfn.dobj.catId, 0, agginfo->aggfn.dobj.dumpId); |
| |
| /* |
| * Since there is no GRANT ON AGGREGATE syntax, we have to make the ACL |
| * command look like a function's GRANT; in particular this affects the |
| * syntax for zero-argument aggregates and ordered-set aggregates. |
| */ |
| free(aggsig); |
| |
| aggsig = format_function_signature(fout, &agginfo->aggfn, true); |
| |
| if (agginfo->aggfn.dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, agginfo->aggfn.dobj.dumpId, InvalidDumpId, |
| "FUNCTION", aggsig, NULL, |
| agginfo->aggfn.dobj.namespace->dobj.name, |
| agginfo->aggfn.rolname, &agginfo->aggfn.dacl); |
| |
| free(aggsig); |
| free(aggfullsig); |
| free(aggsig_tag); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(details); |
| } |
| |
| /* |
| * getFunctionName - retrieves a function name from an oid |
| * |
| */ |
| static char * |
| getFunctionName(Archive *fout, Oid oid) |
| { |
| char *result; |
| PQExpBuffer query; |
| PGresult *res; |
| int ntups; |
| |
| if (oid == InvalidOid) |
| { |
| return NULL; |
| } |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "SELECT proname FROM pg_proc WHERE oid = %u;",oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| /* Expecting a single result only */ |
| ntups = PQntuples(res); |
| if (ntups != 1) |
| { |
| pg_log_warning("query yielded %d rows instead of one: %s", |
| ntups, query->data); |
| exit_nicely(1); |
| } |
| |
| /* already quoted */ |
| result = pg_strdup(PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| return result; |
| } |
| |
| /* |
| * dumpExtProtocol |
| * write out a single external protocol definition |
| */ |
| static void |
| dumpExtProtocol(Archive *fout, const ExtProtInfo *ptcinfo) |
| { |
| #define FCOUNT 3 |
| #define READFN_IDX 0 |
| #define WRITEFN_IDX 1 |
| #define VALIDFN_IDX 2 |
| |
| typedef struct |
| { |
| Oid oid; /* func's oid */ |
| char* name; /* func name */ |
| FuncInfo* pfuncinfo; /* FuncInfo ptr */ |
| bool dumpable; /* should we dump this function */ |
| bool internal; /* is it an internal function */ |
| } ProtoFunc; |
| |
| ProtoFunc protoFuncs[FCOUNT]; |
| |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer nsq; |
| char *prev_ns; |
| char *namecopy; |
| int i; |
| bool has_internal = false; |
| |
| /* Skip if not to be dumped */ |
| if (!ptcinfo->dobj.dump || fout->dopt->dataOnly) |
| return; |
| |
| /* init and fill the protoFuncs array */ |
| memset(protoFuncs, 0, sizeof(protoFuncs)); |
| protoFuncs[READFN_IDX].oid = ptcinfo->ptcreadid; |
| protoFuncs[WRITEFN_IDX].oid = ptcinfo->ptcwriteid; |
| protoFuncs[VALIDFN_IDX].oid = ptcinfo->ptcvalidid; |
| |
| for (i = 0; i < FCOUNT; i++) |
| { |
| if (protoFuncs[i].oid == InvalidOid) |
| { |
| protoFuncs[i].dumpable = false; |
| protoFuncs[i].internal = true; |
| /* |
| * We have at least one internal function, signal that we need the |
| * public schema in the search_path |
| */ |
| has_internal = true; |
| } |
| else |
| { |
| protoFuncs[i].pfuncinfo = findFuncByOid(protoFuncs[i].oid); |
| if (protoFuncs[i].pfuncinfo != NULL) |
| { |
| protoFuncs[i].dumpable = true; |
| protoFuncs[i].name = pg_strdup(protoFuncs[i].pfuncinfo->dobj.name); |
| protoFuncs[i].internal = false; |
| } |
| else |
| protoFuncs[i].internal = true; |
| } |
| } |
| |
| /* if all funcs are internal then we do not need to dump this protocol */ |
| if (protoFuncs[READFN_IDX].internal && protoFuncs[WRITEFN_IDX].internal |
| && protoFuncs[VALIDFN_IDX].internal) |
| return; |
| |
| /* obtain the function name for internal functions (if any) */ |
| for (i = 0; i < FCOUNT; i++) |
| { |
| if (protoFuncs[i].internal && protoFuncs[i].oid) |
| { |
| protoFuncs[i].name = getFunctionName(fout, protoFuncs[i].oid); |
| if (protoFuncs[i].name) |
| protoFuncs[i].dumpable = true; |
| } |
| } |
| |
| nsq = createPQExpBuffer(); |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| /* |
| * Since the function parameters to the external protocol cannot be fully |
| * qualified with namespace, we must ensure that we have the search_path |
| * set with the namespaces of the referenced functions. We only need the |
| * dump file to have the search_path so inject a SET search_path = .. ; |
| * into the output stream instead of calling selectSourceSchema(). |
| * |
| * GPDB_96_MERGE_FIXME: update the above comment because selectSourceSchema |
| * has been removed in upstream 9f6e5296a Security: CVE-2018-1058 |
| */ |
| prev_ns = NULL; |
| for (i = 0; i < FCOUNT; i++) |
| { |
| if (!protoFuncs[i].pfuncinfo || protoFuncs[i].internal) |
| continue; |
| |
| if (prev_ns && strcmp(prev_ns, protoFuncs[i].pfuncinfo->dobj.namespace->dobj.name) == 0) |
| continue; |
| |
| appendPQExpBuffer(nsq, "%s%s", (prev_ns ? "," : ""), protoFuncs[i].pfuncinfo->dobj.namespace->dobj.name); |
| prev_ns = protoFuncs[i].pfuncinfo->dobj.namespace->dobj.name; |
| |
| /* |
| * If we are adding public to the search_path, then we don't need to do |
| * so again for any internal functions |
| */ |
| if (strcmp(prev_ns, "public") == 0) |
| has_internal = false; |
| } |
| |
| if (prev_ns) |
| { |
| appendPQExpBufferStr(q, "-- Set the search_path required to look up the functions\n"); |
| appendPQExpBuffer(q, "SET search_path = %s%s;\n\n", |
| nsq->data, (has_internal ? ", public" : "")); |
| } |
| destroyPQExpBuffer(nsq); |
| |
| appendPQExpBuffer(q, "CREATE %s PROTOCOL %s (", |
| ptcinfo->ptctrusted == true ? "TRUSTED" : "", |
| fmtId(ptcinfo->dobj.name)); |
| |
| if (protoFuncs[READFN_IDX].dumpable) |
| { |
| appendPQExpBuffer(q, " readfunc = '%s'%s", |
| protoFuncs[READFN_IDX].name, |
| (protoFuncs[WRITEFN_IDX].dumpable ? "," : "")); |
| } |
| |
| if (protoFuncs[WRITEFN_IDX].dumpable) |
| { |
| appendPQExpBuffer(q, " writefunc = '%s'%s", |
| protoFuncs[WRITEFN_IDX].name, |
| (protoFuncs[VALIDFN_IDX].dumpable ? "," : "")); |
| } |
| |
| if (protoFuncs[VALIDFN_IDX].dumpable) |
| { |
| appendPQExpBuffer(q, " validatorfunc = '%s'", |
| protoFuncs[VALIDFN_IDX].name); |
| } |
| appendPQExpBufferStr(q, ");\n"); |
| |
| appendPQExpBuffer(delq, "DROP PROTOCOL %s;\n", |
| fmtId(ptcinfo->dobj.name)); |
| |
| ArchiveEntry(fout, ptcinfo->dobj.catId, ptcinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = ptcinfo->dobj.name, |
| .owner = ptcinfo->ptcowner, |
| .description = "PROTOCOL", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data, |
| .deps = ptcinfo->dobj.dependencies, |
| .nDeps = ptcinfo->dobj.nDeps)); |
| |
| /* Handle the ACL */ |
| namecopy = pg_strdup(fmtId(ptcinfo->dobj.name)); |
| dumpACL(fout, ptcinfo->dobj.dumpId, InvalidDumpId, |
| "PROTOCOL", |
| namecopy, NULL, |
| NULL, ptcinfo->ptcowner, |
| ptcinfo->ptcacl, ptcinfo->rproacl, |
| ptcinfo->initproacl, ptcinfo->initrproacl); |
| free(namecopy); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| |
| for (i = 0; i < FCOUNT; i++) |
| { |
| if (protoFuncs[i].name) |
| free(protoFuncs[i].name); |
| } |
| } |
| |
| /* |
| * dumpTSParser |
| * write out a single text search parser |
| */ |
| static void |
| dumpTSParser(Archive *fout, const TSParserInfo *prsinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qprsname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qprsname = pg_strdup(fmtId(prsinfo->dobj.name)); |
| |
| appendPQExpBuffer(q, "CREATE TEXT SEARCH PARSER %s (\n", |
| fmtQualifiedDumpable(prsinfo)); |
| |
| appendPQExpBuffer(q, " START = %s,\n", |
| convertTSFunction(fout, prsinfo->prsstart)); |
| appendPQExpBuffer(q, " GETTOKEN = %s,\n", |
| convertTSFunction(fout, prsinfo->prstoken)); |
| appendPQExpBuffer(q, " END = %s,\n", |
| convertTSFunction(fout, prsinfo->prsend)); |
| if (prsinfo->prsheadline != InvalidOid) |
| appendPQExpBuffer(q, " HEADLINE = %s,\n", |
| convertTSFunction(fout, prsinfo->prsheadline)); |
| appendPQExpBuffer(q, " LEXTYPES = %s );\n", |
| convertTSFunction(fout, prsinfo->prslextype)); |
| |
| appendPQExpBuffer(delq, "DROP TEXT SEARCH PARSER %s;\n", |
| fmtQualifiedDumpable(prsinfo)); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &prsinfo->dobj, |
| "TEXT SEARCH PARSER", qprsname, |
| prsinfo->dobj.namespace->dobj.name); |
| |
| if (prsinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, prsinfo->dobj.catId, prsinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = prsinfo->dobj.name, |
| .namespace = prsinfo->dobj.namespace->dobj.name, |
| .description = "TEXT SEARCH PARSER", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Parser Comments */ |
| if (prsinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TEXT SEARCH PARSER", qprsname, |
| prsinfo->dobj.namespace->dobj.name, "", |
| prsinfo->dobj.catId, 0, prsinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qprsname); |
| } |
| |
| /* |
| * dumpTSDictionary |
| * write out a single text search dictionary |
| */ |
| static void |
| dumpTSDictionary(Archive *fout, const TSDictInfo *dictinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| char *qdictname; |
| PGresult *res; |
| char *nspname; |
| char *tmplname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| qdictname = pg_strdup(fmtId(dictinfo->dobj.name)); |
| |
| /* Fetch name and namespace of the dictionary's template */ |
| appendPQExpBuffer(query, "SELECT nspname, tmplname " |
| "FROM pg_ts_template p, pg_namespace n " |
| "WHERE p.oid = '%u' AND n.oid = tmplnamespace", |
| dictinfo->dicttemplate); |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| nspname = PQgetvalue(res, 0, 0); |
| tmplname = PQgetvalue(res, 0, 1); |
| |
| appendPQExpBuffer(q, "CREATE TEXT SEARCH DICTIONARY %s (\n", |
| fmtQualifiedDumpable(dictinfo)); |
| |
| appendPQExpBufferStr(q, " TEMPLATE = "); |
| appendPQExpBuffer(q, "%s.", fmtId(nspname)); |
| appendPQExpBufferStr(q, fmtId(tmplname)); |
| |
| PQclear(res); |
| |
| /* the dictinitoption can be dumped straight into the command */ |
| if (dictinfo->dictinitoption) |
| appendPQExpBuffer(q, ",\n %s", dictinfo->dictinitoption); |
| |
| appendPQExpBufferStr(q, " );\n"); |
| |
| appendPQExpBuffer(delq, "DROP TEXT SEARCH DICTIONARY %s;\n", |
| fmtQualifiedDumpable(dictinfo)); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &dictinfo->dobj, |
| "TEXT SEARCH DICTIONARY", qdictname, |
| dictinfo->dobj.namespace->dobj.name); |
| |
| if (dictinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, dictinfo->dobj.catId, dictinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = dictinfo->dobj.name, |
| .namespace = dictinfo->dobj.namespace->dobj.name, |
| .owner = dictinfo->rolname, |
| .description = "TEXT SEARCH DICTIONARY", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Dictionary Comments */ |
| if (dictinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TEXT SEARCH DICTIONARY", qdictname, |
| dictinfo->dobj.namespace->dobj.name, dictinfo->rolname, |
| dictinfo->dobj.catId, 0, dictinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qdictname); |
| } |
| |
| /* |
| * dumpTSTemplate |
| * write out a single text search template |
| */ |
| static void |
| dumpTSTemplate(Archive *fout, const TSTemplateInfo *tmplinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qtmplname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qtmplname = pg_strdup(fmtId(tmplinfo->dobj.name)); |
| |
| appendPQExpBuffer(q, "CREATE TEXT SEARCH TEMPLATE %s (\n", |
| fmtQualifiedDumpable(tmplinfo)); |
| |
| if (tmplinfo->tmplinit != InvalidOid) |
| appendPQExpBuffer(q, " INIT = %s,\n", |
| convertTSFunction(fout, tmplinfo->tmplinit)); |
| appendPQExpBuffer(q, " LEXIZE = %s );\n", |
| convertTSFunction(fout, tmplinfo->tmpllexize)); |
| |
| appendPQExpBuffer(delq, "DROP TEXT SEARCH TEMPLATE %s;\n", |
| fmtQualifiedDumpable(tmplinfo)); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tmplinfo->dobj, |
| "TEXT SEARCH TEMPLATE", qtmplname, |
| tmplinfo->dobj.namespace->dobj.name); |
| |
| if (tmplinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tmplinfo->dobj.catId, tmplinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tmplinfo->dobj.name, |
| .namespace = tmplinfo->dobj.namespace->dobj.name, |
| .description = "TEXT SEARCH TEMPLATE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Template Comments */ |
| if (tmplinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TEXT SEARCH TEMPLATE", qtmplname, |
| tmplinfo->dobj.namespace->dobj.name, "", |
| tmplinfo->dobj.catId, 0, tmplinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qtmplname); |
| } |
| |
| /* |
| * dumpTSConfig |
| * write out a single text search configuration |
| */ |
| static void |
| dumpTSConfig(Archive *fout, const TSConfigInfo *cfginfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| char *qcfgname; |
| PGresult *res; |
| char *nspname; |
| char *prsname; |
| int ntups, |
| i; |
| int i_tokenname; |
| int i_dictname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| qcfgname = pg_strdup(fmtId(cfginfo->dobj.name)); |
| |
| /* Fetch name and namespace of the config's parser */ |
| appendPQExpBuffer(query, "SELECT nspname, prsname " |
| "FROM pg_ts_parser p, pg_namespace n " |
| "WHERE p.oid = '%u' AND n.oid = prsnamespace", |
| cfginfo->cfgparser); |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| nspname = PQgetvalue(res, 0, 0); |
| prsname = PQgetvalue(res, 0, 1); |
| |
| appendPQExpBuffer(q, "CREATE TEXT SEARCH CONFIGURATION %s (\n", |
| fmtQualifiedDumpable(cfginfo)); |
| |
| appendPQExpBuffer(q, " PARSER = %s.", fmtId(nspname)); |
| appendPQExpBuffer(q, "%s );\n", fmtId(prsname)); |
| |
| PQclear(res); |
| |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, |
| "SELECT\n" |
| " ( SELECT alias FROM pg_catalog.ts_token_type('%u'::pg_catalog.oid) AS t\n" |
| " WHERE t.tokid = m.maptokentype ) AS tokenname,\n" |
| " m.mapdict::pg_catalog.regdictionary AS dictname\n" |
| "FROM pg_catalog.pg_ts_config_map AS m\n" |
| "WHERE m.mapcfg = '%u'\n" |
| "ORDER BY m.mapcfg, m.maptokentype, m.mapseqno", |
| cfginfo->cfgparser, cfginfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| ntups = PQntuples(res); |
| |
| i_tokenname = PQfnumber(res, "tokenname"); |
| i_dictname = PQfnumber(res, "dictname"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| char *tokenname = PQgetvalue(res, i, i_tokenname); |
| char *dictname = PQgetvalue(res, i, i_dictname); |
| |
| if (i == 0 || |
| strcmp(tokenname, PQgetvalue(res, i - 1, i_tokenname)) != 0) |
| { |
| /* starting a new token type, so start a new command */ |
| if (i > 0) |
| appendPQExpBufferStr(q, ";\n"); |
| appendPQExpBuffer(q, "\nALTER TEXT SEARCH CONFIGURATION %s\n", |
| fmtQualifiedDumpable(cfginfo)); |
| /* tokenname needs quoting, dictname does NOT */ |
| appendPQExpBuffer(q, " ADD MAPPING FOR %s WITH %s", |
| fmtId(tokenname), dictname); |
| } |
| else |
| appendPQExpBuffer(q, ", %s", dictname); |
| } |
| |
| if (ntups > 0) |
| appendPQExpBufferStr(q, ";\n"); |
| |
| PQclear(res); |
| |
| appendPQExpBuffer(delq, "DROP TEXT SEARCH CONFIGURATION %s;\n", |
| fmtQualifiedDumpable(cfginfo)); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &cfginfo->dobj, |
| "TEXT SEARCH CONFIGURATION", qcfgname, |
| cfginfo->dobj.namespace->dobj.name); |
| |
| if (cfginfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, cfginfo->dobj.catId, cfginfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = cfginfo->dobj.name, |
| .namespace = cfginfo->dobj.namespace->dobj.name, |
| .owner = cfginfo->rolname, |
| .description = "TEXT SEARCH CONFIGURATION", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Configuration Comments */ |
| if (cfginfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "TEXT SEARCH CONFIGURATION", qcfgname, |
| cfginfo->dobj.namespace->dobj.name, cfginfo->rolname, |
| cfginfo->dobj.catId, 0, cfginfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qcfgname); |
| } |
| |
| /* |
| * dumpForeignDataWrapper |
| * write out a single foreign-data wrapper definition |
| */ |
| static void |
| dumpForeignDataWrapper(Archive *fout, const FdwInfo *fdwinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qfdwname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qfdwname = pg_strdup(fmtId(fdwinfo->dobj.name)); |
| |
| appendPQExpBuffer(q, "CREATE FOREIGN DATA WRAPPER %s", |
| qfdwname); |
| |
| if (strcmp(fdwinfo->fdwhandler, "-") != 0) |
| appendPQExpBuffer(q, " HANDLER %s", fdwinfo->fdwhandler); |
| |
| if (strcmp(fdwinfo->fdwvalidator, "-") != 0) |
| appendPQExpBuffer(q, " VALIDATOR %s", fdwinfo->fdwvalidator); |
| |
| if (strlen(fdwinfo->fdwoptions) > 0) |
| appendPQExpBuffer(q, " OPTIONS (\n %s\n)", fdwinfo->fdwoptions); |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| appendPQExpBuffer(delq, "DROP FOREIGN DATA WRAPPER %s;\n", |
| qfdwname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &fdwinfo->dobj, |
| "FOREIGN DATA WRAPPER", qfdwname, |
| NULL); |
| |
| if (fdwinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, fdwinfo->dobj.catId, fdwinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = fdwinfo->dobj.name, |
| .owner = fdwinfo->rolname, |
| .description = "FOREIGN DATA WRAPPER", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Foreign Data Wrapper Comments */ |
| if (fdwinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "FOREIGN DATA WRAPPER", qfdwname, |
| NULL, fdwinfo->rolname, |
| fdwinfo->dobj.catId, 0, fdwinfo->dobj.dumpId); |
| |
| /* Handle the ACL */ |
| if (fdwinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, fdwinfo->dobj.dumpId, InvalidDumpId, |
| "FOREIGN DATA WRAPPER", qfdwname, NULL, |
| NULL, fdwinfo->rolname, &fdwinfo->dacl); |
| |
| free(qfdwname); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| } |
| |
| /* |
| * dumpForeignServer |
| * write out a foreign server definition |
| */ |
| static void |
| dumpForeignServer(Archive *fout, const ForeignServerInfo *srvinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| PGresult *res; |
| char *qsrvname; |
| char *fdwname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| qsrvname = pg_strdup(fmtId(srvinfo->dobj.name)); |
| |
| /* look up the foreign-data wrapper */ |
| appendPQExpBuffer(query, "SELECT fdwname " |
| "FROM pg_foreign_data_wrapper w " |
| "WHERE w.oid = '%u'", |
| srvinfo->srvfdw); |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| fdwname = PQgetvalue(res, 0, 0); |
| |
| appendPQExpBuffer(q, "CREATE SERVER %s", qsrvname); |
| if (srvinfo->srvtype && strlen(srvinfo->srvtype) > 0) |
| { |
| appendPQExpBufferStr(q, " TYPE "); |
| appendStringLiteralAH(q, srvinfo->srvtype, fout); |
| } |
| if (srvinfo->srvversion && strlen(srvinfo->srvversion) > 0) |
| { |
| appendPQExpBufferStr(q, " VERSION "); |
| appendStringLiteralAH(q, srvinfo->srvversion, fout); |
| } |
| |
| appendPQExpBufferStr(q, " FOREIGN DATA WRAPPER "); |
| appendPQExpBufferStr(q, fmtId(fdwname)); |
| |
| if (srvinfo->srvoptions && strlen(srvinfo->srvoptions) > 0) |
| appendPQExpBuffer(q, " OPTIONS (\n %s\n)", srvinfo->srvoptions); |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| appendPQExpBuffer(delq, "DROP SERVER %s;\n", |
| qsrvname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &srvinfo->dobj, |
| "SERVER", qsrvname, NULL); |
| |
| if (srvinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, srvinfo->dobj.catId, srvinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = srvinfo->dobj.name, |
| .owner = srvinfo->rolname, |
| .description = "SERVER", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Foreign Server Comments */ |
| if (srvinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "SERVER", qsrvname, |
| NULL, srvinfo->rolname, |
| srvinfo->dobj.catId, 0, srvinfo->dobj.dumpId); |
| |
| /* Handle the ACL */ |
| if (srvinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| dumpACL(fout, srvinfo->dobj.dumpId, InvalidDumpId, |
| "FOREIGN SERVER", qsrvname, NULL, |
| NULL, srvinfo->rolname, &srvinfo->dacl); |
| |
| /* Dump user mappings */ |
| if (srvinfo->dobj.dump & DUMP_COMPONENT_USERMAP) |
| dumpUserMappings(fout, |
| srvinfo->dobj.name, NULL, |
| srvinfo->rolname, |
| srvinfo->dobj.catId, srvinfo->dobj.dumpId); |
| |
| PQclear(res); |
| |
| free(qsrvname); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpUserMappings |
| * |
| * This routine is used to dump any user mappings associated with the |
| * server handed to this routine. Should be called after ArchiveEntry() |
| * for the server. |
| */ |
| static void |
| dumpUserMappings(Archive *fout, |
| const char *servername, const char *namespace, |
| const char *owner, |
| CatalogId catalogId, DumpId dumpId) |
| { |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| PQExpBuffer tag; |
| PGresult *res; |
| int ntups; |
| int i_usename; |
| int i_umoptions; |
| int i; |
| |
| q = createPQExpBuffer(); |
| tag = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| /* |
| * We read from the publicly accessible view pg_user_mappings, so as not |
| * to fail if run by a non-superuser. Note that the view will show |
| * umoptions as null if the user hasn't got privileges for the associated |
| * server; this means that pg_dump will dump such a mapping, but with no |
| * OPTIONS clause. A possible alternative is to skip such mappings |
| * altogether, but it's not clear that that's an improvement. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT usename, " |
| "array_to_string(ARRAY(" |
| "SELECT quote_ident(option_name) || ' ' || " |
| "quote_literal(option_value) " |
| "FROM pg_options_to_table(umoptions) " |
| "ORDER BY option_name" |
| "), E',\n ') AS umoptions " |
| "FROM pg_user_mappings " |
| "WHERE srvid = '%u' " |
| "ORDER BY usename", |
| catalogId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| i_usename = PQfnumber(res, "usename"); |
| i_umoptions = PQfnumber(res, "umoptions"); |
| |
| for (i = 0; i < ntups; i++) |
| { |
| char *usename; |
| char *umoptions; |
| |
| usename = PQgetvalue(res, i, i_usename); |
| umoptions = PQgetvalue(res, i, i_umoptions); |
| |
| resetPQExpBuffer(q); |
| appendPQExpBuffer(q, "CREATE USER MAPPING FOR %s", fmtId(usename)); |
| appendPQExpBuffer(q, " SERVER %s", fmtId(servername)); |
| |
| if (umoptions && strlen(umoptions) > 0) |
| appendPQExpBuffer(q, " OPTIONS (\n %s\n)", umoptions); |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| resetPQExpBuffer(delq); |
| appendPQExpBuffer(delq, "DROP USER MAPPING FOR %s", fmtId(usename)); |
| appendPQExpBuffer(delq, " SERVER %s;\n", fmtId(servername)); |
| |
| resetPQExpBuffer(tag); |
| appendPQExpBuffer(tag, "USER MAPPING %s SERVER %s", |
| usename, servername); |
| |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = namespace, |
| .owner = owner, |
| .description = "USER MAPPING", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(tag); |
| destroyPQExpBuffer(q); |
| } |
| |
| /* |
| * Write out default privileges information |
| */ |
| static void |
| dumpDefaultACL(Archive *fout, const DefaultACLInfo *daclinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer tag; |
| const char *type; |
| |
| /* Do nothing in data-only dump, or if we're skipping ACLs */ |
| if (dopt->dataOnly || dopt->aclsSkip) |
| return; |
| |
| q = createPQExpBuffer(); |
| tag = createPQExpBuffer(); |
| |
| switch (daclinfo->defaclobjtype) |
| { |
| case DEFACLOBJ_RELATION: |
| type = "TABLES"; |
| break; |
| case DEFACLOBJ_SEQUENCE: |
| type = "SEQUENCES"; |
| break; |
| case DEFACLOBJ_FUNCTION: |
| type = "FUNCTIONS"; |
| break; |
| case DEFACLOBJ_TYPE: |
| type = "TYPES"; |
| break; |
| case DEFACLOBJ_NAMESPACE: |
| type = "SCHEMAS"; |
| break; |
| default: |
| /* shouldn't get here */ |
| pg_fatal("unrecognized object type in default privileges: %d", |
| (int) daclinfo->defaclobjtype); |
| type = ""; /* keep compiler quiet */ |
| } |
| |
| appendPQExpBuffer(tag, "DEFAULT PRIVILEGES FOR %s", type); |
| |
| /* build the actual command(s) for this tuple */ |
| if (!buildDefaultACLCommands(type, |
| daclinfo->dobj.namespace != NULL ? |
| daclinfo->dobj.namespace->dobj.name : NULL, |
| daclinfo->dacl.acl, |
| daclinfo->dacl.acldefault, |
| daclinfo->defaclrole, |
| fout->remoteVersion, |
| q)) |
| pg_fatal("could not parse default ACL list (%s)", |
| daclinfo->dacl.acl); |
| |
| if (daclinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| ArchiveEntry(fout, daclinfo->dobj.catId, daclinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = daclinfo->dobj.namespace ? |
| daclinfo->dobj.namespace->dobj.name : NULL, |
| .owner = daclinfo->defaclrole, |
| .description = "DEFAULT ACL", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data)); |
| |
| destroyPQExpBuffer(tag); |
| destroyPQExpBuffer(q); |
| } |
| |
| /*---------- |
| * Write out grant/revoke information |
| * |
| * 'objDumpId' is the dump ID of the underlying object. |
| * 'altDumpId' can be a second dumpId that the ACL entry must also depend on, |
| * or InvalidDumpId if there is no need for a second dependency. |
| * 'type' must be one of |
| * TABLE, SEQUENCE, FUNCTION, LANGUAGE, SCHEMA, DATABASE, TABLESPACE, |
| * FOREIGN DATA WRAPPER, SERVER, or LARGE OBJECT. |
| * 'name' is the formatted name of the object. Must be quoted etc. already. |
| * 'subname' is the formatted name of the sub-object, if any. Must be quoted. |
| * (Currently we assume that subname is only provided for table columns.) |
| * 'nspname' is the namespace the object is in (NULL if none). |
| * 'owner' is the owner, NULL if there is no owner (for languages). |
| * 'dacl' is the DumpableAcl struct fpr the object. |
| * |
| * Returns the dump ID assigned to the ACL TocEntry, or InvalidDumpId if |
| * no ACL entry was created. |
| *---------- |
| */ |
| static DumpId |
| dumpACL(Archive *fout, DumpId objDumpId, DumpId altDumpId, |
| const char *type, const char *name, const char *subname, |
| const char *nspname, const char *owner, |
| const DumpableAcl *dacl) |
| { |
| DumpId aclDumpId = InvalidDumpId; |
| DumpOptions *dopt = fout->dopt; |
| const char *acls = dacl->acl; |
| const char *acldefault = dacl->acldefault; |
| char privtype = dacl->privtype; |
| const char *initprivs = dacl->initprivs; |
| const char *baseacls; |
| PQExpBuffer sql; |
| |
| /* Do nothing if ACL dump is not enabled */ |
| if (dopt->aclsSkip) |
| return InvalidDumpId; |
| |
| /* --data-only skips ACLs *except* large object ACLs */ |
| if (dopt->dataOnly && strcmp(type, "LARGE OBJECT") != 0) |
| return InvalidDumpId; |
| |
| sql = createPQExpBuffer(); |
| |
| /* |
| * In binary upgrade mode, we don't run an extension's script but instead |
| * dump out the objects independently and then recreate them. To preserve |
| * any initial privileges which were set on extension objects, we need to |
| * compute the set of GRANT and REVOKE commands necessary to get from the |
| * default privileges of an object to its initial privileges as recorded |
| * in pg_init_privs. |
| * |
| * At restore time, we apply these commands after having called |
| * binary_upgrade_set_record_init_privs(true). That tells the backend to |
| * copy the results into pg_init_privs. This is how we preserve the |
| * contents of that catalog across binary upgrades. |
| */ |
| if (dopt->binary_upgrade && privtype == 'e' && |
| initprivs && *initprivs != '\0') |
| { |
| appendPQExpBufferStr(sql, "SELECT pg_catalog.binary_upgrade_set_record_init_privs(true);\n"); |
| if (!buildACLCommands(name, subname, nspname, type, |
| initprivs, acldefault, owner, |
| "", fout->remoteVersion, sql)) |
| pg_fatal("could not parse initial ACL list (%s) or default (%s) for object \"%s\" (%s)", |
| initprivs, acldefault, name, type); |
| appendPQExpBufferStr(sql, "SELECT pg_catalog.binary_upgrade_set_record_init_privs(false);\n"); |
| } |
| |
| /* |
| * Now figure the GRANT and REVOKE commands needed to get to the object's |
| * actual current ACL, starting from the initprivs if given, else from the |
| * object-type-specific default. Also, while buildACLCommands will assume |
| * that a NULL/empty acls string means it needn't do anything, what that |
| * actually represents is the object-type-specific default; so we need to |
| * substitute the acldefault string to get the right results in that case. |
| */ |
| if (initprivs && *initprivs != '\0') |
| { |
| baseacls = initprivs; |
| if (acls == NULL || *acls == '\0') |
| acls = acldefault; |
| } |
| else |
| baseacls = acldefault; |
| |
| if (!buildACLCommands(name, subname, nspname, type, |
| acls, baseacls, owner, |
| "", fout->remoteVersion, sql)) |
| pg_fatal("could not parse ACL list (%s) or default (%s) for object \"%s\" (%s)", |
| acls, baseacls, name, type); |
| |
| if (sql->len > 0) |
| { |
| PQExpBuffer tag = createPQExpBuffer(); |
| DumpId aclDeps[2]; |
| int nDeps = 0; |
| |
| if (subname) |
| appendPQExpBuffer(tag, "COLUMN %s.%s", name, subname); |
| else |
| appendPQExpBuffer(tag, "%s %s", type, name); |
| |
| aclDeps[nDeps++] = objDumpId; |
| if (altDumpId != InvalidDumpId) |
| aclDeps[nDeps++] = altDumpId; |
| |
| aclDumpId = createDumpId(); |
| |
| ArchiveEntry(fout, nilCatalogId, aclDumpId, |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = nspname, |
| .owner = owner, |
| .description = "ACL", |
| .section = SECTION_NONE, |
| .createStmt = sql->data, |
| .deps = aclDeps, |
| .nDeps = nDeps)); |
| |
| destroyPQExpBuffer(tag); |
| } |
| |
| destroyPQExpBuffer(sql); |
| |
| return aclDumpId; |
| } |
| |
| /* |
| * dumpSecLabel |
| * |
| * This routine is used to dump any security labels associated with the |
| * object handed to this routine. The routine takes the object type |
| * and object name (ready to print, except for schema decoration), plus |
| * the namespace and owner of the object (for labeling the ArchiveEntry), |
| * plus catalog ID and subid which are the lookup key for pg_seclabel, |
| * plus the dump ID for the object (for setting a dependency). |
| * If a matching pg_seclabel entry is found, it is dumped. |
| * |
| * Note: although this routine takes a dumpId for dependency purposes, |
| * that purpose is just to mark the dependency in the emitted dump file |
| * for possible future use by pg_restore. We do NOT use it for determining |
| * ordering of the label in the dump file, because this routine is called |
| * after dependency sorting occurs. This routine should be called just after |
| * calling ArchiveEntry() for the specified object. |
| */ |
| static void |
| dumpSecLabel(Archive *fout, const char *type, const char *name, |
| const char *namespace, const char *owner, |
| CatalogId catalogId, int subid, DumpId dumpId) |
| { |
| DumpOptions *dopt = fout->dopt; |
| SecLabelItem *labels; |
| int nlabels; |
| int i; |
| PQExpBuffer query; |
| |
| /* do nothing, if --no-security-labels is supplied */ |
| if (dopt->no_security_labels) |
| return; |
| |
| /* |
| * Security labels are schema not data ... except large object labels are |
| * data |
| */ |
| if (strcmp(type, "LARGE OBJECT") != 0) |
| { |
| if (dopt->dataOnly) |
| return; |
| } |
| else |
| { |
| /* We do dump large object security labels in binary-upgrade mode */ |
| if (dopt->schemaOnly && !dopt->binary_upgrade) |
| return; |
| } |
| |
| /* Search for security labels associated with catalogId, using table */ |
| nlabels = findSecLabels(catalogId.tableoid, catalogId.oid, &labels); |
| |
| query = createPQExpBuffer(); |
| |
| for (i = 0; i < nlabels; i++) |
| { |
| /* |
| * Ignore label entries for which the subid doesn't match. |
| */ |
| if (labels[i].objsubid != subid) |
| continue; |
| |
| appendPQExpBuffer(query, |
| "SECURITY LABEL FOR %s ON %s ", |
| fmtId(labels[i].provider), type); |
| if (namespace && *namespace) |
| appendPQExpBuffer(query, "%s.", fmtId(namespace)); |
| appendPQExpBuffer(query, "%s IS ", name); |
| appendStringLiteralAH(query, labels[i].label, fout); |
| appendPQExpBufferStr(query, ";\n"); |
| } |
| |
| if (query->len > 0) |
| { |
| PQExpBuffer tag = createPQExpBuffer(); |
| |
| appendPQExpBuffer(tag, "%s %s", type, name); |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tag->data, |
| .namespace = namespace, |
| .owner = owner, |
| .description = "SECURITY LABEL", |
| .section = SECTION_NONE, |
| .createStmt = query->data, |
| .deps = &dumpId, |
| .nDeps = 1)); |
| destroyPQExpBuffer(tag); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpTableSecLabel |
| * |
| * As above, but dump security label for both the specified table (or view) |
| * and its columns. |
| */ |
| static void |
| dumpTableSecLabel(Archive *fout, const TableInfo *tbinfo, const char *reltypename) |
| { |
| DumpOptions *dopt = fout->dopt; |
| SecLabelItem *labels; |
| int nlabels; |
| int i; |
| PQExpBuffer query; |
| PQExpBuffer target; |
| |
| /* do nothing, if --no-security-labels is supplied */ |
| if (dopt->no_security_labels) |
| return; |
| |
| /* SecLabel are SCHEMA not data */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* Search for comments associated with relation, using table */ |
| nlabels = findSecLabels(tbinfo->dobj.catId.tableoid, |
| tbinfo->dobj.catId.oid, |
| &labels); |
| |
| /* If security labels exist, build SECURITY LABEL statements */ |
| if (nlabels <= 0) |
| return; |
| |
| query = createPQExpBuffer(); |
| target = createPQExpBuffer(); |
| |
| for (i = 0; i < nlabels; i++) |
| { |
| const char *colname; |
| const char *provider = labels[i].provider; |
| const char *label = labels[i].label; |
| int objsubid = labels[i].objsubid; |
| |
| resetPQExpBuffer(target); |
| if (objsubid == 0) |
| { |
| appendPQExpBuffer(target, "%s %s", reltypename, |
| fmtQualifiedDumpable(tbinfo)); |
| } |
| else |
| { |
| colname = getAttrName(objsubid, tbinfo); |
| /* first fmtXXX result must be consumed before calling again */ |
| appendPQExpBuffer(target, "COLUMN %s", |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(target, ".%s", fmtId(colname)); |
| } |
| appendPQExpBuffer(query, "SECURITY LABEL FOR %s ON %s IS ", |
| fmtId(provider), target->data); |
| appendStringLiteralAH(query, label, fout); |
| appendPQExpBufferStr(query, ";\n"); |
| } |
| if (query->len > 0) |
| { |
| resetPQExpBuffer(target); |
| appendPQExpBuffer(target, "%s %s", reltypename, |
| fmtId(tbinfo->dobj.name)); |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = target->data, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "SECURITY LABEL", |
| .section = SECTION_NONE, |
| .createStmt = query->data, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1)); |
| } |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(target); |
| } |
| |
| /* |
| * findSecLabels |
| * |
| * Find the security label(s), if any, associated with the given object. |
| * All the objsubid values associated with the given classoid/objoid are |
| * found with one search. |
| */ |
| static int |
| findSecLabels(Oid classoid, Oid objoid, SecLabelItem **items) |
| { |
| SecLabelItem *middle = NULL; |
| SecLabelItem *low; |
| SecLabelItem *high; |
| int nmatch; |
| |
| if (nseclabels <= 0) /* no labels, so no match is possible */ |
| { |
| *items = NULL; |
| return 0; |
| } |
| |
| /* |
| * Do binary search to find some item matching the object. |
| */ |
| low = &seclabels[0]; |
| high = &seclabels[nseclabels - 1]; |
| while (low <= high) |
| { |
| middle = low + (high - low) / 2; |
| |
| if (classoid < middle->classoid) |
| high = middle - 1; |
| else if (classoid > middle->classoid) |
| low = middle + 1; |
| else if (objoid < middle->objoid) |
| high = middle - 1; |
| else if (objoid > middle->objoid) |
| low = middle + 1; |
| else |
| break; /* found a match */ |
| } |
| |
| if (low > high) /* no matches */ |
| { |
| *items = NULL; |
| return 0; |
| } |
| |
| /* |
| * Now determine how many items match the object. The search loop |
| * invariant still holds: only items between low and high inclusive could |
| * match. |
| */ |
| nmatch = 1; |
| while (middle > low) |
| { |
| if (classoid != middle[-1].classoid || |
| objoid != middle[-1].objoid) |
| break; |
| middle--; |
| nmatch++; |
| } |
| |
| *items = middle; |
| |
| middle += nmatch; |
| while (middle <= high) |
| { |
| if (classoid != middle->classoid || |
| objoid != middle->objoid) |
| break; |
| middle++; |
| nmatch++; |
| } |
| |
| return nmatch; |
| } |
| |
| /* |
| * collectSecLabels |
| * |
| * Construct a table of all security labels available for database objects; |
| * also set the has-seclabel component flag for each relevant object. |
| * |
| * The table is sorted by classoid/objid/objsubid for speed in lookup. |
| */ |
| static void |
| collectSecLabels(Archive *fout) |
| { |
| PGresult *res; |
| PQExpBuffer query; |
| int i_label; |
| int i_provider; |
| int i_classoid; |
| int i_objoid; |
| int i_objsubid; |
| int ntups; |
| int i; |
| DumpableObject *dobj; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBufferStr(query, |
| "SELECT label, provider, classoid, objoid, objsubid " |
| "FROM pg_catalog.pg_seclabel " |
| "ORDER BY classoid, objoid, objsubid"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| /* Construct lookup table containing OIDs in numeric form */ |
| i_label = PQfnumber(res, "label"); |
| i_provider = PQfnumber(res, "provider"); |
| i_classoid = PQfnumber(res, "classoid"); |
| i_objoid = PQfnumber(res, "objoid"); |
| i_objsubid = PQfnumber(res, "objsubid"); |
| |
| ntups = PQntuples(res); |
| |
| seclabels = (SecLabelItem *) pg_malloc(ntups * sizeof(SecLabelItem)); |
| nseclabels = 0; |
| dobj = NULL; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| CatalogId objId; |
| int subid; |
| |
| objId.tableoid = atooid(PQgetvalue(res, i, i_classoid)); |
| objId.oid = atooid(PQgetvalue(res, i, i_objoid)); |
| subid = atoi(PQgetvalue(res, i, i_objsubid)); |
| |
| /* We needn't remember labels that don't match any dumpable object */ |
| if (dobj == NULL || |
| dobj->catId.tableoid != objId.tableoid || |
| dobj->catId.oid != objId.oid) |
| dobj = findObjectByCatalogId(objId); |
| if (dobj == NULL) |
| continue; |
| |
| /* |
| * Labels on columns of composite types are linked to the type's |
| * pg_class entry, but we need to set the DUMP_COMPONENT_SECLABEL flag |
| * in the type's own DumpableObject. |
| */ |
| if (subid != 0 && dobj->objType == DO_TABLE && |
| ((TableInfo *) dobj)->relkind == RELKIND_COMPOSITE_TYPE) |
| { |
| TypeInfo *cTypeInfo; |
| |
| cTypeInfo = findTypeByOid(((TableInfo *) dobj)->reltype); |
| if (cTypeInfo) |
| cTypeInfo->dobj.components |= DUMP_COMPONENT_SECLABEL; |
| } |
| else |
| dobj->components |= DUMP_COMPONENT_SECLABEL; |
| |
| seclabels[nseclabels].label = pg_strdup(PQgetvalue(res, i, i_label)); |
| seclabels[nseclabels].provider = pg_strdup(PQgetvalue(res, i, i_provider)); |
| seclabels[nseclabels].classoid = objId.tableoid; |
| seclabels[nseclabels].objoid = objId.oid; |
| seclabels[nseclabels].objsubid = subid; |
| nseclabels++; |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpTable |
| * write out to fout the declarations (not data) of a user-defined table |
| */ |
| static void |
| dumpTable(Archive *fout, const TableInfo *tbinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| DumpId tableAclDumpId = InvalidDumpId; |
| char *namecopy; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| { |
| if (tbinfo->relkind == RELKIND_SEQUENCE) |
| dumpSequence(fout, tbinfo); |
| else |
| dumpTableSchema(fout, tbinfo); |
| } |
| |
| /* Handle the ACL here */ |
| namecopy = pg_strdup(fmtId(tbinfo->dobj.name)); |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_ACL) |
| { |
| const char *objtype = |
| (tbinfo->relkind == RELKIND_SEQUENCE) ? "SEQUENCE" : "TABLE"; |
| |
| tableAclDumpId = |
| dumpACL(fout, tbinfo->dobj.dumpId, InvalidDumpId, |
| objtype, namecopy, NULL, |
| tbinfo->dobj.namespace->dobj.name, tbinfo->rolname, |
| &tbinfo->dacl); |
| } |
| |
| /* |
| * Handle column ACLs, if any. Note: we pull these with a separate query |
| * rather than trying to fetch them during getTableAttrs, so that we won't |
| * miss ACLs on system columns. Doing it this way also allows us to dump |
| * ACLs for catalogs that we didn't mark "interesting" back in getTables. |
| */ |
| if ((tbinfo->dobj.dump & DUMP_COMPONENT_ACL) && tbinfo->hascolumnACLs) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| int i; |
| |
| if (!fout->is_prepared[PREPQUERY_GETCOLUMNACLS]) |
| { |
| /* Set up query for column ACLs */ |
| appendPQExpBufferStr(query, |
| "PREPARE getColumnACLs(pg_catalog.oid) AS\n"); |
| |
| if (fout->remoteVersion >= 90600) |
| { |
| /* |
| * In principle we should call acldefault('c', relowner) to |
| * get the default ACL for a column. However, we don't |
| * currently store the numeric OID of the relowner in |
| * TableInfo. We could convert the owner name using regrole, |
| * but that creates a risk of failure due to concurrent role |
| * renames. Given that the default ACL for columns is empty |
| * and is likely to stay that way, it's not worth extra cycles |
| * and risk to avoid hard-wiring that knowledge here. |
| */ |
| appendPQExpBufferStr(query, |
| "SELECT at.attname, " |
| "at.attacl, " |
| "'{}' AS acldefault, " |
| "pip.privtype, pip.initprivs " |
| "FROM pg_catalog.pg_attribute at " |
| "LEFT JOIN pg_catalog.pg_init_privs pip ON " |
| "(at.attrelid = pip.objoid " |
| "AND pip.classoid = 'pg_catalog.pg_class'::pg_catalog.regclass " |
| "AND at.attnum = pip.objsubid) " |
| "WHERE at.attrelid = $1 AND " |
| "NOT at.attisdropped " |
| "AND (at.attacl IS NOT NULL OR pip.initprivs IS NOT NULL) " |
| "ORDER BY at.attnum"); |
| } |
| else |
| { |
| appendPQExpBufferStr(query, |
| "SELECT attname, attacl, '{}' AS acldefault, " |
| "NULL AS privtype, NULL AS initprivs " |
| "FROM pg_catalog.pg_attribute " |
| "WHERE attrelid = $1 AND NOT attisdropped " |
| "AND attacl IS NOT NULL " |
| "ORDER BY attnum"); |
| } |
| |
| ExecuteSqlStatement(fout, query->data); |
| |
| fout->is_prepared[PREPQUERY_GETCOLUMNACLS] = true; |
| } |
| |
| printfPQExpBuffer(query, |
| "EXECUTE getColumnACLs('%u')", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| for (i = 0; i < PQntuples(res); i++) |
| { |
| char *attname = PQgetvalue(res, i, 0); |
| char *attacl = PQgetvalue(res, i, 1); |
| char *acldefault = PQgetvalue(res, i, 2); |
| char privtype = *(PQgetvalue(res, i, 3)); |
| char *initprivs = PQgetvalue(res, i, 4); |
| DumpableAcl coldacl; |
| char *attnamecopy; |
| |
| coldacl.acl = attacl; |
| coldacl.acldefault = acldefault; |
| coldacl.privtype = privtype; |
| coldacl.initprivs = initprivs; |
| attnamecopy = pg_strdup(fmtId(attname)); |
| |
| /* |
| * Column's GRANT type is always TABLE. Each column ACL depends |
| * on the table-level ACL, since we can restore column ACLs in |
| * parallel but the table-level ACL has to be done first. |
| */ |
| dumpACL(fout, tbinfo->dobj.dumpId, tableAclDumpId, |
| "TABLE", namecopy, attnamecopy, |
| tbinfo->dobj.namespace->dobj.name, tbinfo->rolname, |
| &coldacl); |
| free(attnamecopy); |
| } |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| free(namecopy); |
| } |
| |
| static void |
| dumpExternal(Archive *fout, const TableInfo *tbinfo, PQExpBuffer q, PQExpBuffer delq) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PGresult *res; |
| char *urilocations; |
| char *execlocations; |
| char *location; |
| char *fmttype; |
| char *fmtopts; |
| char *command = NULL; |
| char *rejlim; |
| char *rejlimtype; |
| char *extencoding; |
| char *writable = NULL; |
| char *tmpstring = NULL; |
| char *tabfmt = NULL; |
| char *customfmt = NULL; |
| bool isweb = false; |
| bool iswritable = false; |
| char *options; |
| bool gpdb6OrLater = isGPDB6000OrLater(fout); |
| char *logerrors = NULL; |
| char *on_clause; |
| char *qualrelname; |
| PQExpBuffer query = createPQExpBuffer(); |
| |
| qualrelname = pg_strdup(fmtQualifiedDumpable(tbinfo)); |
| |
| /* |
| * DROP must be fully qualified in case same name appears in |
| * pg_catalog |
| */ |
| appendPQExpBuffer(delq, "DROP EXTERNAL TABLE %s.", |
| qualrelname); |
| |
| /* Now get required information from pg_exttable */ |
| if (gpdb6OrLater) |
| { |
| appendPQExpBuffer(query, |
| "SELECT x.urilocation, x.execlocation, x.fmttype, x.fmtopts, x.command, " |
| "x.rejectlimit, x.rejectlimittype, " |
| "CASE WHEN x.logerrors='f' THEN null ELSE x.logerrors::text END AS logerrors, " |
| "pg_catalog.pg_encoding_to_char(x.encoding), " |
| "x.writable, " |
| "array_to_string(ARRAY( " |
| "SELECT pg_catalog.quote_ident(option_name) || ' ' || " |
| "pg_catalog.quote_literal(option_value) " |
| "FROM pg_options_to_table(x.options) " |
| "ORDER BY option_name" |
| "), E',\n ') AS options " |
| "FROM pg_catalog.pg_exttable x, pg_catalog.pg_class c " |
| "WHERE x.reloid = c.oid AND c.oid = '%u'::oid ", tbinfo->dobj.catId.oid); |
| } |
| else |
| { |
| appendPQExpBuffer(query, |
| "SELECT x.urilocation, x.execlocation, x.fmttype, x.fmtopts, x.command, " |
| "x.rejectlimit, x.rejectlimittype, " |
| "x.fmterrtbl AS logerrors, " |
| "pg_catalog.pg_encoding_to_char(x.encoding), " |
| "x.writable, " |
| "array_to_string(ARRAY( " |
| "SELECT pg_catalog.quote_ident(option_name) || ' ' || " |
| "pg_catalog.quote_literal(option_value) " |
| "FROM pg_options_to_table(x.options) " |
| "ORDER BY option_name" |
| "), E',\n ') AS options " |
| "FROM pg_catalog.pg_exttable x, pg_catalog.pg_class c " |
| "WHERE x.reloid = c.oid AND c.oid = '%u'::oid ", tbinfo->dobj.catId.oid); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| { |
| if (PQntuples(res) < 1) |
| pg_log_warning("query to obtain definition of external table \"%s\" returned no data", |
| tbinfo->dobj.name); |
| else |
| pg_log_warning("query to obtain definition of external table \"%s\" returned more than one definition", |
| tbinfo->dobj.name); |
| exit_nicely(1); |
| |
| } |
| |
| |
| urilocations = PQgetvalue(res, 0, 0); |
| execlocations = PQgetvalue(res, 0, 1); |
| fmttype = PQgetvalue(res, 0, 2); |
| fmtopts = PQgetvalue(res, 0, 3); |
| command = PQgetvalue(res, 0, 4); |
| rejlim = PQgetvalue(res, 0, 5); |
| rejlimtype = PQgetvalue(res, 0, 6); |
| logerrors = PQgetvalue(res, 0, 7); |
| extencoding = PQgetvalue(res, 0, 8); |
| writable = PQgetvalue(res, 0, 9); |
| options = PQgetvalue(res, 0, 10); |
| |
| on_clause = execlocations; |
| |
| if ((command && strlen(command) > 0) || |
| (strncmp(urilocations + 1, "http", strlen("http")) == 0)) |
| isweb = true; |
| |
| if (writable && writable[0] == 't') |
| iswritable = true; |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_pg_class_oids(fout, q, tbinfo->dobj.catId.oid, false); |
| |
| appendPQExpBuffer(q, "CREATE %sEXTERNAL %sTABLE %s (", |
| (iswritable ? "WRITABLE " : ""), |
| (isweb ? "WEB " : ""), |
| qualrelname); |
| |
| int actual_atts = 0; |
| int j; |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| /* Is the attribute not dropped? */ |
| if (shouldPrintColumn(dopt, tbinfo, j)) |
| { |
| /* Format properly if not first attr */ |
| if (actual_atts > 0) |
| appendPQExpBufferChar(q, ','); |
| appendPQExpBufferStr(q, "\n "); |
| |
| /* Attribute name */ |
| appendPQExpBuffer(q, "%s ", fmtId(tbinfo->attnames[j])); |
| |
| /* Attribute type */ |
| if (tbinfo->attisdropped[j]) |
| { |
| /* |
| * ALTER TABLE DROP COLUMN clears |
| * pg_attribute.atttypid, so we will not have gotten a |
| * valid type name; insert INTEGER as a stopgap. We'll |
| * clean things up later. |
| */ |
| appendPQExpBufferStr(q, " INTEGER /* dummy */"); |
| } |
| else |
| { |
| appendPQExpBufferStr(q, tbinfo->atttypnames[j]); |
| } |
| |
| actual_atts++; |
| } |
| } |
| |
| appendPQExpBufferStr(q, "\n)"); |
| |
| if (command && strlen(command) > 0) |
| { |
| /* add EXECUTE clause */ |
| tmpstring = escape_backslashes(command, true); |
| appendPQExpBuffer(q, " EXECUTE E'%s' ", tmpstring); |
| free(tmpstring); |
| tmpstring = NULL; |
| } |
| else |
| { |
| /* add LOCATION clause, remove '{"' and '"}' */ |
| urilocations[strlen(urilocations) - 1] = '\0'; |
| urilocations++; |
| |
| /* the URI of custom protocol will contains \"\" and need to be removed */ |
| |
| location = nextToken(&urilocations, ","); |
| |
| if (location[0] == '\"') |
| { |
| location++; |
| location[strlen(location) - 1] = '\0'; |
| } |
| appendPQExpBuffer(q, " LOCATION (\n '%s'", location); |
| for (; (location = nextToken(&urilocations, ",")) != NULL;) |
| { |
| if (location[0] == '\"') |
| { |
| location++; |
| location[strlen(location) - 1] = '\0'; |
| } |
| appendPQExpBuffer(q, ",\n '%s'", location); |
| } |
| appendPQExpBufferStr(q, "\n) "); |
| } |
| |
| /* |
| * Add ON clause (unless WRITABLE table, which doesn't allow ON). |
| * ON clauses were up until 5.0 supported only on EXECUTE, in 5.0 |
| * and thereafter they are allowed on all external tables. |
| */ |
| if (!iswritable) |
| { |
| /* remove curly braces */ |
| on_clause[strlen(on_clause) - 1] = '\0'; |
| on_clause++; |
| |
| if (strncmp(on_clause, "HOST:", strlen("HOST:")) == 0) |
| appendPQExpBuffer(q, "ON HOST '%s' ", on_clause + strlen("HOST:")); |
| else if (strncmp(on_clause, "PER_HOST", strlen("PER_HOST")) == 0) |
| appendPQExpBufferStr(q, "ON HOST "); |
| else if (strncmp(on_clause, "MASTER_ONLY", strlen("MASTER_ONLY")) == 0) |
| appendPQExpBufferStr(q, "ON COORDINATOR "); |
| else if (strncmp(on_clause, "COORDINATOR_ONLY", strlen("COORDINATOR_ONLY")) == 0) |
| appendPQExpBufferStr(q, "ON COORDINATOR "); |
| else if (strncmp(on_clause, "SEGMENT_ID:", strlen("SEGMENT_ID:")) == 0) |
| appendPQExpBuffer(q, "ON SEGMENT %s ", on_clause + strlen("SEGMENT_ID:")); |
| else if (strncmp(on_clause, "TOTAL_SEGS:", strlen("TOTAL_SEGS:")) == 0) |
| appendPQExpBuffer(q, "ON %s ", on_clause + strlen("TOTAL_SEGS:")); |
| else if (strncmp(on_clause, "ALL_SEGMENTS", strlen("ALL_SEGMENTS")) == 0) |
| appendPQExpBufferStr(q, "ON ALL "); |
| else |
| { |
| pg_log_warning("illegal ON clause catalog information \"%s\" for command '%s' on table \"%s\"", |
| on_clause, command, fmtId(tbinfo->dobj.name)); |
| exit_nicely(1); |
| } |
| } |
| appendPQExpBufferChar(q, '\n'); |
| |
| /* add FORMAT clause */ |
| tmpstring = escape_fmtopts_string((const char *) fmtopts); |
| |
| switch (fmttype[0]) |
| { |
| case 't': |
| tabfmt = "text"; |
| break; |
| case 'b': |
| /* |
| * b denotes that a custom format is used. |
| * the fmtopts string should be formatted as: |
| * a1 = 'val1',...,an = 'valn' |
| * |
| */ |
| tabfmt = "custom"; |
| customfmt = custom_fmtopts_string(tmpstring); |
| break; |
| case 'a': |
| tabfmt = "avro"; |
| customfmt = custom_fmtopts_string(tmpstring); |
| break; |
| case 'p': |
| tabfmt = "parquet"; |
| customfmt = custom_fmtopts_string(tmpstring); |
| break; |
| default: |
| tabfmt = "csv"; |
| } |
| appendPQExpBuffer(q, "FORMAT '%s' (%s)\n", |
| tabfmt, |
| customfmt ? customfmt : tmpstring); |
| pg_free(tmpstring); |
| if (customfmt) |
| { |
| free(customfmt); |
| customfmt = NULL; |
| } |
| |
| if (options && options[0] != '\0') |
| { |
| appendPQExpBuffer(q, "OPTIONS (\n %s\n )\n", options); |
| } |
| |
| /* add ENCODING clause */ |
| appendPQExpBuffer(q, "ENCODING '%s'", extencoding); |
| |
| /* add Single Row Error Handling clause (if any) */ |
| if (rejlim && strlen(rejlim) > 0) |
| { |
| appendPQExpBufferChar(q, '\n'); |
| |
| /* |
| * Error tables were removed and replaced with file error |
| * logging. |
| */ |
| if (logerrors && strlen(logerrors) > 0) |
| { |
| appendPQExpBufferStr(q, "LOG ERRORS "); |
| if (logerrors[0] == 'p' && strlen(logerrors) == 1) |
| appendPQExpBufferStr(q, "PERSISTENTLY "); |
| } |
| |
| /* reject limit */ |
| appendPQExpBuffer(q, "SEGMENT REJECT LIMIT %s", rejlim); |
| |
| /* reject limit type */ |
| if (rejlimtype[0] == 'r') |
| appendPQExpBufferStr(q, " ROWS"); |
| else |
| appendPQExpBufferStr(q, " PERCENT"); |
| } |
| |
| /* DISTRIBUTED BY clause (if WRITABLE table) */ |
| if (iswritable) |
| addDistributedBy(fout, q, tbinfo, actual_atts); |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| PQclear(res); |
| |
| |
| destroyPQExpBuffer(query); |
| free(qualrelname); |
| } |
| |
| /* |
| * Create the AS clause for a view or materialized view. The semicolon is |
| * stripped because a materialized view must add a WITH NO DATA clause. |
| * |
| * This returns a new buffer which must be freed by the caller. |
| */ |
| static PQExpBuffer |
| createViewAsClause(Archive *fout, const TableInfo *tbinfo) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PQExpBuffer result = createPQExpBuffer(); |
| PGresult *res; |
| int len; |
| |
| /* Fetch the view definition */ |
| appendPQExpBuffer(query, |
| "SELECT pg_catalog.pg_get_viewdef('%u'::pg_catalog.oid) AS viewdef", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| { |
| if (PQntuples(res) < 1) |
| pg_fatal("query to obtain definition of view \"%s\" returned no data", |
| tbinfo->dobj.name); |
| else |
| pg_fatal("query to obtain definition of view \"%s\" returned more than one definition", |
| tbinfo->dobj.name); |
| } |
| |
| len = PQgetlength(res, 0, 0); |
| |
| if (len == 0) |
| pg_fatal("definition of view \"%s\" appears to be empty (length zero)", |
| tbinfo->dobj.name); |
| |
| /* Strip off the trailing semicolon so that other things may follow. */ |
| Assert(PQgetvalue(res, 0, 0)[len - 1] == ';'); |
| appendBinaryPQExpBuffer(result, PQgetvalue(res, 0, 0), len - 1); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| return result; |
| } |
| |
| /* |
| * Create a dummy AS clause for a view. This is used when the real view |
| * definition has to be postponed because of circular dependencies. |
| * We must duplicate the view's external properties -- column names and types |
| * (including collation) -- so that it works for subsequent references. |
| * |
| * This returns a new buffer which must be freed by the caller. |
| */ |
| static PQExpBuffer |
| createDummyViewAsClause(Archive *fout, const TableInfo *tbinfo) |
| { |
| PQExpBuffer result = createPQExpBuffer(); |
| int j; |
| |
| appendPQExpBufferStr(result, "SELECT"); |
| |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| if (j > 0) |
| appendPQExpBufferChar(result, ','); |
| appendPQExpBufferStr(result, "\n "); |
| |
| appendPQExpBuffer(result, "NULL::%s", tbinfo->atttypnames[j]); |
| |
| /* |
| * Must add collation if not default for the type, because CREATE OR |
| * REPLACE VIEW won't change it |
| */ |
| if (OidIsValid(tbinfo->attcollation[j])) |
| { |
| CollInfo *coll; |
| |
| coll = findCollationByOid(tbinfo->attcollation[j]); |
| if (coll) |
| appendPQExpBuffer(result, " COLLATE %s", |
| fmtQualifiedDumpable(coll)); |
| } |
| |
| appendPQExpBuffer(result, " AS %s", fmtId(tbinfo->attnames[j])); |
| } |
| |
| return result; |
| } |
| |
| /* |
| * dumpTableSchema |
| * write the declaration (not data) of one user-defined table or view |
| */ |
| static void |
| dumpTableSchema(Archive *fout, const TableInfo *tbinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q = createPQExpBuffer(); |
| PQExpBuffer delq = createPQExpBuffer(); |
| char *qrelname; |
| char *qualrelname; |
| int numParents; |
| TableInfo **parents; |
| int actual_atts; /* number of attrs in this CREATE statement */ |
| const char *reltypename; |
| char *storage; |
| int j, |
| k; |
| bool hasExternalPartitions = false; |
| char *ftoptions = NULL; |
| char *srvname = NULL; |
| char *foreign = ""; |
| |
| |
| /* We had better have loaded per-column details about this table */ |
| Assert(tbinfo->interesting); |
| |
| qrelname = pg_strdup(fmtId(tbinfo->dobj.name)); |
| qualrelname = pg_strdup(fmtQualifiedDumpable(tbinfo)); |
| |
| if (tbinfo->hasoids) |
| pg_log_warning("WITH OIDS is not supported anymore (table \"%s\")", |
| qrelname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_type_oids_by_rel(fout, q, tbinfo); |
| |
| /* Is it a table or a view? */ |
| if (tbinfo->relkind == RELKIND_VIEW) |
| { |
| PQExpBuffer result; |
| |
| /* |
| * Note: keep this code in sync with the is_view case in dumpRule() |
| */ |
| |
| reltypename = "VIEW"; |
| |
| appendPQExpBuffer(delq, "DROP VIEW %s;\n", qualrelname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_pg_class_oids(fout, q, |
| tbinfo->dobj.catId.oid, false); |
| |
| appendPQExpBuffer(q, "CREATE VIEW %s", qualrelname); |
| |
| if (tbinfo->dummy_view) |
| result = createDummyViewAsClause(fout, tbinfo); |
| else |
| { |
| if (nonemptyReloptions(tbinfo->reloptions)) |
| { |
| appendPQExpBufferStr(q, " WITH ("); |
| appendReloptionsArrayAH(q, tbinfo->reloptions, "", fout); |
| appendPQExpBufferChar(q, ')'); |
| } |
| result = createViewAsClause(fout, tbinfo); |
| } |
| appendPQExpBuffer(q, " AS\n%s", result->data); |
| destroyPQExpBuffer(result); |
| |
| if (tbinfo->checkoption != NULL && !tbinfo->dummy_view) |
| appendPQExpBuffer(q, "\n WITH %s CHECK OPTION", tbinfo->checkoption); |
| appendPQExpBufferStr(q, ";\n"); |
| } |
| else |
| { |
| <<<<<<< HEAD |
| ======= |
| char *partkeydef = NULL; |
| char *ftoptions = NULL; |
| char *srvname = NULL; |
| char *foreign = ""; |
| |
| /* |
| * Set reltypename, and collect any relkind-specific data that we |
| * didn't fetch during getTables(). |
| */ |
| >>>>>>> REL_16_9 |
| switch (tbinfo->relkind) |
| { |
| case RELKIND_PARTITIONED_TABLE: |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| |
| reltypename = "TABLE"; |
| |
| /* retrieve partition key definition */ |
| appendPQExpBuffer(query, |
| "SELECT pg_get_partkeydef('%u')", |
| tbinfo->dobj.catId.oid); |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| partkeydef = pg_strdup(PQgetvalue(res, 0, 0)); |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| break; |
| } |
| case RELKIND_FOREIGN_TABLE: |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| int i_srvname; |
| int i_ftoptions; |
| |
| reltypename = "FOREIGN TABLE"; |
| |
| /* retrieve name of foreign server and generic options */ |
| appendPQExpBuffer(query, |
| "SELECT fs.srvname, " |
| "pg_catalog.array_to_string(ARRAY(" |
| "SELECT pg_catalog.quote_ident(option_name) || " |
| "' ' || pg_catalog.quote_literal(option_value) " |
| "FROM pg_catalog.pg_options_to_table(ftoptions) " |
| "ORDER BY option_name" |
| "), E',\n ') AS ftoptions " |
| "FROM pg_catalog.pg_foreign_table ft " |
| "JOIN pg_catalog.pg_foreign_server fs " |
| "ON (fs.oid = ft.ftserver) " |
| "WHERE ft.ftrelid = '%u'", |
| tbinfo->dobj.catId.oid); |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| i_srvname = PQfnumber(res, "srvname"); |
| i_ftoptions = PQfnumber(res, "ftoptions"); |
| srvname = pg_strdup(PQgetvalue(res, 0, i_srvname)); |
| ftoptions = pg_strdup(PQgetvalue(res, 0, i_ftoptions)); |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| foreign = "FOREIGN "; |
| break; |
| } |
| case RELKIND_MATVIEW: |
| if (tbinfo->isdynamic) |
| reltypename = "DYNAMIC TABLE"; |
| else |
| reltypename = "MATERIALIZED VIEW"; |
| break; |
| default: |
| reltypename = "TABLE"; |
| <<<<<<< HEAD |
| ======= |
| break; |
| } |
| >>>>>>> REL_16_9 |
| |
| /* Is it an external table (server GPDB 6.x and below.) */ |
| if (tbinfo->relstorage == RELSTORAGE_EXTERNAL) |
| reltypename = "EXTERNAL TABLE"; |
| } |
| } |
| |
| if (strcmp(reltypename, "EXTERNAL TABLE") == 0) |
| { |
| dumpExternal(fout, tbinfo, q, delq); |
| } |
| else if (strcmp(reltypename, "VIEW") == 0) |
| { |
| /* handled above already */ |
| } |
| else |
| { |
| numParents = tbinfo->numParents; |
| parents = tbinfo->parents; |
| |
| appendPQExpBuffer(delq, "DROP %s %s;\n", reltypename, qualrelname); |
| |
| if (dopt->binary_upgrade) |
| { |
| binary_upgrade_set_pg_class_oids(fout, q, |
| tbinfo->dobj.catId.oid, false); |
| |
| /* |
| * If this is a partition hierarchy parent, dump the Oid |
| * preassignments for each partition member individually. |
| * |
| * GPDB_12_MERGE_FIXME: Disabled when dumping GPDB 7. But do |
| * we need it when dumping older versions, either? The presumed |
| * *target* is the current latest version anyway. |
| */ |
| if (tbinfo->parparent && fout->remoteVersion < 100000) |
| { |
| PQExpBuffer partquery = createPQExpBuffer(); |
| PGresult *partres; |
| |
| appendPQExpBuffer(partquery, "SELECT DISTINCT(child.oid) " |
| "FROM pg_catalog.pg_partition part, " |
| " pg_catalog.pg_partition_rule rule, " |
| " pg_catalog.pg_class child " |
| "WHERE part.parrelid = '%u'::pg_catalog.oid " |
| " AND rule.paroid = part.oid " |
| " AND child.oid = rule.parchildrelid", |
| tbinfo->dobj.catId.oid); |
| partres = ExecuteSqlQuery(fout, partquery->data, PGRES_TUPLES_OK); |
| |
| /* It really should.. */ |
| if (PQntuples(partres) > 0) |
| { |
| int i; |
| |
| for (i = 0; i < PQntuples(partres); i++) |
| { |
| Oid part_oid = atooid(PQgetvalue(partres, i, 0)); |
| |
| binary_upgrade_set_pg_class_oids(fout, q, part_oid, false); |
| binary_upgrade_set_type_oids_by_rel_oid(fout, q, part_oid); |
| } |
| } |
| |
| PQclear(partres); |
| destroyPQExpBuffer(partquery); |
| } |
| } |
| |
| if (tbinfo->relkind == RELKIND_MATVIEW && tbinfo->isdynamic) |
| { |
| /* We'r sure there is no UNLOGGED and this is a DYNAMIC TABLE. */ |
| appendPQExpBuffer(q, "CREATE DYNAMIC TABLE %s", qualrelname); |
| addSchedule(fout, q, tbinfo); |
| } |
| else |
| { |
| appendPQExpBuffer(q, "CREATE %s%s%s %s", |
| tbinfo->relpersistence == RELPERSISTENCE_UNLOGGED ? |
| "UNLOGGED " : "", |
| tbinfo->relkind == RELKIND_MATVIEW && tbinfo->isivm ? |
| "INCREMENTAL " : "", |
| reltypename, |
| qualrelname); |
| |
| } |
| /* |
| * Attach to type, if reloftype; except in case of a binary upgrade, |
| * we dump the table normally and attach it to the type afterward. |
| */ |
| if (OidIsValid(tbinfo->reloftype) && !dopt->binary_upgrade) |
| appendPQExpBuffer(q, " OF %s", |
| getFormattedTypeName(fout, tbinfo->reloftype, |
| zeroIsError)); |
| |
| if (tbinfo->relkind != RELKIND_MATVIEW) |
| { |
| /* Dump the attributes */ |
| actual_atts = 0; |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| /* |
| * Normally, dump if it's locally defined in this table, and |
| * not dropped. But for binary upgrade, we'll dump all the |
| * columns, and then fix up the dropped and nonlocal cases |
| * below. |
| */ |
| if (shouldPrintColumn(dopt, tbinfo, j)) |
| { |
| bool print_default; |
| bool print_notnull; |
| |
| /* |
| * Default value --- suppress if to be printed separately |
| * or not at all. |
| */ |
| print_default = (tbinfo->attrdefs[j] != NULL && |
| tbinfo->attrdefs[j]->dobj.dump && |
| !tbinfo->attrdefs[j]->separate); |
| |
| /* |
| * Not Null constraint --- suppress if inherited, except |
| * if partition, or in binary-upgrade case where that |
| * won't work. |
| */ |
| print_notnull = (tbinfo->notnull[j] && |
| (!tbinfo->inhNotNull[j] || |
| tbinfo->ispartition || dopt->binary_upgrade)); |
| |
| /* |
| * Skip column if fully defined by reloftype, except in |
| * binary upgrade |
| */ |
| if (OidIsValid(tbinfo->reloftype) && |
| !print_default && !print_notnull && |
| !dopt->binary_upgrade) |
| continue; |
| |
| /* Format properly if not first attr */ |
| if (actual_atts == 0) |
| appendPQExpBufferStr(q, " ("); |
| else |
| appendPQExpBufferChar(q, ','); |
| appendPQExpBufferStr(q, "\n "); |
| actual_atts++; |
| |
| /* Attribute name */ |
| appendPQExpBufferStr(q, fmtId(tbinfo->attnames[j])); |
| |
| if (tbinfo->attisdropped[j]) |
| { |
| /* |
| * ALTER TABLE DROP COLUMN clears |
| * pg_attribute.atttypid, so we will not have gotten a |
| * valid type name; insert INTEGER as a stopgap. We'll |
| * clean things up later. |
| */ |
| appendPQExpBufferStr(q, " INTEGER /* dummy */"); |
| |
| /* Dropped columns are dumped during binary upgrade. |
| * Dump the encoding clause also to maintain a consistent |
| * catalog entry in pg_attribute_encoding post upgrade. |
| */ |
| if (tbinfo->attencoding[j] != NULL) |
| appendPQExpBuffer(q, " ENCODING (%s)", tbinfo->attencoding[j]); |
| |
| /* Skip all the rest */ |
| continue; |
| } |
| |
| /* |
| * Attribute type; print it except when creating a typed |
| * table ('OF type_name'), but in binary-upgrade mode, |
| * print it in that case too. |
| */ |
| if (dopt->binary_upgrade || !OidIsValid(tbinfo->reloftype)) |
| { |
| appendPQExpBuffer(q, " %s", |
| tbinfo->atttypnames[j]); |
| } |
| |
| if (print_default) |
| { |
| if (tbinfo->attgenerated[j] == ATTRIBUTE_GENERATED_STORED) |
| appendPQExpBuffer(q, " GENERATED ALWAYS AS (%s) STORED", |
| tbinfo->attrdefs[j]->adef_expr); |
| else |
| appendPQExpBuffer(q, " DEFAULT %s", |
| tbinfo->attrdefs[j]->adef_expr); |
| } |
| |
| |
| if (print_notnull) |
| appendPQExpBufferStr(q, " NOT NULL"); |
| |
| /* Add collation if not default for the type */ |
| if (OidIsValid(tbinfo->attcollation[j])) |
| { |
| CollInfo *coll; |
| |
| coll = findCollationByOid(tbinfo->attcollation[j]); |
| if (coll) |
| appendPQExpBuffer(q, " COLLATE %s", |
| fmtQualifiedDumpable(coll)); |
| } |
| } |
| } |
| |
| /* |
| * Add non-inherited CHECK constraints, if any. |
| * |
| * For partitions, we need to include check constraints even if |
| * they're not defined locally, because the ALTER TABLE ATTACH |
| * PARTITION that we'll emit later expects the constraint to be |
| * there. (No need to fix conislocal: ATTACH PARTITION does that) |
| */ |
| for (j = 0; j < tbinfo->ncheck; j++) |
| { |
| ConstraintInfo *constr = &(tbinfo->checkexprs[j]); |
| |
| if (constr->separate || |
| (!constr->conislocal && !tbinfo->ispartition)) |
| continue; |
| |
| if (actual_atts == 0) |
| appendPQExpBufferStr(q, " (\n "); |
| else |
| appendPQExpBufferStr(q, ",\n "); |
| |
| appendPQExpBuffer(q, "CONSTRAINT %s ", |
| fmtId(constr->dobj.name)); |
| appendPQExpBufferStr(q, constr->condef); |
| |
| actual_atts++; |
| } |
| |
| /* |
| * Add AOCO ENCODING directives, if any. |
| * |
| * We dump these as separate "COLUMN <col> ENCODING ..." clauses, |
| * instead of tacking the ENCODING at the column definition, so |
| * that this works for inherited columns, too. Inherited columns |
| * are not listed in the column list. |
| */ |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| if (tbinfo->attisdropped[j]) |
| continue; |
| |
| if (tbinfo->attencoding[j] != NULL) |
| { |
| if (actual_atts == 0) |
| appendPQExpBufferStr(q, " (\n "); |
| else |
| appendPQExpBufferStr(q, ",\n "); |
| |
| appendPQExpBuffer(q, "COLUMN %s ENCODING (%s)", |
| fmtId(tbinfo->attnames[j]), |
| tbinfo->attencoding[j]); |
| actual_atts++; |
| } |
| } |
| |
| if (actual_atts) |
| appendPQExpBufferStr(q, "\n)"); |
| else if (!(OidIsValid(tbinfo->reloftype) && !dopt->binary_upgrade)) |
| { |
| /* |
| * No attributes? we must have a parenthesized attribute list, |
| * even though empty, when not using the OF TYPE syntax. |
| */ |
| appendPQExpBufferStr(q, " (\n)"); |
| } |
| |
| /* |
| * Emit the INHERITS clause (not for partitions), except in |
| * binary-upgrade mode. |
| */ |
| if (numParents > 0 && !tbinfo->ispartition && |
| !dopt->binary_upgrade) |
| { |
| appendPQExpBufferStr(q, "\nINHERITS ("); |
| for (k = 0; k < numParents; k++) |
| { |
| TableInfo *parentRel = parents[k]; |
| |
| if (k > 0) |
| appendPQExpBufferStr(q, ", "); |
| appendPQExpBufferStr(q, fmtQualifiedDumpable(parentRel)); |
| } |
| appendPQExpBufferChar(q, ')'); |
| } |
| |
| if (tbinfo->relkind == RELKIND_PARTITIONED_TABLE) |
| appendPQExpBuffer(q, "\nPARTITION BY %s", partkeydef); |
| |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE) |
| appendPQExpBuffer(q, "\nSERVER %s", fmtId(srvname)); |
| } |
| else |
| actual_atts = 0; |
| |
| if (nonemptyReloptions(tbinfo->reloptions) || |
| nonemptyReloptions(tbinfo->toast_reloptions)) |
| { |
| bool addcomma = false; |
| |
| appendPQExpBufferStr(q, "\nWITH ("); |
| if (nonemptyReloptions(tbinfo->reloptions)) |
| { |
| addcomma = true; |
| appendReloptionsArrayAH(q, tbinfo->reloptions, "", fout); |
| } |
| if (nonemptyReloptions(tbinfo->toast_reloptions)) |
| { |
| if (addcomma) |
| appendPQExpBufferStr(q, ", "); |
| appendReloptionsArrayAH(q, tbinfo->toast_reloptions, "toast.", |
| fout); |
| } |
| appendPQExpBufferChar(q, ')'); |
| } |
| |
| |
| /* |
| * For materialized views, create the AS clause just like a view. At |
| * this point, we always mark the view as not populated. |
| */ |
| if (tbinfo->relkind == RELKIND_MATVIEW) |
| { |
| PQExpBuffer result; |
| |
| result = createViewAsClause(fout, tbinfo); |
| appendPQExpBuffer(q, " AS\n%s\n WITH NO DATA", |
| result->data); |
| destroyPQExpBuffer(result); |
| } |
| |
| /* START MPP ADDITION */ |
| |
| /* |
| * Dump distributed by clause. |
| */ |
| if (dumpGpPolicy && tbinfo->relkind != RELKIND_FOREIGN_TABLE) |
| addDistributedBy(fout, q, tbinfo, actual_atts); |
| |
| /* |
| * If GP partitioning is supported add the partitioning constraints to |
| * the table definition. |
| */ |
| if (gp_partitioning_available) |
| { |
| bool isTemplatesSupported = fout->remoteVersion >= 80214; |
| bool isPartitioned = false; |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| |
| /* does support GP partitioning. */ |
| resetPQExpBuffer(query); |
| /* MPP-6297: dump by tablename */ |
| if (isTemplatesSupported) |
| /* use 4.x version of function */ |
| appendPQExpBuffer(query, "SELECT " |
| "pg_get_partition_def('%u'::pg_catalog.oid, true, true) ", |
| tbinfo->dobj.catId.oid); |
| else /* use 3.x version of function */ |
| appendPQExpBuffer(query, "SELECT " |
| "pg_get_partition_def('%u'::pg_catalog.oid, true) ", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| { |
| if (PQntuples(res) < 1) |
| pg_log_warning("query to obtain definition of table \"%s\" returned no data", |
| tbinfo->dobj.name); |
| else |
| pg_log_warning("query to obtain definition of table \"%s\" returned more than one definition", |
| tbinfo->dobj.name); |
| exit_nicely(1); |
| } |
| isPartitioned = !PQgetisnull(res, 0, 0); |
| if (isPartitioned) |
| appendPQExpBuffer(q, " %s", PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| |
| /* |
| * MPP-6095: dump ALTER TABLE statements for subpartition |
| * templates |
| */ |
| if (isTemplatesSupported) |
| { |
| resetPQExpBuffer(query); |
| |
| appendPQExpBuffer( |
| query, "SELECT " |
| "pg_get_partition_template_def('%u'::pg_catalog.oid, true, true) ", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| { |
| if (PQntuples(res) < 1) |
| pg_log_warning("query to obtain definition of table \"%s\" returned no data", |
| tbinfo->dobj.name); |
| else |
| pg_log_warning("query to obtain definition of table \"%s\" returned more than one definition", |
| tbinfo->dobj.name); |
| exit_nicely(1); |
| } |
| |
| /* |
| * MPP-9537: terminate (with semicolon) the previous |
| * statement, and dump the template definitions |
| */ |
| if (!PQgetisnull(res, 0, 0) && |
| PQgetlength(res, 0, 0)) |
| appendPQExpBuffer(q, ";\n %s", PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| } |
| |
| if (isPartitioned) |
| { |
| /* |
| * Find out if there are any external partitions. |
| * |
| * In GPDB 6.X and below, external tables have |
| * relkind=RELKIND_RELATION and relstorage=RELSTORAGE_EXTERNAL. |
| * In later versions, they are just foreign tables. This query |
| * works on all server versions. |
| */ |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, "SELECT EXISTS (SELECT 1 " |
| " FROM pg_class part " |
| " JOIN pg_partition_rule pr ON (part.oid = pr.parchildrelid) " |
| " JOIN pg_partition p ON (pr.paroid = p.oid) " |
| "WHERE p.parrelid = '%u'::pg_catalog.oid " |
| " AND (part.relstorage = '%c' OR part.relkind = '%c')) " |
| "AS has_external_partitions;", |
| tbinfo->dobj.catId.oid, |
| RELSTORAGE_EXTERNAL, |
| RELKIND_FOREIGN_TABLE); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| hasExternalPartitions = (PQgetvalue(res, 0, 0)[0] == 't'); |
| PQclear(res); |
| } |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* END MPP ADDITION */ |
| |
| /* Dump generic options if any */ |
| if (ftoptions && ftoptions[0]) |
| appendPQExpBuffer(q, "\nOPTIONS (\n %s\n)", ftoptions); |
| |
| if (dumpGpPolicy && tbinfo->relkind == RELKIND_FOREIGN_TABLE && strcmp(srvname, GP_EXTTABLE_SERVER_NAME) == 0) |
| addDistributedBy(fout, q, tbinfo, actual_atts); |
| |
| appendPQExpBufferStr(q, ";\n"); |
| |
| /* |
| * Exchange external partitions. This is an expensive process, so only |
| * run it if we've found evidence of external partitions up above. |
| */ |
| if (hasExternalPartitions) |
| { |
| int i = 0; |
| int ntups = 0; |
| char *relname = NULL; |
| int i_relname = 0; |
| int i_parname = 0; |
| int i_partitionrank = 0; |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| |
| /* |
| * The multiple JOINs below trigger an apparent planner bug which |
| * may effectively hang the backend. This bug is present in both |
| * released versions of GPDB and the current development tip at time |
| * of writing. Disable nestloops temporarily as a workaround. |
| * |
| * TODO: when this bug is fixed, version-gate this code so that we |
| * don't run it on well-behaved backends. |
| */ |
| ExecuteSqlStatement(fout, "SET enable_nestloop TO off"); |
| |
| appendPQExpBuffer(query, "SELECT DISTINCT cc.relname, ps.partitionrank, pp.parname " |
| "FROM pg_partition p " |
| "JOIN pg_class c on (p.parrelid = c.oid) " |
| "JOIN pg_partitions ps on (c.relname = ps.tablename) " |
| "JOIN pg_class cc on (ps.partitiontablename = cc.relname) " |
| "JOIN pg_partition_rule pp on (cc.oid = pp.parchildrelid) " |
| "WHERE p.parrelid = %u AND (cc.relstorage='%c' OR cc.relkind = '%c');", |
| tbinfo->dobj.catId.oid, |
| RELSTORAGE_EXTERNAL, RELKIND_FOREIGN_TABLE); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| i_relname = PQfnumber(res, "relname"); |
| i_parname = PQfnumber(res, "parname"); |
| i_partitionrank = PQfnumber(res, "partitionrank"); |
| |
| /* FIXME: does this code handle external SUBPARTITIONs correctly? */ |
| for (i = 0; i < ntups; i++) |
| { |
| char tmpExtTable[500] = {0}; |
| relname = pg_strdup(PQgetvalue(res, i, i_relname)); |
| snprintf(tmpExtTable, sizeof(tmpExtTable), "%s%s", relname, EXT_PARTITION_NAME_POSTFIX); |
| char *qualTmpExtTable = pg_strdup(fmtQualifiedId(tbinfo->dobj.namespace->dobj.name, |
| tmpExtTable)); |
| |
| appendPQExpBuffer(q, "ALTER TABLE %s ", qualrelname); |
| /* |
| * If it is an anonymous range partition we must exchange for |
| * the rank rather than the parname. |
| */ |
| if (PQgetisnull(res, i, i_parname) || !strlen(PQgetvalue(res, i, i_parname))) |
| { |
| appendPQExpBuffer(q, "EXCHANGE PARTITION FOR (RANK(%s)) ", |
| PQgetvalue(res, i, i_partitionrank)); |
| } |
| else |
| { |
| appendPQExpBuffer(q, "EXCHANGE PARTITION %s ", |
| fmtId(PQgetvalue(res, i, i_parname))); |
| } |
| appendPQExpBuffer(q, "WITH TABLE %s WITHOUT VALIDATION; ", qualTmpExtTable); |
| |
| appendPQExpBuffer(q, "\n"); |
| |
| appendPQExpBuffer(q, "DROP TABLE %s; ", qualTmpExtTable); |
| |
| appendPQExpBuffer(q, "\n"); |
| free(relname); |
| free(qualTmpExtTable); |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| /* TODO: version-gate this when the planner bug is fixed; see above. */ |
| ExecuteSqlStatement(fout, "SET enable_nestloop TO on"); |
| } |
| |
| /* Materialized views can depend on extensions */ |
| if (tbinfo->relkind == RELKIND_MATVIEW) |
| append_depends_on_extension(fout, q, &tbinfo->dobj, |
| "pg_catalog.pg_class", |
| "MATERIALIZED VIEW", |
| qualrelname); |
| |
| /* |
| * in binary upgrade mode, update the catalog with any missing values |
| * that might be present. |
| */ |
| if (dopt->binary_upgrade) |
| { |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| if (tbinfo->attmissingval[j][0] != '\0') |
| { |
| appendPQExpBufferStr(q, "\n-- set missing value.\n"); |
| appendPQExpBufferStr(q, |
| "SELECT pg_catalog.binary_upgrade_set_missing_value("); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass,"); |
| appendStringLiteralAH(q, tbinfo->attnames[j], fout); |
| appendPQExpBufferChar(q, ','); |
| appendStringLiteralAH(q, tbinfo->attmissingval[j], fout); |
| appendPQExpBufferStr(q, ");\n\n"); |
| } |
| } |
| } |
| |
| /* |
| * To create binary-compatible heap files, we have to ensure the same |
| * physical column order, including dropped columns, as in the |
| * original. Therefore, we create dropped columns above and drop them |
| * here, also updating their attlen/attalign values so that the |
| * dropped column can be skipped properly. (We do not bother with |
| * restoring the original attbyval setting.) Also, inheritance |
| * relationships are set up by doing ALTER TABLE INHERIT rather than |
| * using an INHERITS clause --- the latter would possibly mess up the |
| * column order. That also means we have to take care about setting |
| * attislocal correctly, plus fix up any inherited CHECK constraints. |
| * Analogously, we set up typed tables using ALTER TABLE / OF here. |
| * |
| * We process foreign and partitioned tables here, even though they |
| * lack heap storage, because they can participate in inheritance |
| * relationships and we want this stuff to be consistent across the |
| * inheritance tree. We can exclude indexes, toast tables, sequences |
| * and matviews, even though they have storage, because we don't |
| * support altering or dropping columns in them, nor can they be part |
| * of inheritance trees. |
| */ |
| if (dopt->binary_upgrade && |
| (tbinfo->relkind == RELKIND_RELATION || |
| tbinfo->relkind == RELKIND_FOREIGN_TABLE || |
| tbinfo->relkind == RELKIND_PARTITIONED_TABLE)) |
| { |
| /* |
| * Cloudberry doesn't allow altering system catalogs without |
| * setting the allow_system_table_mods GUC first. |
| */ |
| appendPQExpBuffer(q, "SET allow_system_table_mods = true;\n"); |
| |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| if (tbinfo->attisdropped[j]) |
| { |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, recreate dropped column.\n"); |
| appendPQExpBuffer(q, "UPDATE pg_catalog.pg_attribute\n" |
| "SET attlen = %d, " |
| "attalign = '%c', attbyval = false\n" |
| "WHERE attname = ", |
| tbinfo->attlen[j], |
| tbinfo->attalign[j]); |
| appendStringLiteralAH(q, tbinfo->attnames[j], fout); |
| |
| if (gp_partitioning_available) |
| { |
| /* |
| * Do for all descendants of a partition table. |
| * No hurt if this is not a table with partitions. |
| */ |
| appendPQExpBufferStr(q, "\n AND attrelid IN (SELECT "); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass "); |
| appendPQExpBufferStr(q, "UNION SELECT pr.parchildrelid FROM " |
| "pg_catalog.pg_partition_rule pr, " |
| "pg_catalog.pg_partition p WHERE " |
| "pr.parchildrelid != 0 AND " |
| "pr.paroid = p.oid AND p.parrelid = "); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass);\n"); |
| } |
| else |
| { |
| appendPQExpBufferStr(q, "\n AND attrelid = "); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass;\n"); |
| } |
| |
| /* |
| * GPDB: Upstream uses ALTER TABLE ONLY below. Because we |
| * need to cascade the DROP down into partitions as well, |
| * we use ALTER TABLE instead. |
| * |
| * At the moment, we believe this does not cause problems |
| * for vanilla inherited tables, because the tables aren't |
| * plugged into the inheritance hierarchy until after this |
| * code is run (see the ALTER TABLE ... INHERIT below), and |
| * therefore ALTER TABLE and ALTER TABLE ONLY are |
| * effectively the same. If that changes, this will need to |
| * be revisited. |
| */ |
| if (tbinfo->relkind == RELKIND_RELATION) |
| appendPQExpBuffer(q, "ALTER TABLE %s ", |
| qualrelname); |
| else if (tbinfo->relkind == RELKIND_PARTITIONED_TABLE) |
| appendPQExpBuffer(q, "ALTER TABLE ONLY %s ", |
| qualrelname); |
| else |
| appendPQExpBuffer(q, "ALTER FOREIGN TABLE ONLY %s ", |
| qualrelname); |
| appendPQExpBuffer(q, "DROP COLUMN %s;\n", |
| fmtId(tbinfo->attnames[j])); |
| } |
| else if (!tbinfo->attislocal[j]) |
| { |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, recreate inherited column.\n"); |
| appendPQExpBufferStr(q, "UPDATE pg_catalog.pg_attribute\n" |
| "SET attislocal = false\n" |
| "WHERE attname = "); |
| appendStringLiteralAH(q, tbinfo->attnames[j], fout); |
| appendPQExpBufferStr(q, "\n AND attrelid = "); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass;\n"); |
| } |
| } |
| |
| /* |
| * Add inherited CHECK constraints, if any. |
| * |
| * For partitions, they were already dumped, and conislocal |
| * doesn't need fixing. |
| */ |
| for (k = 0; k < tbinfo->ncheck; k++) |
| { |
| ConstraintInfo *constr = &(tbinfo->checkexprs[k]); |
| |
| if (constr->separate || constr->conislocal || tbinfo->ispartition) |
| continue; |
| |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, set up inherited constraint.\n"); |
| appendPQExpBuffer(q, "ALTER TABLE ONLY %s ADD CONSTRAINT %s %s;\n", |
| qualrelname, |
| fmtId(constr->dobj.name), |
| constr->condef); |
| appendPQExpBufferStr(q, "UPDATE pg_catalog.pg_constraint\n" |
| "SET conislocal = false\n" |
| "WHERE contype = 'c' AND conname = "); |
| appendStringLiteralAH(q, constr->dobj.name, fout); |
| appendPQExpBufferStr(q, "\n AND conrelid = "); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass;\n"); |
| } |
| |
| if (numParents > 0 && !tbinfo->ispartition) |
| { |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, set up inheritance this way.\n"); |
| for (k = 0; k < numParents; k++) |
| { |
| TableInfo *parentRel = parents[k]; |
| |
| appendPQExpBuffer(q, "ALTER %sTABLE ONLY %s INHERIT %s;\n", foreign, |
| qualrelname, |
| fmtQualifiedDumpable(parentRel)); |
| } |
| } |
| |
| if (OidIsValid(tbinfo->reloftype)) |
| { |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, set up typed tables this way.\n"); |
| appendPQExpBuffer(q, "ALTER TABLE ONLY %s OF %s;\n", |
| qualrelname, |
| getFormattedTypeName(fout, tbinfo->reloftype, |
| zeroIsError)); |
| } |
| appendPQExpBuffer(q, "RESET allow_system_table_mods;\n"); |
| } |
| |
| /* |
| * In binary_upgrade mode, arrange to restore the old relfrozenxid and |
| * relminmxid of all vacuumable relations. (While vacuum.c processes |
| * TOAST tables semi-independently, here we see them only as children |
| * of other relations; so this "if" lacks RELKIND_TOASTVALUE, and the |
| * child toast table is handled below.) |
| */ |
| if (dopt->binary_upgrade && |
| (tbinfo->relkind == RELKIND_RELATION || |
| tbinfo->relkind == RELKIND_MATVIEW)) |
| { |
| appendPQExpBufferStr(q, "SET allow_system_table_mods = true;\n"); |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, set heap's relfrozenxid and relminmxid\n"); |
| appendPQExpBuffer(q, "UPDATE pg_catalog.pg_class\n" |
| "SET relfrozenxid = '%u', relminmxid = '%u'\n" |
| "WHERE oid = ", |
| tbinfo->frozenxid, tbinfo->minmxid); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass;\n"); |
| |
| if (tbinfo->toast_oid) |
| { |
| /* |
| * The toast table will have the same OID at restore, so we |
| * can safely target it by OID. |
| */ |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, set toast's relfrozenxid and relminmxid\n"); |
| appendPQExpBuffer(q, "UPDATE pg_catalog.pg_class\n" |
| "SET relfrozenxid = '%u', relminmxid = '%u'\n" |
| "WHERE oid = '%u';\n", |
| tbinfo->toast_frozenxid, |
| tbinfo->toast_minmxid, tbinfo->toast_oid); |
| } |
| |
| /* |
| * We have probably bumped allow_system_table_mods to true in the |
| * above processing, but even we didn't let's just reset it here |
| * since it doesn't to do any harm to. |
| */ |
| appendPQExpBuffer(q, "RESET allow_system_table_mods;\n"); |
| } |
| |
| /* |
| * In binary_upgrade mode, restore matviews' populated status by |
| * poking pg_class directly. This is pretty ugly, but we can't use |
| * REFRESH MATERIALIZED VIEW since it's possible that some underlying |
| * matview is not populated even though this matview is; in any case, |
| * we want to transfer the matview's heap storage, not run REFRESH. |
| */ |
| if (dopt->binary_upgrade && tbinfo->relkind == RELKIND_MATVIEW && |
| tbinfo->relispopulated) |
| { |
| appendPQExpBuffer(q, "SET allow_system_table_mods = true;\n"); |
| |
| appendPQExpBufferStr(q, "\n-- For binary upgrade, mark materialized view as populated\n"); |
| appendPQExpBufferStr(q, "UPDATE pg_catalog.pg_class\n" |
| "SET relispopulated = 't'\n" |
| "WHERE oid = "); |
| appendStringLiteralAH(q, qualrelname, fout); |
| appendPQExpBufferStr(q, "::pg_catalog.regclass;\n"); |
| |
| appendPQExpBuffer(q, "RESET allow_system_table_mods;\n"); |
| } |
| |
| /* |
| * Dump additional per-column properties that we can't handle in the |
| * main CREATE TABLE command. |
| */ |
| for (j = 0; j < tbinfo->numatts; j++) |
| { |
| /* None of this applies to dropped columns */ |
| if (tbinfo->attisdropped[j]) |
| continue; |
| |
| /* |
| * If we didn't dump the column definition explicitly above, and |
| * it is NOT NULL and did not inherit that property from a parent, |
| * we have to mark it separately. |
| */ |
| if (!shouldPrintColumn(dopt, tbinfo, j) && |
| tbinfo->notnull[j] && !tbinfo->inhNotNull[j]) |
| appendPQExpBuffer(q, |
| "ALTER %sTABLE ONLY %s ALTER COLUMN %s SET NOT NULL;\n", |
| foreign, qualrelname, |
| fmtId(tbinfo->attnames[j])); |
| |
| /* |
| * Dump per-column statistics information. We only issue an ALTER |
| * TABLE statement if the attstattarget entry for this column is |
| * non-negative (i.e. it's not the default value) |
| */ |
| if (tbinfo->attstattarget[j] >= 0) |
| appendPQExpBuffer(q, "ALTER %sTABLE ONLY %s ALTER COLUMN %s SET STATISTICS %d;\n", |
| foreign, qualrelname, |
| fmtId(tbinfo->attnames[j]), |
| tbinfo->attstattarget[j]); |
| |
| /* |
| * Dump per-column storage information. The statement is only |
| * dumped if the storage has been changed from the type's default. |
| * An inherited column can have |
| * its storage type changed independently from the parent |
| * specification. |
| */ |
| if (tbinfo->attstorage[j] != tbinfo->typstorage[j]) |
| { |
| switch (tbinfo->attstorage[j]) |
| { |
| case TYPSTORAGE_PLAIN: |
| storage = "PLAIN"; |
| break; |
| case TYPSTORAGE_EXTERNAL: |
| storage = "EXTERNAL"; |
| break; |
| case TYPSTORAGE_EXTENDED: |
| storage = "EXTENDED"; |
| break; |
| case TYPSTORAGE_MAIN: |
| storage = "MAIN"; |
| break; |
| default: |
| storage = NULL; |
| } |
| |
| /* |
| * Only dump the statement if it's a storage type we recognize |
| */ |
| if (storage != NULL) |
| appendPQExpBuffer(q, "ALTER %sTABLE ONLY %s ALTER COLUMN %s SET STORAGE %s;\n", |
| foreign, qualrelname, |
| fmtId(tbinfo->attnames[j]), |
| storage); |
| } |
| |
| /* |
| * Dump per-column compression, if it's been set. |
| */ |
| if (!dopt->no_toast_compression) |
| { |
| const char *cmname; |
| |
| switch (tbinfo->attcompression[j]) |
| { |
| case 'p': |
| cmname = "pglz"; |
| break; |
| case 'l': |
| cmname = "lz4"; |
| break; |
| default: |
| cmname = NULL; |
| break; |
| } |
| |
| if (cmname != NULL) |
| appendPQExpBuffer(q, "ALTER %sTABLE ONLY %s ALTER COLUMN %s SET COMPRESSION %s;\n", |
| foreign, qualrelname, |
| fmtId(tbinfo->attnames[j]), |
| cmname); |
| } |
| |
| /* |
| * Dump per-column attributes. |
| */ |
| if (tbinfo->attoptions[j][0] != '\0') |
| appendPQExpBuffer(q, "ALTER %sTABLE ONLY %s ALTER COLUMN %s SET (%s);\n", |
| foreign, qualrelname, |
| fmtId(tbinfo->attnames[j]), |
| tbinfo->attoptions[j]); |
| |
| /* |
| * Dump per-column fdw options. |
| */ |
| if (tbinfo->relkind == RELKIND_FOREIGN_TABLE && |
| tbinfo->attfdwoptions[j][0] != '\0') |
| appendPQExpBuffer(q, |
| "ALTER FOREIGN TABLE %s ALTER COLUMN %s OPTIONS (\n" |
| " %s\n" |
| ");\n", |
| qualrelname, |
| fmtId(tbinfo->attnames[j]), |
| tbinfo->attfdwoptions[j]); |
| } /* end loop over columns */ |
| |
| free(partkeydef); |
| free(ftoptions); |
| free(srvname); |
| } |
| |
| /* |
| * dump properties we only have ALTER TABLE syntax for |
| */ |
| if ((tbinfo->relkind == RELKIND_RELATION || |
| tbinfo->relkind == RELKIND_PARTITIONED_TABLE || |
| tbinfo->relkind == RELKIND_MATVIEW) && |
| tbinfo->relreplident != REPLICA_IDENTITY_DEFAULT) |
| { |
| if (tbinfo->relreplident == REPLICA_IDENTITY_INDEX) |
| { |
| /* nothing to do, will be set when the index is dumped */ |
| } |
| else if (tbinfo->relreplident == REPLICA_IDENTITY_NOTHING) |
| { |
| appendPQExpBuffer(q, "\nALTER TABLE ONLY %s REPLICA IDENTITY NOTHING;\n", |
| qualrelname); |
| } |
| else if (tbinfo->relreplident == REPLICA_IDENTITY_FULL) |
| { |
| appendPQExpBuffer(q, "\nALTER TABLE ONLY %s REPLICA IDENTITY FULL;\n", |
| qualrelname); |
| } |
| } |
| |
| if (tbinfo->forcerowsec) |
| appendPQExpBuffer(q, "\nALTER TABLE ONLY %s FORCE ROW LEVEL SECURITY;\n", |
| qualrelname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(q, &tbinfo->dobj, |
| reltypename, qrelname, |
| tbinfo->dobj.namespace->dobj.name); |
| |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| { |
| char *tablespace = NULL; |
| char *tableam = NULL; |
| |
| /* |
| * _selectTablespace() relies on tablespace-enabled objects in the |
| * default tablespace to have a tablespace of "" (empty string) versus |
| * non-tablespace-enabled objects to have a tablespace of NULL. |
| * getTables() sets tbinfo->reltablespace to "" for the default |
| * tablespace (not NULL). |
| */ |
| if (RELKIND_HAS_TABLESPACE(tbinfo->relkind)) |
| tablespace = tbinfo->reltablespace; |
| |
| if (RELKIND_HAS_TABLE_AM(tbinfo->relkind)) |
| tableam = tbinfo->amname; |
| |
| ArchiveEntry(fout, tbinfo->dobj.catId, tbinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tbinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .tablespace = tablespace, |
| .tableam = tableam, |
| .owner = tbinfo->rolname, |
| .description = reltypename, |
| .section = tbinfo->postponed_def ? |
| SECTION_POST_DATA : SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| } |
| |
| /* Dump Table Comments */ |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpTableComment(fout, tbinfo, reltypename); |
| |
| /* Dump Table Security Labels */ |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpTableSecLabel(fout, tbinfo, reltypename); |
| |
| /* Dump comments on inlined table constraints */ |
| for (j = 0; j < tbinfo->ncheck; j++) |
| { |
| ConstraintInfo *constr = &(tbinfo->checkexprs[j]); |
| |
| if (constr->separate || !constr->conislocal) |
| continue; |
| |
| if (constr->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpTableConstraintComment(fout, constr); |
| } |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qrelname); |
| free(qualrelname); |
| } |
| |
| /* |
| * dumpTableAttach |
| * write to fout the commands to attach a child partition |
| * |
| * Child partitions are always made by creating them separately |
| * and then using ATTACH PARTITION, rather than using |
| * CREATE TABLE ... PARTITION OF. This is important for preserving |
| * any possible discrepancy in column layout, to allow assigning the |
| * correct tablespace if different, and so that it's possible to restore |
| * a partition without restoring its parent. (You'll get an error from |
| * the ATTACH PARTITION command, but that can be ignored, or skipped |
| * using "pg_restore -L" if you prefer.) The last point motivates |
| * treating ATTACH PARTITION as a completely separate ArchiveEntry |
| * rather than emitting it within the child partition's ArchiveEntry. |
| */ |
| static void |
| dumpTableAttach(Archive *fout, const TableAttachInfo *attachinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PGresult *res; |
| char *partbound; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| |
| if (!fout->is_prepared[PREPQUERY_DUMPTABLEATTACH]) |
| { |
| /* Set up query for partbound details */ |
| appendPQExpBufferStr(q, |
| "PREPARE dumpTableAttach(pg_catalog.oid) AS\n"); |
| |
| appendPQExpBufferStr(q, |
| "SELECT pg_get_expr(c.relpartbound, c.oid) " |
| "FROM pg_class c " |
| "WHERE c.oid = $1"); |
| |
| ExecuteSqlStatement(fout, q->data); |
| |
| fout->is_prepared[PREPQUERY_DUMPTABLEATTACH] = true; |
| } |
| |
| printfPQExpBuffer(q, |
| "EXECUTE dumpTableAttach('%u')", |
| attachinfo->partitionTbl->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, q->data); |
| partbound = PQgetvalue(res, 0, 0); |
| |
| /* Perform ALTER TABLE on the parent */ |
| printfPQExpBuffer(q, |
| "ALTER TABLE ONLY %s ", |
| fmtQualifiedDumpable(attachinfo->parentTbl)); |
| appendPQExpBuffer(q, |
| "ATTACH PARTITION %s %s;\n", |
| fmtQualifiedDumpable(attachinfo->partitionTbl), |
| partbound); |
| |
| /* |
| * There is no point in creating a drop query as the drop is done by table |
| * drop. (If you think to change this, see also _printTocEntry().) |
| * Although this object doesn't really have ownership as such, set the |
| * owner field anyway to ensure that the command is run by the correct |
| * role at restore time. |
| */ |
| ArchiveEntry(fout, attachinfo->dobj.catId, attachinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = attachinfo->dobj.name, |
| .namespace = attachinfo->dobj.namespace->dobj.name, |
| .owner = attachinfo->partitionTbl->rolname, |
| .description = "TABLE ATTACH", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data)); |
| |
| PQclear(res); |
| destroyPQExpBuffer(q); |
| } |
| |
| /* |
| * dumpAttrDef --- dump an attribute's default-value declaration |
| */ |
| static void |
| dumpAttrDef(Archive *fout, const AttrDefInfo *adinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = adinfo->adtable; |
| int adnum = adinfo->adnum; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qualrelname; |
| char *tag; |
| char *foreign; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* Skip if not "separate"; it was dumped in the table's definition */ |
| if (!adinfo->separate) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qualrelname = pg_strdup(fmtQualifiedDumpable(tbinfo)); |
| |
| /* |
| * GPDB: Upstream uses ALTER TABLE ONLY below. If the table |
| * is the parent of a GPDB partitioning hierarchy, the default |
| * constraint must cascade to all children as well. Because of |
| * this, we use ALTER TABLE instead when acting on a partition |
| * parent. |
| */ |
| foreign = tbinfo->relkind == RELKIND_FOREIGN_TABLE ? "FOREIGN " : ""; |
| |
| appendPQExpBuffer(q, |
| "ALTER %sTABLE %s%s ALTER COLUMN %s SET DEFAULT %s;\n", |
| foreign, tbinfo->parparent ? "" : "ONLY ", |
| qualrelname, fmtId(tbinfo->attnames[adnum - 1]), |
| adinfo->adef_expr); |
| |
| appendPQExpBuffer(delq, "ALTER %sTABLE %s ALTER COLUMN %s DROP DEFAULT;\n", |
| foreign, qualrelname, |
| fmtId(tbinfo->attnames[adnum - 1])); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, tbinfo->attnames[adnum - 1]); |
| |
| if (adinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, adinfo->dobj.catId, adinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "DEFAULT", |
| .section = SECTION_PRE_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| free(tag); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qualrelname); |
| } |
| |
| /* |
| * getAttrName: extract the correct name for an attribute |
| * |
| * The array tblInfo->attnames[] only provides names of user attributes; |
| * if a system attribute number is supplied, we have to fake it. |
| * We also do a little bit of bounds checking for safety's sake. |
| */ |
| static const char * |
| getAttrName(int attrnum, const TableInfo *tblInfo) |
| { |
| if (attrnum > 0 && attrnum <= tblInfo->numatts) |
| return tblInfo->attnames[attrnum - 1]; |
| switch (attrnum) |
| { |
| case SelfItemPointerAttributeNumber: |
| return "ctid"; |
| case MinTransactionIdAttributeNumber: |
| return "xmin"; |
| case MinCommandIdAttributeNumber: |
| return "cmin"; |
| case MaxTransactionIdAttributeNumber: |
| return "xmax"; |
| case MaxCommandIdAttributeNumber: |
| return "cmax"; |
| case TableOidAttributeNumber: |
| return "tableoid"; |
| } |
| pg_fatal("invalid column number %d for table \"%s\"", |
| attrnum, tblInfo->dobj.name); |
| return NULL; /* keep compiler quiet */ |
| } |
| |
| /* |
| * dumpIndex |
| * write out to fout a user-defined index |
| */ |
| static void |
| dumpIndex(Archive *fout, const IndxInfo *indxinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = indxinfo->indextable; |
| bool is_constraint = (indxinfo->indexconstraint != 0); |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *qindxname; |
| char *qqindxname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| qindxname = pg_strdup(fmtId(indxinfo->dobj.name)); |
| qqindxname = pg_strdup(fmtQualifiedDumpable(indxinfo)); |
| |
| /* |
| * If there's an associated constraint, don't dump the index per se, but |
| * do dump any comment for it. (This is safe because dependency ordering |
| * will have ensured the constraint is emitted first.) Note that the |
| * emitted comment has to be shown as depending on the constraint, not the |
| * index, in such cases. |
| */ |
| if (!is_constraint) |
| { |
| char *indstatcols = indxinfo->indstatcols; |
| char *indstatvals = indxinfo->indstatvals; |
| char **indstatcolsarray = NULL; |
| char **indstatvalsarray = NULL; |
| int nstatcols = 0; |
| int nstatvals = 0; |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_pg_class_oids(fout, q, indxinfo->dobj.catId.oid, true); |
| |
| /* Plain secondary index */ |
| appendPQExpBuffer(q, "%s;\n", indxinfo->indexdef); |
| |
| /* |
| * Append ALTER TABLE commands as needed to set properties that we |
| * only have ALTER TABLE syntax for. Keep this in sync with the |
| * similar code in dumpConstraint! |
| */ |
| |
| /* If the index is clustered, we need to record that. */ |
| if (indxinfo->indisclustered) |
| { |
| appendPQExpBuffer(q, "\nALTER TABLE %s CLUSTER", |
| fmtQualifiedDumpable(tbinfo)); |
| /* index name is not qualified in this syntax */ |
| appendPQExpBuffer(q, " ON %s;\n", |
| qindxname); |
| } |
| |
| /* |
| * If the index has any statistics on some of its columns, generate |
| * the associated ALTER INDEX queries. |
| */ |
| if (strlen(indstatcols) != 0 || strlen(indstatvals) != 0) |
| { |
| int j; |
| |
| if (!parsePGArray(indstatcols, &indstatcolsarray, &nstatcols)) |
| pg_fatal("could not parse index statistic columns"); |
| if (!parsePGArray(indstatvals, &indstatvalsarray, &nstatvals)) |
| pg_fatal("could not parse index statistic values"); |
| if (nstatcols != nstatvals) |
| pg_fatal("mismatched number of columns and values for index statistics"); |
| |
| for (j = 0; j < nstatcols; j++) |
| { |
| appendPQExpBuffer(q, "ALTER INDEX %s ", qqindxname); |
| |
| /* |
| * Note that this is a column number, so no quotes should be |
| * used. |
| */ |
| appendPQExpBuffer(q, "ALTER COLUMN %s ", |
| indstatcolsarray[j]); |
| appendPQExpBuffer(q, "SET STATISTICS %s;\n", |
| indstatvalsarray[j]); |
| } |
| } |
| |
| /* Indexes can depend on extensions */ |
| append_depends_on_extension(fout, q, &indxinfo->dobj, |
| "pg_catalog.pg_class", |
| "INDEX", qqindxname); |
| |
| /* If the index defines identity, we need to record that. */ |
| if (indxinfo->indisreplident) |
| { |
| appendPQExpBuffer(q, "\nALTER TABLE ONLY %s REPLICA IDENTITY USING", |
| fmtQualifiedDumpable(tbinfo)); |
| /* index name is not qualified in this syntax */ |
| appendPQExpBuffer(q, " INDEX %s;\n", |
| qindxname); |
| } |
| |
| /* |
| * If this index is a member of a partitioned index, the backend will |
| * not allow us to drop it separately, so don't try. It will go away |
| * automatically when we drop either the index's table or the |
| * partitioned index. (If, in a selective restore with --clean, we |
| * drop neither of those, then this index will not be dropped either. |
| * But that's fine, and even if you think it's not, the backend won't |
| * let us do differently.) |
| */ |
| if (indxinfo->parentidx == 0) |
| appendPQExpBuffer(delq, "DROP INDEX %s;\n", qqindxname); |
| |
| if (indxinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, indxinfo->dobj.catId, indxinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = indxinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .tablespace = indxinfo->tablespace, |
| .owner = tbinfo->rolname, |
| .description = "INDEX", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| free(indstatcolsarray); |
| free(indstatvalsarray); |
| } |
| |
| /* Dump Index Comments */ |
| if (indxinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "INDEX", qindxname, |
| tbinfo->dobj.namespace->dobj.name, |
| tbinfo->rolname, |
| indxinfo->dobj.catId, 0, |
| is_constraint ? indxinfo->indexconstraint : |
| indxinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| free(qindxname); |
| free(qqindxname); |
| } |
| |
| /* |
| * dumpIndexAttach |
| * write out to fout a partitioned-index attachment clause |
| */ |
| static void |
| dumpIndexAttach(Archive *fout, const IndexAttachInfo *attachinfo) |
| { |
| /* Do nothing in data-only dump */ |
| if (fout->dopt->dataOnly) |
| return; |
| |
| if (attachinfo->partitionIdx->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| { |
| PQExpBuffer q = createPQExpBuffer(); |
| |
| appendPQExpBuffer(q, "ALTER INDEX %s ", |
| fmtQualifiedDumpable(attachinfo->parentIdx)); |
| appendPQExpBuffer(q, "ATTACH PARTITION %s;\n", |
| fmtQualifiedDumpable(attachinfo->partitionIdx)); |
| |
| /* |
| * There is no need for a dropStmt since the drop is done implicitly |
| * when we drop either the index's table or the partitioned index. |
| * Moreover, since there's no ALTER INDEX DETACH PARTITION command, |
| * there's no way to do it anyway. (If you think to change this, |
| * consider also what to do with --if-exists.) |
| * |
| * Although this object doesn't really have ownership as such, set the |
| * owner field anyway to ensure that the command is run by the correct |
| * role at restore time. |
| */ |
| ArchiveEntry(fout, attachinfo->dobj.catId, attachinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = attachinfo->dobj.name, |
| .namespace = attachinfo->dobj.namespace->dobj.name, |
| .owner = attachinfo->parentIdx->indextable->rolname, |
| .description = "INDEX ATTACH", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data)); |
| |
| destroyPQExpBuffer(q); |
| } |
| } |
| |
| /* |
| * dumpStatisticsExt |
| * write out to fout an extended statistics object |
| */ |
| static void |
| dumpStatisticsExt(Archive *fout, const StatsExtInfo *statsextinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| PQExpBuffer query; |
| char *qstatsextname; |
| PGresult *res; |
| char *stxdef; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| query = createPQExpBuffer(); |
| |
| qstatsextname = pg_strdup(fmtId(statsextinfo->dobj.name)); |
| |
| appendPQExpBuffer(query, "SELECT " |
| "pg_catalog.pg_get_statisticsobjdef('%u'::pg_catalog.oid)", |
| statsextinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| stxdef = PQgetvalue(res, 0, 0); |
| |
| /* Result of pg_get_statisticsobjdef is complete except for semicolon */ |
| appendPQExpBuffer(q, "%s;\n", stxdef); |
| |
| /* |
| * We only issue an ALTER STATISTICS statement if the stxstattarget entry |
| * for this statistics object is non-negative (i.e. it's not the default |
| * value). |
| */ |
| if (statsextinfo->stattarget >= 0) |
| { |
| appendPQExpBuffer(q, "ALTER STATISTICS %s ", |
| fmtQualifiedDumpable(statsextinfo)); |
| appendPQExpBuffer(q, "SET STATISTICS %d;\n", |
| statsextinfo->stattarget); |
| } |
| |
| appendPQExpBuffer(delq, "DROP STATISTICS %s;\n", |
| fmtQualifiedDumpable(statsextinfo)); |
| |
| if (statsextinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, statsextinfo->dobj.catId, |
| statsextinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = statsextinfo->dobj.name, |
| .namespace = statsextinfo->dobj.namespace->dobj.name, |
| .owner = statsextinfo->rolname, |
| .description = "STATISTICS", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| |
| /* Dump Statistics Comments */ |
| if (statsextinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "STATISTICS", qstatsextname, |
| statsextinfo->dobj.namespace->dobj.name, |
| statsextinfo->rolname, |
| statsextinfo->dobj.catId, 0, |
| statsextinfo->dobj.dumpId); |
| |
| PQclear(res); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| destroyPQExpBuffer(query); |
| free(qstatsextname); |
| } |
| |
| /* |
| * dumpConstraint |
| * write out to fout a user-defined constraint |
| */ |
| static void |
| dumpConstraint(Archive *fout, const ConstraintInfo *coninfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = coninfo->contable; |
| PQExpBuffer q; |
| PQExpBuffer delq; |
| char *tag = NULL; |
| char *foreign; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| q = createPQExpBuffer(); |
| delq = createPQExpBuffer(); |
| |
| foreign = tbinfo && |
| tbinfo->relkind == RELKIND_FOREIGN_TABLE ? "FOREIGN " : ""; |
| |
| if (coninfo->contype == 'p' || |
| coninfo->contype == 'u' || |
| coninfo->contype == 'x') |
| { |
| /* Index-related constraint */ |
| IndxInfo *indxinfo; |
| int k; |
| |
| indxinfo = (IndxInfo *) findObjectByDumpId(coninfo->conindex); |
| |
| if (indxinfo == NULL) |
| pg_fatal("missing index for constraint \"%s\"", |
| coninfo->dobj.name); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_set_pg_class_oids(fout, q, indxinfo->dobj.catId.oid, true); |
| |
| appendPQExpBuffer(q, "ALTER %sTABLE ONLY %s\n", foreign, |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(q, " ADD CONSTRAINT %s ", |
| fmtId(coninfo->dobj.name)); |
| |
| if (coninfo->condef) |
| { |
| /* pg_get_constraintdef should have provided everything */ |
| appendPQExpBuffer(q, "%s;\n", coninfo->condef); |
| } |
| else |
| { |
| appendPQExpBufferStr(q, |
| coninfo->contype == 'p' ? "PRIMARY KEY" : "UNIQUE"); |
| |
| /* |
| * PRIMARY KEY constraints should not be using NULLS NOT DISTINCT |
| * indexes. Being able to create this was fixed, but we need to |
| * make the index distinct in order to be able to restore the |
| * dump. |
| */ |
| if (indxinfo->indnullsnotdistinct && coninfo->contype != 'p') |
| appendPQExpBufferStr(q, " NULLS NOT DISTINCT"); |
| appendPQExpBufferStr(q, " ("); |
| for (k = 0; k < indxinfo->indnkeyattrs; k++) |
| { |
| int indkey = (int) indxinfo->indkeys[k]; |
| const char *attname; |
| |
| if (indkey == InvalidAttrNumber) |
| break; |
| attname = getAttrName(indkey, tbinfo); |
| |
| appendPQExpBuffer(q, "%s%s", |
| (k == 0) ? "" : ", ", |
| fmtId(attname)); |
| } |
| |
| if (indxinfo->indnkeyattrs < indxinfo->indnattrs) |
| appendPQExpBufferStr(q, ") INCLUDE ("); |
| |
| for (k = indxinfo->indnkeyattrs; k < indxinfo->indnattrs; k++) |
| { |
| int indkey = (int) indxinfo->indkeys[k]; |
| const char *attname; |
| |
| if (indkey == InvalidAttrNumber) |
| break; |
| attname = getAttrName(indkey, tbinfo); |
| |
| appendPQExpBuffer(q, "%s%s", |
| (k == indxinfo->indnkeyattrs) ? "" : ", ", |
| fmtId(attname)); |
| } |
| |
| appendPQExpBufferChar(q, ')'); |
| |
| if (nonemptyReloptions(indxinfo->indreloptions)) |
| { |
| appendPQExpBufferStr(q, " WITH ("); |
| appendReloptionsArrayAH(q, indxinfo->indreloptions, "", fout); |
| appendPQExpBufferChar(q, ')'); |
| } |
| |
| if (coninfo->condeferrable) |
| { |
| appendPQExpBufferStr(q, " DEFERRABLE"); |
| if (coninfo->condeferred) |
| appendPQExpBufferStr(q, " INITIALLY DEFERRED"); |
| } |
| |
| appendPQExpBufferStr(q, ";\n"); |
| } |
| |
| /* |
| * Append ALTER TABLE commands as needed to set properties that we |
| * only have ALTER TABLE syntax for. Keep this in sync with the |
| * similar code in dumpIndex! |
| */ |
| |
| /* If the index is clustered, we need to record that. */ |
| if (indxinfo->indisclustered) |
| { |
| appendPQExpBuffer(q, "\nALTER TABLE %s CLUSTER", |
| fmtQualifiedDumpable(tbinfo)); |
| /* index name is not qualified in this syntax */ |
| appendPQExpBuffer(q, " ON %s;\n", |
| fmtId(indxinfo->dobj.name)); |
| } |
| |
| /* If the index defines identity, we need to record that. */ |
| if (indxinfo->indisreplident) |
| { |
| appendPQExpBuffer(q, "\nALTER TABLE ONLY %s REPLICA IDENTITY USING", |
| fmtQualifiedDumpable(tbinfo)); |
| /* index name is not qualified in this syntax */ |
| appendPQExpBuffer(q, " INDEX %s;\n", |
| fmtId(indxinfo->dobj.name)); |
| } |
| |
| /* Indexes can depend on extensions */ |
| append_depends_on_extension(fout, q, &indxinfo->dobj, |
| "pg_catalog.pg_class", "INDEX", |
| fmtQualifiedDumpable(indxinfo)); |
| |
| appendPQExpBuffer(delq, "ALTER %sTABLE ONLY %s ", foreign, |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(delq, "DROP CONSTRAINT %s;\n", |
| fmtId(coninfo->dobj.name)); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, coninfo->dobj.name); |
| |
| if (coninfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, coninfo->dobj.catId, coninfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .tablespace = indxinfo->tablespace, |
| .owner = tbinfo->rolname, |
| .description = "CONSTRAINT", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| } |
| else if (coninfo->contype == 'f') |
| { |
| char *only; |
| |
| /* |
| * Foreign keys on partitioned tables are always declared as |
| * inheriting to partitions; for all other cases, emit them as |
| * applying ONLY directly to the named table, because that's how they |
| * work for regular inherited tables. |
| */ |
| only = tbinfo->relkind == RELKIND_PARTITIONED_TABLE ? "" : "ONLY "; |
| |
| /* |
| * XXX Potentially wrap in a 'SET CONSTRAINTS OFF' block so that the |
| * current table data is not processed |
| */ |
| appendPQExpBuffer(q, "ALTER %sTABLE %s%s\n", foreign, |
| only, fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(q, " ADD CONSTRAINT %s %s;\n", |
| fmtId(coninfo->dobj.name), |
| coninfo->condef); |
| |
| appendPQExpBuffer(delq, "ALTER %sTABLE %s%s ", foreign, |
| only, fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(delq, "DROP CONSTRAINT %s;\n", |
| fmtId(coninfo->dobj.name)); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, coninfo->dobj.name); |
| |
| if (coninfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, coninfo->dobj.catId, coninfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "FK CONSTRAINT", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| } |
| else if (coninfo->contype == 'c' && tbinfo) |
| { |
| /* CHECK constraint on a table */ |
| |
| /* Ignore if not to be dumped separately, or if it was inherited */ |
| if (coninfo->separate && coninfo->conislocal) |
| { |
| /* not ONLY since we want it to propagate to children */ |
| appendPQExpBuffer(q, "ALTER %sTABLE %s\n", foreign, |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(q, " ADD CONSTRAINT %s %s;\n", |
| fmtId(coninfo->dobj.name), |
| coninfo->condef); |
| |
| appendPQExpBuffer(delq, "ALTER %sTABLE %s ", foreign, |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(delq, "DROP CONSTRAINT %s;\n", |
| fmtId(coninfo->dobj.name)); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, coninfo->dobj.name); |
| |
| if (coninfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, coninfo->dobj.catId, coninfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "CHECK CONSTRAINT", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| } |
| } |
| else if (coninfo->contype == 'c' && tbinfo == NULL) |
| { |
| /* CHECK constraint on a domain */ |
| TypeInfo *tyinfo = coninfo->condomain; |
| |
| /* Ignore if not to be dumped separately */ |
| if (coninfo->separate) |
| { |
| appendPQExpBuffer(q, "ALTER DOMAIN %s\n", |
| fmtQualifiedDumpable(tyinfo)); |
| appendPQExpBuffer(q, " ADD CONSTRAINT %s %s;\n", |
| fmtId(coninfo->dobj.name), |
| coninfo->condef); |
| |
| appendPQExpBuffer(delq, "ALTER DOMAIN %s ", |
| fmtQualifiedDumpable(tyinfo)); |
| appendPQExpBuffer(delq, "DROP CONSTRAINT %s;\n", |
| fmtId(coninfo->dobj.name)); |
| |
| tag = psprintf("%s %s", tyinfo->dobj.name, coninfo->dobj.name); |
| |
| if (coninfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, coninfo->dobj.catId, coninfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tyinfo->dobj.namespace->dobj.name, |
| .owner = tyinfo->rolname, |
| .description = "CHECK CONSTRAINT", |
| .section = SECTION_POST_DATA, |
| .createStmt = q->data, |
| .dropStmt = delq->data)); |
| } |
| } |
| else |
| { |
| pg_fatal("unrecognized constraint type: %c", |
| coninfo->contype); |
| } |
| |
| /* Dump Constraint Comments --- only works for table constraints */ |
| if (tbinfo && coninfo->separate && |
| coninfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpTableConstraintComment(fout, coninfo); |
| |
| free(tag); |
| destroyPQExpBuffer(q); |
| destroyPQExpBuffer(delq); |
| } |
| |
| /* |
| * dumpTableConstraintComment --- dump a constraint's comment if any |
| * |
| * This is split out because we need the function in two different places |
| * depending on whether the constraint is dumped as part of CREATE TABLE |
| * or as a separate ALTER command. |
| */ |
| static void |
| dumpTableConstraintComment(Archive *fout, const ConstraintInfo *coninfo) |
| { |
| TableInfo *tbinfo = coninfo->contable; |
| PQExpBuffer conprefix = createPQExpBuffer(); |
| char *qtabname; |
| |
| qtabname = pg_strdup(fmtId(tbinfo->dobj.name)); |
| |
| appendPQExpBuffer(conprefix, "CONSTRAINT %s ON", |
| fmtId(coninfo->dobj.name)); |
| |
| if (coninfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, conprefix->data, qtabname, |
| tbinfo->dobj.namespace->dobj.name, |
| tbinfo->rolname, |
| coninfo->dobj.catId, 0, |
| coninfo->separate ? coninfo->dobj.dumpId : tbinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(conprefix); |
| free(qtabname); |
| } |
| |
| /* |
| * dumpSequence |
| * write the declaration (not data) of one user-defined sequence |
| */ |
| static void |
| dumpSequence(Archive *fout, const TableInfo *tbinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PGresult *res; |
| char *startv, |
| *incby, |
| *maxv, |
| *minv, |
| *cache, |
| *seqtype; |
| bool cycled; |
| bool is_ascending; |
| int64 default_minv, |
| default_maxv; |
| char bufm[32], |
| bufx[32]; |
| PQExpBuffer query = createPQExpBuffer(); |
| PQExpBuffer delqry = createPQExpBuffer(); |
| char *qseqname; |
| TableInfo *owning_tab = NULL; |
| |
| qseqname = pg_strdup(fmtId(tbinfo->dobj.name)); |
| |
| if (fout->remoteVersion >= 100000) |
| { |
| appendPQExpBuffer(query, |
| "SELECT format_type(seqtypid, NULL), " |
| "seqstart, seqincrement, " |
| "seqmax, seqmin, " |
| "seqcache, seqcycle " |
| "FROM pg_catalog.pg_sequence " |
| "WHERE seqrelid = '%u'::oid", |
| tbinfo->dobj.catId.oid); |
| } |
| else |
| { |
| /* |
| * Before PostgreSQL 10, sequence metadata is in the sequence itself. |
| * |
| * Note: it might seem that 'bigint' potentially needs to be |
| * schema-qualified, but actually that's a keyword. |
| */ |
| appendPQExpBuffer(query, |
| "SELECT 'bigint' AS sequence_type, " |
| "start_value, increment_by, max_value, min_value, " |
| "cache_value, is_cycled FROM %s", |
| fmtQualifiedDumpable(tbinfo)); |
| } |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| pg_fatal(ngettext("query to get data of sequence \"%s\" returned %d row (expected 1)", |
| "query to get data of sequence \"%s\" returned %d rows (expected 1)", |
| PQntuples(res)), |
| tbinfo->dobj.name, PQntuples(res)); |
| |
| seqtype = PQgetvalue(res, 0, 0); |
| startv = PQgetvalue(res, 0, 1); |
| incby = PQgetvalue(res, 0, 2); |
| maxv = PQgetvalue(res, 0, 3); |
| minv = PQgetvalue(res, 0, 4); |
| cache = PQgetvalue(res, 0, 5); |
| cycled = (strcmp(PQgetvalue(res, 0, 6), "t") == 0); |
| |
| /* Calculate default limits for a sequence of this type */ |
| is_ascending = (incby[0] != '-'); |
| if (strcmp(seqtype, "smallint") == 0) |
| { |
| default_minv = is_ascending ? 1 : PG_INT16_MIN; |
| default_maxv = is_ascending ? PG_INT16_MAX : -1; |
| } |
| else if (strcmp(seqtype, "integer") == 0) |
| { |
| default_minv = is_ascending ? 1 : PG_INT32_MIN; |
| default_maxv = is_ascending ? PG_INT32_MAX : -1; |
| } |
| else if (strcmp(seqtype, "bigint") == 0) |
| { |
| default_minv = is_ascending ? 1 : PG_INT64_MIN; |
| default_maxv = is_ascending ? PG_INT64_MAX : -1; |
| } |
| else |
| { |
| pg_fatal("unrecognized sequence type: %s", seqtype); |
| default_minv = default_maxv = 0; /* keep compiler quiet */ |
| } |
| |
| /* |
| * 64-bit strtol() isn't very portable, so convert the limits to strings |
| * and compare that way. |
| */ |
| snprintf(bufm, sizeof(bufm), INT64_FORMAT, default_minv); |
| snprintf(bufx, sizeof(bufx), INT64_FORMAT, default_maxv); |
| |
| /* Don't print minv/maxv if they match the respective default limit */ |
| if (strcmp(minv, bufm) == 0) |
| minv = NULL; |
| if (strcmp(maxv, bufx) == 0) |
| maxv = NULL; |
| |
| /* |
| * Identity sequences are not to be dropped separately. |
| */ |
| if (!tbinfo->is_identity_sequence) |
| { |
| appendPQExpBuffer(delqry, "DROP SEQUENCE %s;\n", |
| fmtQualifiedDumpable(tbinfo)); |
| } |
| |
| resetPQExpBuffer(query); |
| |
| if (dopt->binary_upgrade) |
| { |
| binary_upgrade_set_pg_class_oids(fout, query, |
| tbinfo->dobj.catId.oid, false); |
| |
| /* |
| * In older PG versions a sequence will have a pg_type entry, but v14 |
| * and up don't use that, so don't attempt to preserve the type OID. |
| */ |
| } |
| |
| if (tbinfo->is_identity_sequence) |
| { |
| owning_tab = findTableByOid(tbinfo->owning_tab); |
| |
| appendPQExpBuffer(query, |
| "ALTER TABLE %s ", |
| fmtQualifiedDumpable(owning_tab)); |
| appendPQExpBuffer(query, |
| "ALTER COLUMN %s ADD GENERATED ", |
| fmtId(owning_tab->attnames[tbinfo->owning_col - 1])); |
| if (owning_tab->attidentity[tbinfo->owning_col - 1] == ATTRIBUTE_IDENTITY_ALWAYS) |
| appendPQExpBufferStr(query, "ALWAYS"); |
| else if (owning_tab->attidentity[tbinfo->owning_col - 1] == ATTRIBUTE_IDENTITY_BY_DEFAULT) |
| appendPQExpBufferStr(query, "BY DEFAULT"); |
| appendPQExpBuffer(query, " AS IDENTITY (\n SEQUENCE NAME %s\n", |
| fmtQualifiedDumpable(tbinfo)); |
| |
| /* |
| * Emit persistence option only if it's different from the owning |
| * table's. This avoids using this new syntax unnecessarily. |
| */ |
| if (tbinfo->relpersistence != owning_tab->relpersistence) |
| appendPQExpBuffer(query, " %s\n", |
| tbinfo->relpersistence == RELPERSISTENCE_UNLOGGED ? |
| "UNLOGGED" : "LOGGED"); |
| } |
| else |
| { |
| appendPQExpBuffer(query, |
| "CREATE %sSEQUENCE %s\n", |
| tbinfo->relpersistence == RELPERSISTENCE_UNLOGGED ? |
| "UNLOGGED " : "", |
| fmtQualifiedDumpable(tbinfo)); |
| |
| if (strcmp(seqtype, "bigint") != 0) |
| appendPQExpBuffer(query, " AS %s\n", seqtype); |
| } |
| |
| appendPQExpBuffer(query, " START WITH %s\n", startv); |
| |
| appendPQExpBuffer(query, " INCREMENT BY %s\n", incby); |
| |
| if (minv) |
| appendPQExpBuffer(query, " MINVALUE %s\n", minv); |
| else |
| appendPQExpBufferStr(query, " NO MINVALUE\n"); |
| |
| if (maxv) |
| appendPQExpBuffer(query, " MAXVALUE %s\n", maxv); |
| else |
| appendPQExpBufferStr(query, " NO MAXVALUE\n"); |
| |
| appendPQExpBuffer(query, |
| " CACHE %s%s", |
| cache, (cycled ? "\n CYCLE" : "")); |
| |
| if (tbinfo->is_identity_sequence) |
| appendPQExpBufferStr(query, "\n);\n"); |
| else |
| appendPQExpBufferStr(query, ";\n"); |
| |
| /* binary_upgrade: no need to clear TOAST table oid */ |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(query, &tbinfo->dobj, |
| "SEQUENCE", qseqname, |
| tbinfo->dobj.namespace->dobj.name); |
| |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tbinfo->dobj.catId, tbinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tbinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "SEQUENCE", |
| .section = SECTION_PRE_DATA, |
| .createStmt = query->data, |
| .dropStmt = delqry->data)); |
| |
| /* |
| * If the sequence is owned by a table column, emit the ALTER for it as a |
| * separate TOC entry immediately following the sequence's own entry. It's |
| * OK to do this rather than using full sorting logic, because the |
| * dependency that tells us it's owned will have forced the table to be |
| * created first. We can't just include the ALTER in the TOC entry |
| * because it will fail if we haven't reassigned the sequence owner to |
| * match the table's owner. |
| * |
| * We need not schema-qualify the table reference because both sequence |
| * and table must be in the same schema. |
| */ |
| if (OidIsValid(tbinfo->owning_tab) && !tbinfo->is_identity_sequence) |
| { |
| owning_tab = findTableByOid(tbinfo->owning_tab); |
| |
| if (owning_tab == NULL) |
| pg_fatal("failed sanity check, parent table with OID %u of sequence with OID %u not found", |
| tbinfo->owning_tab, tbinfo->dobj.catId.oid); |
| |
| if (owning_tab->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| { |
| resetPQExpBuffer(query); |
| appendPQExpBuffer(query, "ALTER SEQUENCE %s", |
| fmtQualifiedDumpable(tbinfo)); |
| appendPQExpBuffer(query, " OWNED BY %s", |
| fmtQualifiedDumpable(owning_tab)); |
| appendPQExpBuffer(query, ".%s;\n", |
| fmtId(owning_tab->attnames[tbinfo->owning_col - 1])); |
| |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tbinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "SEQUENCE OWNED BY", |
| .section = SECTION_PRE_DATA, |
| .createStmt = query->data, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1)); |
| } |
| } |
| |
| /* Dump Sequence Comments and Security Labels */ |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "SEQUENCE", qseqname, |
| tbinfo->dobj.namespace->dobj.name, tbinfo->rolname, |
| tbinfo->dobj.catId, 0, tbinfo->dobj.dumpId); |
| |
| if (tbinfo->dobj.dump & DUMP_COMPONENT_SECLABEL) |
| dumpSecLabel(fout, "SEQUENCE", qseqname, |
| tbinfo->dobj.namespace->dobj.name, tbinfo->rolname, |
| tbinfo->dobj.catId, 0, tbinfo->dobj.dumpId); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(delqry); |
| free(qseqname); |
| } |
| |
| /* |
| * dumpSequenceData |
| * write the data of one user-defined sequence |
| */ |
| static void |
| dumpSequenceData(Archive *fout, const TableDataInfo *tdinfo) |
| { |
| TableInfo *tbinfo = tdinfo->tdtable; |
| PGresult *res; |
| char *last; |
| bool called; |
| PQExpBuffer query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, |
| "SELECT last_value, is_called FROM %s", |
| fmtQualifiedDumpable(tbinfo)); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| pg_fatal(ngettext("query to get data of sequence \"%s\" returned %d row (expected 1)", |
| "query to get data of sequence \"%s\" returned %d rows (expected 1)", |
| PQntuples(res)), |
| tbinfo->dobj.name, PQntuples(res)); |
| |
| last = PQgetvalue(res, 0, 0); |
| called = (strcmp(PQgetvalue(res, 0, 1), "t") == 0); |
| |
| resetPQExpBuffer(query); |
| appendPQExpBufferStr(query, "SELECT pg_catalog.setval("); |
| appendStringLiteralAH(query, fmtQualifiedDumpable(tbinfo), fout); |
| appendPQExpBuffer(query, ", %s, %s);\n", |
| last, (called ? "true" : "false")); |
| |
| if (tdinfo->dobj.dump & DUMP_COMPONENT_DATA) |
| ArchiveEntry(fout, nilCatalogId, createDumpId(), |
| ARCHIVE_OPTS(.tag = tbinfo->dobj.name, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "SEQUENCE SET", |
| .section = SECTION_DATA, |
| .createStmt = query->data, |
| .deps = &(tbinfo->dobj.dumpId), |
| .nDeps = 1)); |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * dumpTrigger |
| * write the declaration of one user-defined table trigger |
| */ |
| static void |
| dumpTrigger(Archive *fout, const TriggerInfo *tginfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = tginfo->tgtable; |
| PQExpBuffer query; |
| PQExpBuffer delqry; |
| PQExpBuffer trigprefix; |
| PQExpBuffer trigidentity; |
| char *qtabname; |
| char *tgargs; |
| size_t lentgargs; |
| const char *p; |
| int findx; |
| char *tag; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| trigprefix = createPQExpBuffer(); |
| trigidentity = createPQExpBuffer(); |
| |
| qtabname = pg_strdup(fmtId(tbinfo->dobj.name)); |
| |
| appendPQExpBuffer(trigidentity, "%s ", fmtId(tginfo->dobj.name)); |
| appendPQExpBuffer(trigidentity, "ON %s", fmtQualifiedDumpable(tbinfo)); |
| |
| appendPQExpBuffer(delqry, "DROP TRIGGER %s;\n", trigidentity->data); |
| |
| if (tginfo->tgdef) |
| { |
| appendPQExpBuffer(query, "%s;\n", tginfo->tgdef); |
| } |
| else |
| { |
| if (tginfo->tgisconstraint) |
| { |
| appendPQExpBufferStr(query, "CREATE CONSTRAINT TRIGGER "); |
| appendPQExpBufferStr(query, fmtId(tginfo->tgconstrname)); |
| } |
| else |
| { |
| appendPQExpBufferStr(query, "CREATE TRIGGER "); |
| appendPQExpBufferStr(query, fmtId(tginfo->dobj.name)); |
| } |
| appendPQExpBufferStr(query, "\n "); |
| |
| /* Trigger type */ |
| if (TRIGGER_FOR_BEFORE(tginfo->tgtype)) |
| appendPQExpBufferStr(query, "BEFORE"); |
| else if (TRIGGER_FOR_AFTER(tginfo->tgtype)) |
| appendPQExpBufferStr(query, "AFTER"); |
| else if (TRIGGER_FOR_INSTEAD(tginfo->tgtype)) |
| appendPQExpBufferStr(query, "INSTEAD OF"); |
| else |
| pg_fatal("unexpected tgtype value: %d", tginfo->tgtype); |
| |
| findx = 0; |
| if (TRIGGER_FOR_INSERT(tginfo->tgtype)) |
| { |
| appendPQExpBufferStr(query, " INSERT"); |
| findx++; |
| } |
| if (TRIGGER_FOR_DELETE(tginfo->tgtype)) |
| { |
| if (findx > 0) |
| appendPQExpBufferStr(query, " OR DELETE"); |
| else |
| appendPQExpBufferStr(query, " DELETE"); |
| findx++; |
| } |
| if (TRIGGER_FOR_UPDATE(tginfo->tgtype)) |
| { |
| if (findx > 0) |
| appendPQExpBufferStr(query, " OR UPDATE"); |
| else |
| appendPQExpBufferStr(query, " UPDATE"); |
| findx++; |
| } |
| if (TRIGGER_FOR_TRUNCATE(tginfo->tgtype)) |
| { |
| if (findx > 0) |
| appendPQExpBufferStr(query, " OR TRUNCATE"); |
| else |
| appendPQExpBufferStr(query, " TRUNCATE"); |
| findx++; |
| } |
| appendPQExpBuffer(query, " ON %s\n", |
| fmtQualifiedDumpable(tbinfo)); |
| |
| if (tginfo->tgisconstraint) |
| { |
| if (OidIsValid(tginfo->tgconstrrelid)) |
| { |
| /* regclass output is already quoted */ |
| appendPQExpBuffer(query, " FROM %s\n ", |
| tginfo->tgconstrrelname); |
| } |
| if (!tginfo->tgdeferrable) |
| appendPQExpBufferStr(query, "NOT "); |
| appendPQExpBufferStr(query, "DEFERRABLE INITIALLY "); |
| if (tginfo->tginitdeferred) |
| appendPQExpBufferStr(query, "DEFERRED\n"); |
| else |
| appendPQExpBufferStr(query, "IMMEDIATE\n"); |
| } |
| |
| if (TRIGGER_FOR_ROW(tginfo->tgtype)) |
| appendPQExpBufferStr(query, " FOR EACH ROW\n "); |
| else |
| appendPQExpBufferStr(query, " FOR EACH STATEMENT\n "); |
| |
| /* regproc output is already sufficiently quoted */ |
| appendPQExpBuffer(query, "EXECUTE FUNCTION %s(", |
| tginfo->tgfname); |
| |
| tgargs = (char *) PQunescapeBytea((unsigned char *) tginfo->tgargs, |
| &lentgargs); |
| p = tgargs; |
| for (findx = 0; findx < tginfo->tgnargs; findx++) |
| { |
| /* find the embedded null that terminates this trigger argument */ |
| size_t tlen = strlen(p); |
| |
| if (p + tlen >= tgargs + lentgargs) |
| { |
| /* hm, not found before end of bytea value... */ |
| pg_fatal("invalid argument string (%s) for trigger \"%s\" on table \"%s\"", |
| tginfo->tgargs, |
| tginfo->dobj.name, |
| tbinfo->dobj.name); |
| } |
| |
| if (findx > 0) |
| appendPQExpBufferStr(query, ", "); |
| appendStringLiteralAH(query, p, fout); |
| p += tlen + 1; |
| } |
| free(tgargs); |
| appendPQExpBufferStr(query, ");\n"); |
| } |
| |
| /* Triggers can depend on extensions */ |
| append_depends_on_extension(fout, query, &tginfo->dobj, |
| "pg_catalog.pg_trigger", "TRIGGER", |
| trigidentity->data); |
| |
| <<<<<<< HEAD |
| if (tginfo->tgisinternal) |
| { |
| /* |
| * Triggers marked internal only appear here because their 'tgenabled' |
| * flag differs from its parent's. The trigger is created already, so |
| ======= |
| if (tginfo->tgispartition) |
| { |
| Assert(tbinfo->ispartition); |
| |
| /* |
| * Partition triggers only appear here because their 'tgenabled' flag |
| * differs from its parent's. The trigger is created already, so |
| >>>>>>> REL_16_9 |
| * remove the CREATE and replace it with an ALTER. (Clear out the |
| * DROP query too, so that pg_dump --create does not cause errors.) |
| */ |
| resetPQExpBuffer(query); |
| resetPQExpBuffer(delqry); |
| appendPQExpBuffer(query, "\nALTER %sTABLE %s ", |
| tbinfo->relkind == RELKIND_FOREIGN_TABLE ? "FOREIGN " : "", |
| fmtQualifiedDumpable(tbinfo)); |
| switch (tginfo->tgenabled) |
| { |
| case 'f': |
| case 'D': |
| appendPQExpBufferStr(query, "DISABLE"); |
| break; |
| case 't': |
| case 'O': |
| appendPQExpBufferStr(query, "ENABLE"); |
| break; |
| case 'R': |
| appendPQExpBufferStr(query, "ENABLE REPLICA"); |
| break; |
| case 'A': |
| appendPQExpBufferStr(query, "ENABLE ALWAYS"); |
| break; |
| } |
| appendPQExpBuffer(query, " TRIGGER %s;\n", |
| fmtId(tginfo->dobj.name)); |
| } |
| else if (tginfo->tgenabled != 't' && tginfo->tgenabled != 'O') |
| { |
| appendPQExpBuffer(query, "\nALTER %sTABLE %s ", |
| tbinfo->relkind == RELKIND_FOREIGN_TABLE ? "FOREIGN " : "", |
| fmtQualifiedDumpable(tbinfo)); |
| switch (tginfo->tgenabled) |
| { |
| case 'D': |
| case 'f': |
| appendPQExpBufferStr(query, "DISABLE"); |
| break; |
| case 'A': |
| appendPQExpBufferStr(query, "ENABLE ALWAYS"); |
| break; |
| case 'R': |
| appendPQExpBufferStr(query, "ENABLE REPLICA"); |
| break; |
| default: |
| appendPQExpBufferStr(query, "ENABLE"); |
| break; |
| } |
| appendPQExpBuffer(query, " TRIGGER %s;\n", |
| fmtId(tginfo->dobj.name)); |
| } |
| |
| appendPQExpBuffer(trigprefix, "TRIGGER %s ON", |
| fmtId(tginfo->dobj.name)); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, tginfo->dobj.name); |
| |
| if (tginfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, tginfo->dobj.catId, tginfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "TRIGGER", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data, |
| .dropStmt = delqry->data)); |
| |
| if (tginfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, trigprefix->data, qtabname, |
| tbinfo->dobj.namespace->dobj.name, tbinfo->rolname, |
| tginfo->dobj.catId, 0, tginfo->dobj.dumpId); |
| |
| free(tag); |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(delqry); |
| destroyPQExpBuffer(trigprefix); |
| destroyPQExpBuffer(trigidentity); |
| free(qtabname); |
| } |
| |
| /* |
| * dumpEventTrigger |
| * write the declaration of one user-defined event trigger |
| */ |
| static void |
| dumpEventTrigger(Archive *fout, const EventTriggerInfo *evtinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PQExpBuffer delqry; |
| char *qevtname; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| query = createPQExpBuffer(); |
| delqry = createPQExpBuffer(); |
| |
| qevtname = pg_strdup(fmtId(evtinfo->dobj.name)); |
| |
| appendPQExpBufferStr(query, "CREATE EVENT TRIGGER "); |
| appendPQExpBufferStr(query, qevtname); |
| appendPQExpBufferStr(query, " ON "); |
| appendPQExpBufferStr(query, fmtId(evtinfo->evtevent)); |
| |
| if (strcmp("", evtinfo->evttags) != 0) |
| { |
| appendPQExpBufferStr(query, "\n WHEN TAG IN ("); |
| appendPQExpBufferStr(query, evtinfo->evttags); |
| appendPQExpBufferChar(query, ')'); |
| } |
| |
| appendPQExpBufferStr(query, "\n EXECUTE FUNCTION "); |
| appendPQExpBufferStr(query, evtinfo->evtfname); |
| appendPQExpBufferStr(query, "();\n"); |
| |
| if (evtinfo->evtenabled != 'O') |
| { |
| appendPQExpBuffer(query, "\nALTER EVENT TRIGGER %s ", |
| qevtname); |
| switch (evtinfo->evtenabled) |
| { |
| case 'D': |
| appendPQExpBufferStr(query, "DISABLE"); |
| break; |
| case 'A': |
| appendPQExpBufferStr(query, "ENABLE ALWAYS"); |
| break; |
| case 'R': |
| appendPQExpBufferStr(query, "ENABLE REPLICA"); |
| break; |
| default: |
| appendPQExpBufferStr(query, "ENABLE"); |
| break; |
| } |
| appendPQExpBufferStr(query, ";\n"); |
| } |
| |
| appendPQExpBuffer(delqry, "DROP EVENT TRIGGER %s;\n", |
| qevtname); |
| |
| if (dopt->binary_upgrade) |
| binary_upgrade_extension_member(query, &evtinfo->dobj, |
| "EVENT TRIGGER", qevtname, NULL); |
| |
| if (evtinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, evtinfo->dobj.catId, evtinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = evtinfo->dobj.name, |
| .owner = evtinfo->evtowner, |
| .description = "EVENT TRIGGER", |
| .section = SECTION_POST_DATA, |
| .createStmt = query->data, |
| .dropStmt = delqry->data)); |
| |
| if (evtinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, "EVENT TRIGGER", qevtname, |
| NULL, evtinfo->evtowner, |
| evtinfo->dobj.catId, 0, evtinfo->dobj.dumpId); |
| |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(delqry); |
| free(qevtname); |
| } |
| |
| /* |
| * dumpRule |
| * Dump a rule |
| */ |
| static void |
| dumpRule(Archive *fout, const RuleInfo *rinfo) |
| { |
| DumpOptions *dopt = fout->dopt; |
| TableInfo *tbinfo = rinfo->ruletable; |
| bool is_view; |
| PQExpBuffer query; |
| PQExpBuffer cmd; |
| PQExpBuffer delcmd; |
| PQExpBuffer ruleprefix; |
| char *qtabname; |
| PGresult *res; |
| char *tag; |
| |
| /* Do nothing in data-only dump */ |
| if (dopt->dataOnly) |
| return; |
| |
| /* |
| * If it is an ON SELECT rule that is created implicitly by CREATE VIEW, |
| * we do not want to dump it as a separate object. |
| */ |
| if (!rinfo->separate) |
| return; |
| |
| /* |
| * If it's an ON SELECT rule, we want to print it as a view definition, |
| * instead of a rule. |
| */ |
| is_view = (rinfo->ev_type == '1' && rinfo->is_instead); |
| |
| query = createPQExpBuffer(); |
| cmd = createPQExpBuffer(); |
| delcmd = createPQExpBuffer(); |
| ruleprefix = createPQExpBuffer(); |
| |
| qtabname = pg_strdup(fmtId(tbinfo->dobj.name)); |
| |
| if (is_view) |
| { |
| PQExpBuffer result; |
| |
| /* |
| * We need OR REPLACE here because we'll be replacing a dummy view. |
| * Otherwise this should look largely like the regular view dump code. |
| */ |
| appendPQExpBuffer(cmd, "CREATE OR REPLACE VIEW %s", |
| fmtQualifiedDumpable(tbinfo)); |
| if (nonemptyReloptions(tbinfo->reloptions)) |
| { |
| appendPQExpBufferStr(cmd, " WITH ("); |
| appendReloptionsArrayAH(cmd, tbinfo->reloptions, "", fout); |
| appendPQExpBufferChar(cmd, ')'); |
| } |
| result = createViewAsClause(fout, tbinfo); |
| appendPQExpBuffer(cmd, " AS\n%s", result->data); |
| destroyPQExpBuffer(result); |
| if (tbinfo->checkoption != NULL) |
| appendPQExpBuffer(cmd, "\n WITH %s CHECK OPTION", |
| tbinfo->checkoption); |
| appendPQExpBufferStr(cmd, ";\n"); |
| } |
| else |
| { |
| /* In the rule case, just print pg_get_ruledef's result verbatim */ |
| appendPQExpBuffer(query, |
| "SELECT pg_catalog.pg_get_ruledef('%u'::pg_catalog.oid)", |
| rinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| pg_fatal("query to get rule \"%s\" for table \"%s\" failed: wrong number of rows returned", |
| rinfo->dobj.name, tbinfo->dobj.name); |
| |
| printfPQExpBuffer(cmd, "%s\n", PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| } |
| |
| /* |
| * Add the command to alter the rules replication firing semantics if it |
| * differs from the default. |
| */ |
| if (rinfo->ev_enabled != 'O') |
| { |
| appendPQExpBuffer(cmd, "ALTER TABLE %s ", fmtQualifiedDumpable(tbinfo)); |
| switch (rinfo->ev_enabled) |
| { |
| case 'A': |
| appendPQExpBuffer(cmd, "ENABLE ALWAYS RULE %s;\n", |
| fmtId(rinfo->dobj.name)); |
| break; |
| case 'R': |
| appendPQExpBuffer(cmd, "ENABLE REPLICA RULE %s;\n", |
| fmtId(rinfo->dobj.name)); |
| break; |
| case 'D': |
| appendPQExpBuffer(cmd, "DISABLE RULE %s;\n", |
| fmtId(rinfo->dobj.name)); |
| break; |
| } |
| } |
| |
| if (is_view) |
| { |
| /* |
| * We can't DROP a view's ON SELECT rule. Instead, use CREATE OR |
| * REPLACE VIEW to replace the rule with something with minimal |
| * dependencies. |
| */ |
| PQExpBuffer result; |
| |
| appendPQExpBuffer(delcmd, "CREATE OR REPLACE VIEW %s", |
| fmtQualifiedDumpable(tbinfo)); |
| result = createDummyViewAsClause(fout, tbinfo); |
| appendPQExpBuffer(delcmd, " AS\n%s;\n", result->data); |
| destroyPQExpBuffer(result); |
| } |
| else |
| { |
| appendPQExpBuffer(delcmd, "DROP RULE %s ", |
| fmtId(rinfo->dobj.name)); |
| appendPQExpBuffer(delcmd, "ON %s;\n", |
| fmtQualifiedDumpable(tbinfo)); |
| } |
| |
| appendPQExpBuffer(ruleprefix, "RULE %s ON", |
| fmtId(rinfo->dobj.name)); |
| |
| tag = psprintf("%s %s", tbinfo->dobj.name, rinfo->dobj.name); |
| |
| if (rinfo->dobj.dump & DUMP_COMPONENT_DEFINITION) |
| ArchiveEntry(fout, rinfo->dobj.catId, rinfo->dobj.dumpId, |
| ARCHIVE_OPTS(.tag = tag, |
| .namespace = tbinfo->dobj.namespace->dobj.name, |
| .owner = tbinfo->rolname, |
| .description = "RULE", |
| .section = SECTION_POST_DATA, |
| .createStmt = cmd->data, |
| .dropStmt = delcmd->data)); |
| |
| /* Dump rule comments */ |
| if (rinfo->dobj.dump & DUMP_COMPONENT_COMMENT) |
| dumpComment(fout, ruleprefix->data, qtabname, |
| tbinfo->dobj.namespace->dobj.name, |
| tbinfo->rolname, |
| rinfo->dobj.catId, 0, rinfo->dobj.dumpId); |
| |
| free(tag); |
| destroyPQExpBuffer(query); |
| destroyPQExpBuffer(cmd); |
| destroyPQExpBuffer(delcmd); |
| destroyPQExpBuffer(ruleprefix); |
| free(qtabname); |
| } |
| |
| /* |
| * getExtensionMembership --- obtain extension membership data |
| * |
| * We need to identify objects that are extension members as soon as they're |
| * loaded, so that we can correctly determine whether they need to be dumped. |
| * Generally speaking, extension member objects will get marked as *not* to |
| * be dumped, as they will be recreated by the single CREATE EXTENSION |
| * command. However, in binary upgrade mode we still need to dump the members |
| * individually. |
| */ |
| void |
| getExtensionMembership(Archive *fout, ExtensionInfo extinfo[], |
| int numExtensions) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| int ntups, |
| i; |
| int i_classid, |
| i_objid, |
| i_refobjid; |
| ExtensionInfo *ext; |
| |
| /* Nothing to do if no extensions */ |
| if (numExtensions == 0) |
| return; |
| |
| query = createPQExpBuffer(); |
| |
| /* refclassid constraint is redundant but may speed the search */ |
| appendPQExpBufferStr(query, "SELECT " |
| "classid, objid, refobjid " |
| "FROM pg_depend " |
| "WHERE refclassid = 'pg_extension'::regclass " |
| "AND deptype = 'e' " |
| "ORDER BY 3"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_classid = PQfnumber(res, "classid"); |
| i_objid = PQfnumber(res, "objid"); |
| i_refobjid = PQfnumber(res, "refobjid"); |
| |
| /* |
| * Since we ordered the SELECT by referenced ID, we can expect that |
| * multiple entries for the same extension will appear together; this |
| * saves on searches. |
| */ |
| ext = NULL; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| CatalogId objId; |
| Oid extId; |
| |
| objId.tableoid = atooid(PQgetvalue(res, i, i_classid)); |
| objId.oid = atooid(PQgetvalue(res, i, i_objid)); |
| extId = atooid(PQgetvalue(res, i, i_refobjid)); |
| |
| if (ext == NULL || |
| ext->dobj.catId.oid != extId) |
| ext = findExtensionByOid(extId); |
| |
| if (ext == NULL) |
| { |
| /* shouldn't happen */ |
| pg_log_warning("could not find referenced extension %u", extId); |
| continue; |
| } |
| |
| recordExtensionMembership(objId, ext); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * processExtensionTables --- deal with extension configuration tables |
| * |
| * There are two parts to this process: |
| * |
| * 1. Identify and create dump records for extension configuration tables. |
| * |
| * Extensions can mark tables as "configuration", which means that the user |
| * is able and expected to modify those tables after the extension has been |
| * loaded. For these tables, we dump out only the data- the structure is |
| * expected to be handled at CREATE EXTENSION time, including any indexes or |
| * foreign keys, which brings us to- |
| * |
| * 2. Record FK dependencies between configuration tables. |
| * |
| * Due to the FKs being created at CREATE EXTENSION time and therefore before |
| * the data is loaded, we have to work out what the best order for reloading |
| * the data is, to avoid FK violations when the tables are restored. This is |
| * not perfect- we can't handle circular dependencies and if any exist they |
| * will cause an invalid dump to be produced (though at least all of the data |
| * is included for a user to manually restore). This is currently documented |
| * but perhaps we can provide a better solution in the future. |
| */ |
| void |
| processExtensionTables(Archive *fout, ExtensionInfo extinfo[], |
| int numExtensions) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query; |
| PGresult *res; |
| int ntups, |
| i; |
| int i_conrelid, |
| i_confrelid; |
| |
| /* Nothing to do if no extensions */ |
| if (numExtensions == 0) |
| return; |
| |
| /* |
| * Identify extension configuration tables and create TableDataInfo |
| * objects for them, ensuring their data will be dumped even though the |
| * tables themselves won't be. |
| * |
| * Note that we create TableDataInfo objects even in schemaOnly mode, ie, |
| * user data in a configuration table is treated like schema data. This |
| * seems appropriate since system data in a config table would get |
| * reloaded by CREATE EXTENSION. If the extension is not listed in the |
| * list of extensions to be included, none of its data is dumped. |
| */ |
| for (i = 0; i < numExtensions; i++) |
| { |
| ExtensionInfo *curext = &(extinfo[i]); |
| char *extconfig = curext->extconfig; |
| char *extcondition = curext->extcondition; |
| char **extconfigarray = NULL; |
| char **extconditionarray = NULL; |
| int nconfigitems = 0; |
| int nconditionitems = 0; |
| |
| /* |
| * Check if this extension is listed as to include in the dump. If |
| * not, any table data associated with it is discarded. |
| */ |
| if (extension_include_oids.head != NULL && |
| !simple_oid_list_member(&extension_include_oids, |
| curext->dobj.catId.oid)) |
| continue; |
| |
| if (strlen(extconfig) != 0 || strlen(extcondition) != 0) |
| { |
| int j; |
| |
| if (!parsePGArray(extconfig, &extconfigarray, &nconfigitems)) |
| pg_fatal("could not parse %s array", "extconfig"); |
| if (!parsePGArray(extcondition, &extconditionarray, &nconditionitems)) |
| pg_fatal("could not parse %s array", "extcondition"); |
| if (nconfigitems != nconditionitems) |
| pg_fatal("mismatched number of configurations and conditions for extension"); |
| |
| for (j = 0; j < nconfigitems; j++) |
| { |
| TableInfo *configtbl; |
| Oid configtbloid = atooid(extconfigarray[j]); |
| bool dumpobj = |
| curext->dobj.dump & DUMP_COMPONENT_DEFINITION; |
| |
| configtbl = findTableByOid(configtbloid); |
| if (configtbl == NULL) |
| continue; |
| |
| /* |
| * Tables of not-to-be-dumped extensions shouldn't be dumped |
| * unless the table or its schema is explicitly included |
| */ |
| if (!(curext->dobj.dump & DUMP_COMPONENT_DEFINITION)) |
| { |
| /* check table explicitly requested */ |
| if (table_include_oids.head != NULL && |
| simple_oid_list_member(&table_include_oids, |
| configtbloid)) |
| dumpobj = true; |
| |
| /* check table's schema explicitly requested */ |
| if (configtbl->dobj.namespace->dobj.dump & |
| DUMP_COMPONENT_DATA) |
| dumpobj = true; |
| } |
| |
| /* check table excluded by an exclusion switch */ |
| if (table_exclude_oids.head != NULL && |
| simple_oid_list_member(&table_exclude_oids, |
| configtbloid)) |
| dumpobj = false; |
| |
| /* check schema excluded by an exclusion switch */ |
| if (simple_oid_list_member(&schema_exclude_oids, |
| configtbl->dobj.namespace->dobj.catId.oid)) |
| dumpobj = false; |
| |
| if (dumpobj) |
| { |
| makeTableDataInfo(dopt, configtbl); |
| if (configtbl->dataObj != NULL) |
| { |
| if (strlen(extconditionarray[j]) > 0) |
| configtbl->dataObj->filtercond = pg_strdup(extconditionarray[j]); |
| } |
| } |
| } |
| } |
| if (extconfigarray) |
| free(extconfigarray); |
| if (extconditionarray) |
| free(extconditionarray); |
| } |
| |
| /* |
| * Now that all the TableDataInfo objects have been created for all the |
| * extensions, check their FK dependencies and register them to try and |
| * dump the data out in an order that they can be restored in. |
| * |
| * Note that this is not a problem for user tables as their FKs are |
| * recreated after the data has been loaded. |
| */ |
| |
| query = createPQExpBuffer(); |
| |
| printfPQExpBuffer(query, |
| "SELECT conrelid, confrelid " |
| "FROM pg_constraint " |
| "JOIN pg_depend ON (objid = confrelid) " |
| "WHERE contype = 'f' " |
| "AND refclassid = 'pg_extension'::regclass " |
| "AND classid = 'pg_class'::regclass;"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| ntups = PQntuples(res); |
| |
| i_conrelid = PQfnumber(res, "conrelid"); |
| i_confrelid = PQfnumber(res, "confrelid"); |
| |
| /* Now get the dependencies and register them */ |
| for (i = 0; i < ntups; i++) |
| { |
| Oid conrelid, |
| confrelid; |
| TableInfo *reftable, |
| *contable; |
| |
| conrelid = atooid(PQgetvalue(res, i, i_conrelid)); |
| confrelid = atooid(PQgetvalue(res, i, i_confrelid)); |
| contable = findTableByOid(conrelid); |
| reftable = findTableByOid(confrelid); |
| |
| if (reftable == NULL || |
| reftable->dataObj == NULL || |
| contable == NULL || |
| contable->dataObj == NULL) |
| continue; |
| |
| /* |
| * Make referencing TABLE_DATA object depend on the referenced table's |
| * TABLE_DATA object. |
| */ |
| addObjectDependency(&contable->dataObj->dobj, |
| reftable->dataObj->dobj.dumpId); |
| } |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * setExtPartDependency - |
| */ |
| static void |
| setExtPartDependency(TableInfo *tblinfo, int numTables) |
| { |
| int i; |
| |
| for (i = 0; i < numTables; i++) |
| { |
| TableInfo *tbinfo = &(tblinfo[i]); |
| TableInfo *parent; |
| Oid parrelid = tbinfo->parrelid; |
| |
| if (parrelid == 0) |
| continue; |
| |
| parent = findTableByOid(parrelid); |
| if (!parent) |
| { |
| pg_log_warning("parent table (OID %u) of partition \"%s\" (OID %u) not found", |
| parrelid, tbinfo->dobj.name, tbinfo->dobj.catId.oid); |
| exit_nicely(1); |
| } |
| |
| addObjectDependency(&parent->dobj, tbinfo->dobj.dumpId); |
| removeObjectDependency(&tbinfo->dobj, parent->dobj.dumpId); |
| } |
| } |
| |
| /* |
| * getDependencies --- obtain available dependency data |
| */ |
| static void |
| getDependencies(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| int ntups, |
| i; |
| int i_classid, |
| i_objid, |
| i_refclassid, |
| i_refobjid, |
| i_deptype; |
| DumpableObject *dobj, |
| *refdobj; |
| |
| pg_log_info("reading dependency data"); |
| |
| query = createPQExpBuffer(); |
| |
| /* |
| * Messy query to collect the dependency data we need. Note that we |
| * ignore the sub-object column, so that dependencies of or on a column |
| * look the same as dependencies of or on a whole table. |
| * |
| * PIN dependencies aren't interesting, and EXTENSION dependencies were |
| * already processed by getExtensionMembership. |
| */ |
| appendPQExpBufferStr(query, "SELECT " |
| "classid, objid, refclassid, refobjid, deptype " |
| "FROM pg_depend " |
| "WHERE deptype != 'p' AND deptype != 'e'\n"); |
| |
| /* |
| * Since we don't treat pg_amop entries as separate DumpableObjects, we |
| * have to translate their dependencies into dependencies of their parent |
| * opfamily. Ignore internal dependencies though, as those will point to |
| * their parent opclass, which we needn't consider here (and if we did, |
| * it'd just result in circular dependencies). Also, "loose" opfamily |
| * entries will have dependencies on their parent opfamily, which we |
| * should drop since they'd likewise become useless self-dependencies. |
| * (But be sure to keep deps on *other* opfamilies; see amopsortfamily.) |
| */ |
| appendPQExpBufferStr(query, "UNION ALL\n" |
| "SELECT 'pg_opfamily'::regclass AS classid, amopfamily AS objid, refclassid, refobjid, deptype " |
| "FROM pg_depend d, pg_amop o " |
| "WHERE deptype NOT IN ('p', 'e', 'i') AND " |
| "classid = 'pg_amop'::regclass AND objid = o.oid " |
| "AND NOT (refclassid = 'pg_opfamily'::regclass AND amopfamily = refobjid)\n"); |
| |
| /* Likewise for pg_amproc entries */ |
| appendPQExpBufferStr(query, "UNION ALL\n" |
| "SELECT 'pg_opfamily'::regclass AS classid, amprocfamily AS objid, refclassid, refobjid, deptype " |
| "FROM pg_depend d, pg_amproc p " |
| "WHERE deptype NOT IN ('p', 'e', 'i') AND " |
| "classid = 'pg_amproc'::regclass AND objid = p.oid " |
| "AND NOT (refclassid = 'pg_opfamily'::regclass AND amprocfamily = refobjid)\n"); |
| |
| /* Sort the output for efficiency below */ |
| appendPQExpBufferStr(query, "ORDER BY 1,2"); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| ntups = PQntuples(res); |
| |
| i_classid = PQfnumber(res, "classid"); |
| i_objid = PQfnumber(res, "objid"); |
| i_refclassid = PQfnumber(res, "refclassid"); |
| i_refobjid = PQfnumber(res, "refobjid"); |
| i_deptype = PQfnumber(res, "deptype"); |
| |
| /* |
| * Since we ordered the SELECT by referencing ID, we can expect that |
| * multiple entries for the same object will appear together; this saves |
| * on searches. |
| */ |
| dobj = NULL; |
| |
| for (i = 0; i < ntups; i++) |
| { |
| CatalogId objId; |
| CatalogId refobjId; |
| char deptype; |
| |
| objId.tableoid = atooid(PQgetvalue(res, i, i_classid)); |
| objId.oid = atooid(PQgetvalue(res, i, i_objid)); |
| refobjId.tableoid = atooid(PQgetvalue(res, i, i_refclassid)); |
| refobjId.oid = atooid(PQgetvalue(res, i, i_refobjid)); |
| deptype = *(PQgetvalue(res, i, i_deptype)); |
| |
| if (dobj == NULL || |
| dobj->catId.tableoid != objId.tableoid || |
| dobj->catId.oid != objId.oid) |
| dobj = findObjectByCatalogId(objId); |
| |
| /* |
| * Failure to find objects mentioned in pg_depend is not unexpected, |
| * since for example we don't collect info about TOAST tables. |
| */ |
| if (dobj == NULL) |
| { |
| #ifdef NOT_USED |
| pg_log_warning("no referencing object %u %u", |
| objId.tableoid, objId.oid); |
| #endif |
| continue; |
| } |
| |
| refdobj = findObjectByCatalogId(refobjId); |
| |
| if (refdobj == NULL) |
| { |
| #ifdef NOT_USED |
| pg_log_warning("no referenced object %u %u", |
| refobjId.tableoid, refobjId.oid); |
| #endif |
| continue; |
| } |
| |
| /* |
| * For 'x' dependencies, mark the object for later; we still add the |
| * normal dependency, for possible ordering purposes. Currently |
| * pg_dump_sort.c knows to put extensions ahead of all object types |
| * that could possibly depend on them, but this is safer. |
| */ |
| if (deptype == 'x') |
| dobj->depends_on_ext = true; |
| |
| /* |
| * Ordinarily, table rowtypes have implicit dependencies on their |
| * tables. However, for a composite type the implicit dependency goes |
| * the other way in pg_depend; which is the right thing for DROP but |
| * it doesn't produce the dependency ordering we need. So in that one |
| * case, we reverse the direction of the dependency. |
| */ |
| if (deptype == 'i' && |
| dobj->objType == DO_TABLE && |
| refdobj->objType == DO_TYPE) |
| addObjectDependency(refdobj, dobj->dumpId); |
| else |
| /* normal case */ |
| addObjectDependency(dobj, refdobj->dumpId); |
| } |
| |
| PQclear(res); |
| |
| destroyPQExpBuffer(query); |
| } |
| |
| |
| /* |
| * createBoundaryObjects - create dummy DumpableObjects to represent |
| * dump section boundaries. |
| */ |
| static DumpableObject * |
| createBoundaryObjects(void) |
| { |
| DumpableObject *dobjs; |
| |
| dobjs = (DumpableObject *) pg_malloc(2 * sizeof(DumpableObject)); |
| |
| dobjs[0].objType = DO_PRE_DATA_BOUNDARY; |
| dobjs[0].catId = nilCatalogId; |
| AssignDumpId(dobjs + 0); |
| dobjs[0].name = pg_strdup("PRE-DATA BOUNDARY"); |
| |
| dobjs[1].objType = DO_POST_DATA_BOUNDARY; |
| dobjs[1].catId = nilCatalogId; |
| AssignDumpId(dobjs + 1); |
| dobjs[1].name = pg_strdup("POST-DATA BOUNDARY"); |
| |
| return dobjs; |
| } |
| |
| /* |
| * addBoundaryDependencies - add dependencies as needed to enforce the dump |
| * section boundaries. |
| */ |
| static void |
| addBoundaryDependencies(DumpableObject **dobjs, int numObjs, |
| DumpableObject *boundaryObjs) |
| { |
| DumpableObject *preDataBound = boundaryObjs + 0; |
| DumpableObject *postDataBound = boundaryObjs + 1; |
| int i; |
| |
| for (i = 0; i < numObjs; i++) |
| { |
| DumpableObject *dobj = dobjs[i]; |
| |
| /* |
| * The classification of object types here must match the SECTION_xxx |
| * values assigned during subsequent ArchiveEntry calls! |
| */ |
| switch (dobj->objType) |
| { |
| case DO_NAMESPACE: |
| case DO_EXTENSION: |
| case DO_TYPE: |
| case DO_SHELL_TYPE: |
| case DO_FUNC: |
| case DO_AGG: |
| case DO_OPERATOR: |
| case DO_ACCESS_METHOD: |
| case DO_OPCLASS: |
| case DO_OPFAMILY: |
| case DO_COLLATION: |
| case DO_CONVERSION: |
| case DO_TABLE: |
| case DO_TABLE_ATTACH: |
| case DO_ATTRDEF: |
| case DO_PROCLANG: |
| case DO_CAST: |
| case DO_DUMMY_TYPE: |
| case DO_TSPARSER: |
| case DO_TSDICT: |
| case DO_TSTEMPLATE: |
| case DO_TSCONFIG: |
| case DO_FDW: |
| case DO_FOREIGN_SERVER: |
| case DO_TRANSFORM: |
| <<<<<<< HEAD |
| case DO_BLOB: |
| case DO_EXTPROTOCOL: |
| case DO_TYPE_STORAGE_OPTIONS: |
| case DO_BINARY_UPGRADE: |
| ======= |
| case DO_LARGE_OBJECT: |
| >>>>>>> REL_16_9 |
| /* Pre-data objects: must come before the pre-data boundary */ |
| addObjectDependency(preDataBound, dobj->dumpId); |
| break; |
| case DO_TABLE_DATA: |
| case DO_SEQUENCE_SET: |
| case DO_LARGE_OBJECT_DATA: |
| /* Data objects: must come between the boundaries */ |
| addObjectDependency(dobj, preDataBound->dumpId); |
| addObjectDependency(postDataBound, dobj->dumpId); |
| break; |
| case DO_INDEX: |
| case DO_INDEX_ATTACH: |
| case DO_STATSEXT: |
| case DO_REFRESH_MATVIEW: |
| case DO_TRIGGER: |
| case DO_EVENT_TRIGGER: |
| case DO_DEFAULT_ACL: |
| case DO_POLICY: |
| case DO_PUBLICATION: |
| case DO_PUBLICATION_REL: |
| case DO_PUBLICATION_TABLE_IN_SCHEMA: |
| case DO_SUBSCRIPTION: |
| /* Post-data objects: must come after the post-data boundary */ |
| addObjectDependency(dobj, postDataBound->dumpId); |
| break; |
| case DO_RULE: |
| /* Rules are post-data, but only if dumped separately */ |
| if (((RuleInfo *) dobj)->separate) |
| addObjectDependency(dobj, postDataBound->dumpId); |
| break; |
| case DO_CONSTRAINT: |
| case DO_FK_CONSTRAINT: |
| /* Constraints are post-data, but only if dumped separately */ |
| if (((ConstraintInfo *) dobj)->separate) |
| addObjectDependency(dobj, postDataBound->dumpId); |
| break; |
| case DO_PRE_DATA_BOUNDARY: |
| /* nothing to do */ |
| break; |
| case DO_POST_DATA_BOUNDARY: |
| /* must come after the pre-data boundary */ |
| addObjectDependency(dobj, preDataBound->dumpId); |
| break; |
| } |
| } |
| } |
| |
| |
| /* |
| * BuildArchiveDependencies - create dependency data for archive TOC entries |
| * |
| * The raw dependency data obtained by getDependencies() is not terribly |
| * useful in an archive dump, because in many cases there are dependency |
| * chains linking through objects that don't appear explicitly in the dump. |
| * For example, a view will depend on its _RETURN rule while the _RETURN rule |
| * will depend on other objects --- but the rule will not appear as a separate |
| * object in the dump. We need to adjust the view's dependencies to include |
| * whatever the rule depends on that is included in the dump. |
| * |
| * Just to make things more complicated, there are also "special" dependencies |
| * such as the dependency of a TABLE DATA item on its TABLE, which we must |
| * not rearrange because pg_restore knows that TABLE DATA only depends on |
| * its table. In these cases we must leave the dependencies strictly as-is |
| * even if they refer to not-to-be-dumped objects. |
| * |
| * To handle this, the convention is that "special" dependencies are created |
| * during ArchiveEntry calls, and an archive TOC item that has any such |
| * entries will not be touched here. Otherwise, we recursively search the |
| * DumpableObject data structures to build the correct dependencies for each |
| * archive TOC item. |
| */ |
| static void |
| BuildArchiveDependencies(Archive *fout) |
| { |
| ArchiveHandle *AH = (ArchiveHandle *) fout; |
| TocEntry *te; |
| |
| /* Scan all TOC entries in the archive */ |
| for (te = AH->toc->next; te != AH->toc; te = te->next) |
| { |
| DumpableObject *dobj; |
| DumpId *dependencies; |
| int nDeps; |
| int allocDeps; |
| |
| /* No need to process entries that will not be dumped */ |
| if (te->reqs == 0) |
| continue; |
| /* Ignore entries that already have "special" dependencies */ |
| if (te->nDeps > 0) |
| continue; |
| /* Otherwise, look up the item's original DumpableObject, if any */ |
| dobj = findObjectByDumpId(te->dumpId); |
| if (dobj == NULL) |
| continue; |
| /* No work if it has no dependencies */ |
| if (dobj->nDeps <= 0) |
| continue; |
| /* Set up work array */ |
| allocDeps = 64; |
| dependencies = (DumpId *) pg_malloc(allocDeps * sizeof(DumpId)); |
| nDeps = 0; |
| /* Recursively find all dumpable dependencies */ |
| findDumpableDependencies(AH, dobj, |
| &dependencies, &nDeps, &allocDeps); |
| /* And save 'em ... */ |
| if (nDeps > 0) |
| { |
| dependencies = (DumpId *) pg_realloc(dependencies, |
| nDeps * sizeof(DumpId)); |
| te->dependencies = dependencies; |
| te->nDeps = nDeps; |
| } |
| else |
| free(dependencies); |
| } |
| } |
| |
| /* Recursive search subroutine for BuildArchiveDependencies */ |
| static void |
| findDumpableDependencies(ArchiveHandle *AH, const DumpableObject *dobj, |
| DumpId **dependencies, int *nDeps, int *allocDeps) |
| { |
| int i; |
| |
| /* |
| * Ignore section boundary objects: if we search through them, we'll |
| * report lots of bogus dependencies. |
| */ |
| if (dobj->objType == DO_PRE_DATA_BOUNDARY || |
| dobj->objType == DO_POST_DATA_BOUNDARY) |
| return; |
| |
| for (i = 0; i < dobj->nDeps; i++) |
| { |
| DumpId depid = dobj->dependencies[i]; |
| |
| if (TocIDRequired(AH, depid) != 0) |
| { |
| /* Object will be dumped, so just reference it as a dependency */ |
| if (*nDeps >= *allocDeps) |
| { |
| *allocDeps *= 2; |
| *dependencies = (DumpId *) pg_realloc(*dependencies, |
| *allocDeps * sizeof(DumpId)); |
| } |
| (*dependencies)[*nDeps] = depid; |
| (*nDeps)++; |
| } |
| else |
| { |
| /* |
| * Object will not be dumped, so recursively consider its deps. We |
| * rely on the assumption that sortDumpableObjects already broke |
| * any dependency loops, else we might recurse infinitely. |
| */ |
| DumpableObject *otherdobj = findObjectByDumpId(depid); |
| |
| if (otherdobj) |
| findDumpableDependencies(AH, otherdobj, |
| dependencies, nDeps, allocDeps); |
| } |
| } |
| } |
| |
| /* |
| * isGPbackend - returns true if the connected backend is a GreenPlum DB backend. |
| */ |
| static bool |
| testGPbackend(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| bool isGPbackend; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "SELECT current_setting('gp_role');"); |
| ArchiveHandle *AH = (ArchiveHandle *) fout; |
| res = PQexec(AH->connection, query->data); |
| |
| isGPbackend = (PQresultStatus(res) == PGRES_TUPLES_OK); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| return isGPbackend; |
| } |
| |
| /* |
| * testPartitioningSupport - tests whether or not the current GP |
| * database includes support for partitioning. |
| */ |
| static bool |
| testPartitioningSupport(Archive *fout) |
| { |
| PQExpBuffer query; |
| PGresult *res; |
| bool isSupported; |
| |
| query = createPQExpBuffer(); |
| |
| appendPQExpBuffer(query, "SELECT 1 FROM pg_class WHERE relname = 'pg_partition' and relnamespace = 11;"); |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| isSupported = (PQntuples(res) == 1); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| return isSupported; |
| } |
| /* |
| * addSchedule |
| * |
| * find the SCHEDULE of the job in pg_task with dynamic table |
| * and append the SCHEDULE clause to the passed in dump buffer (q). |
| */ |
| static void |
| addSchedule(Archive *fout, PQExpBuffer q, const TableInfo *tbinfo) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| char *dby; |
| |
| appendPQExpBuffer(query, |
| "SELECT pg_catalog.pg_get_dynamic_table_schedule(%u)", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| dby = PQgetvalue(res, 0, 0); |
| if (strcmp(dby, "") != 0) |
| appendPQExpBuffer(q, " SCHEDULE \'%s\'", PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * addDistributedBy |
| * |
| * find the distribution policy of the passed in relation and append the |
| * DISTRIBUTED BY clause to the passed in dump buffer (q). |
| */ |
| static void |
| addDistributedBy(Archive *fout, PQExpBuffer q, const TableInfo *tbinfo, int actual_atts) |
| { |
| if (isGPDB6000OrLater(fout)) |
| { |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| char *dby; |
| |
| appendPQExpBuffer(query, |
| "SELECT pg_catalog.pg_get_table_distributedby(%u)", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| dby = PQgetvalue(res, 0, 0); |
| if (strcmp(dby, "") != 0) |
| appendPQExpBuffer(q, " %s", PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| else |
| addDistributedByOld(fout, q, tbinfo, actual_atts); |
| } |
| |
| /* |
| * This is used with GPDB 5 and older, where pg_get_table_distributedby() |
| * backend function is not available. |
| */ |
| static void |
| addDistributedByOld(Archive *fout, PQExpBuffer q, const TableInfo *tbinfo, int actual_atts) |
| { |
| DumpOptions *dopt = fout->dopt; |
| PQExpBuffer query = createPQExpBuffer(); |
| PGresult *res; |
| char *policydef; |
| char *policycol; |
| |
| appendPQExpBuffer(query, |
| "SELECT attrnums FROM pg_catalog.gp_distribution_policy as p " |
| "WHERE p.localoid = %u", |
| tbinfo->dobj.catId.oid); |
| |
| res = ExecuteSqlQuery(fout, query->data, PGRES_TUPLES_OK); |
| |
| if (PQntuples(res) != 1) |
| { |
| /* |
| * There is no entry in the policy table for this table. Report an |
| * error unless this is a zero attribute table (actual_atts == 0). |
| * |
| * In binary_upgrade mode, we run directly against segments, and there |
| * are no gp_distribution_policy rows in segments. |
| */ |
| if (PQntuples(res) < 1 && actual_atts > 0 && !dopt->binary_upgrade) |
| { |
| /* if this is a catalog table we allow dumping it, skip the error */ |
| if (strncmp(tbinfo->dobj.namespace->dobj.name, "pg_", 3) != 0) |
| { |
| pg_log_warning("query to obtain distribution policy of table \"%s\" returned no data", |
| tbinfo->dobj.name); |
| exit_nicely(1); |
| } |
| } |
| |
| /* |
| * There is more than 1 entry in the policy table for this table. |
| * Report an error. |
| */ |
| if (PQntuples(res) > 1) |
| { |
| pg_log_warning("query to obtain distribution policy of table \"%s\" returned more than one policy", |
| tbinfo->dobj.name); |
| exit_nicely(1); |
| } |
| } |
| else |
| { |
| /* |
| * There is exactly 1 policy entry for this table (either a concrete |
| * one or NULL). |
| */ |
| policydef = PQgetvalue(res, 0, 0); |
| |
| if (strlen(policydef) > 0) |
| { |
| bool isfirst = true; |
| char *separator; |
| |
| appendPQExpBuffer(q, " DISTRIBUTED BY ("); |
| |
| /* policy indicates one or more columns to distribute on */ |
| policydef[strlen(policydef) - 1] = '\0'; |
| policydef++; |
| separator = ","; |
| |
| while ((policycol = nextToken(&policydef, ",")) != NULL) |
| { |
| if (!isfirst) |
| appendPQExpBuffer(q, ", "); |
| isfirst = false; |
| |
| appendPQExpBuffer(q, "%s", |
| fmtId(tbinfo->attnames[atoi(policycol) - 1])); |
| } |
| appendPQExpBufferChar(q, ')'); |
| } |
| else |
| { |
| /* policy has an empty policy - distribute randomly */ |
| appendPQExpBufferStr(q, " DISTRIBUTED RANDOMLY"); |
| } |
| } |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| } |
| |
| /* |
| * getFormattedTypeName - retrieve a nicely-formatted type name for the |
| * given type OID. |
| * |
| * This does not guarantee to schema-qualify the output, so it should not |
| * be used to create the target object name for CREATE or ALTER commands. |
| * |
| * Note that the result is cached and must not be freed by the caller. |
| */ |
| static const char * |
| getFormattedTypeName(Archive *fout, Oid oid, OidOptions opts) |
| { |
| TypeInfo *typeInfo; |
| char *result; |
| PQExpBuffer query; |
| PGresult *res; |
| |
| if (oid == 0) |
| { |
| if ((opts & zeroAsStar) != 0) |
| return "*"; |
| else if ((opts & zeroAsNone) != 0) |
| return "NONE"; |
| } |
| |
| /* see if we have the result cached in the type's TypeInfo record */ |
| typeInfo = findTypeByOid(oid); |
| if (typeInfo && typeInfo->ftypname) |
| return typeInfo->ftypname; |
| |
| query = createPQExpBuffer(); |
| appendPQExpBuffer(query, "SELECT pg_catalog.format_type('%u'::pg_catalog.oid, NULL)", |
| oid); |
| |
| res = ExecuteSqlQueryForSingleRow(fout, query->data); |
| |
| /* result of format_type is already quoted */ |
| result = pg_strdup(PQgetvalue(res, 0, 0)); |
| |
| PQclear(res); |
| destroyPQExpBuffer(query); |
| |
| /* |
| * Cache the result for re-use in later requests, if possible. If we |
| * don't have a TypeInfo for the type, the string will be leaked once the |
| * caller is done with it ... but that case really should not happen, so |
| * leaking if it does seems acceptable. |
| */ |
| if (typeInfo) |
| typeInfo->ftypname = result; |
| |
| return result; |
| } |
| |
| /* |
| * Return a column list clause for the given relation. |
| * |
| * Special case: if there are no undropped columns in the relation, return |
| * "", not an invalid "()" column list. |
| */ |
| static const char * |
| fmtCopyColumnList(const TableInfo *ti, PQExpBuffer buffer) |
| { |
| int numatts = ti->numatts; |
| char **attnames = ti->attnames; |
| bool *attisdropped = ti->attisdropped; |
| char *attgenerated = ti->attgenerated; |
| bool needComma; |
| int i; |
| |
| appendPQExpBufferChar(buffer, '('); |
| needComma = false; |
| for (i = 0; i < numatts; i++) |
| { |
| if (attisdropped[i]) |
| continue; |
| if (attgenerated[i]) |
| continue; |
| if (needComma) |
| appendPQExpBufferStr(buffer, ", "); |
| appendPQExpBufferStr(buffer, fmtId(attnames[i])); |
| needComma = true; |
| } |
| |
| if (!needComma) |
| return ""; /* no undropped columns */ |
| |
| appendPQExpBufferChar(buffer, ')'); |
| return buffer->data; |
| } |
| |
| /* |
| * Check if a reloptions array is nonempty. |
| */ |
| static bool |
| nonemptyReloptions(const char *reloptions) |
| { |
| /* Don't want to print it if it's just "{}" */ |
| return (reloptions != NULL && strlen(reloptions) > 2); |
| } |
| |
| /* |
| * Format a reloptions array and append it to the given buffer. |
| * |
| * "prefix" is prepended to the option names; typically it's "" or "toast.". |
| */ |
| static void |
| appendReloptionsArrayAH(PQExpBuffer buffer, const char *reloptions, |
| const char *prefix, Archive *fout) |
| { |
| bool res; |
| |
| res = appendReloptionsArray(buffer, reloptions, prefix, fout->encoding, |
| fout->std_strings); |
| if (!res) |
| pg_log_warning("could not parse %s array", "reloptions"); |
| } |
| |
| /* START MPP ADDITION */ |
| /* |
| * Get next token from string *stringp, where tokens are possibly-empty |
| * strings separated by characters from delim. |
| * |
| * Writes NULs into the string at *stringp to end tokens. |
| * delim need not remain constant from call to call. |
| * On return, *stringp points past the last NUL written (if there might |
| * be further tokens), or is NULL (if there are definitely no more tokens). |
| * |
| * If *stringp is NULL, strsep returns NULL. |
| */ |
| static char * |
| nextToken(register char **stringp, register const char *delim) |
| { |
| register char *s; |
| register const char *spanp; |
| register int c, |
| sc; |
| char *tok; |
| |
| if ((s = *stringp) == NULL) |
| return (NULL); |
| for (tok = s;;) |
| { |
| c = *s++; |
| spanp = delim; |
| do |
| { |
| if ((sc = *spanp++) == c) |
| { |
| if (c == 0) |
| s = NULL; |
| else |
| s[-1] = 0; |
| *stringp = s; |
| return (tok); |
| } |
| } while (sc != 0); |
| } |
| /* NOTREACHED */ |
| } |
| |
| /* END MPP ADDITION */ |