]> granicus.if.org Git - postgresql/commitdiff
Final pgindent + perltidy run for 9.6.
authorTom Lane <tgl@sss.pgh.pa.us>
Mon, 15 Aug 2016 17:42:51 +0000 (13:42 -0400)
committerTom Lane <tgl@sss.pgh.pa.us>
Mon, 15 Aug 2016 17:42:51 +0000 (13:42 -0400)
26 files changed:
src/backend/access/brin/brin_validate.c
src/backend/access/gin/ginvalidate.c
src/backend/access/gist/gistvalidate.c
src/backend/access/hash/hashvalidate.c
src/backend/access/heap/heapam.c
src/backend/access/heap/tuptoaster.c
src/backend/commands/variable.c
src/backend/replication/libpqwalreceiver/libpqwalreceiver.c
src/backend/replication/syncrep.c
src/backend/tsearch/spell.c
src/backend/utils/adt/jsonb.c
src/backend/utils/adt/tsquery.c
src/backend/utils/adt/tsquery_cleanup.c
src/backend/utils/sort/tuplesort.c
src/bin/pg_dump/pg_dump.c
src/bin/pg_dump/pg_dumpall.c
src/bin/pg_dump/t/002_pg_dump.pl
src/include/commands/async.h
src/include/tsearch/ts_type.h
src/pl/plpython/plpy_plpymodule.c
src/test/modules/test_pg_dump/t/001_base.pl
src/test/perl/PostgresNode.pm
src/test/recovery/t/003_recovery_targets.pl
src/tools/msvc/Install.pm
src/tools/msvc/vcregress.pl
src/tools/pgindent/typedefs.list

index 2b3a2cb0abc803bdd28e3e2513cc01e116edbec5..1f1011e0acfc609bd3d52db0b828f16353243424 100644 (file)
@@ -264,8 +264,8 @@ brinvalidate(Oid opclassoid)
                        continue;                       /* got it */
                ereport(INFO,
                                (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
-                                errmsg("brin operator class \"%s\" is missing support function %d",
-                                               opclassname, i)));
+                 errmsg("brin operator class \"%s\" is missing support function %d",
+                                opclassname, i)));
                result = false;
        }
 
index 7518ededd95e1a006ad6b6292c6e2167ba0d46a8..032508387d1ac3083f89a49f9ee7ae8917eb169a 100644 (file)
@@ -243,8 +243,8 @@ ginvalidate(Oid opclassoid)
                        continue;                       /* don't need both, see check below loop */
                ereport(INFO,
                                (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
-                                errmsg("gin operator class \"%s\" is missing support function %d",
-                                               opclassname, i)));
+                  errmsg("gin operator class \"%s\" is missing support function %d",
+                                 opclassname, i)));
                result = false;
        }
        if (!opclassgroup ||
index 190a3d2f0cf56ba46e56177aa7c13aecd589eef0..ffd7fd631b8d40195d309ba82ae6d37a1613206a 100644 (file)
@@ -261,8 +261,8 @@ gistvalidate(Oid opclassoid)
                        continue;                       /* optional methods */
                ereport(INFO,
                                (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
-                                errmsg("gist operator class \"%s\" is missing support function %d",
-                                               opclassname, i)));
+                 errmsg("gist operator class \"%s\" is missing support function %d",
+                                opclassname, i)));
                result = false;
        }
 
index b9b7bb0395a3a5413c051d813886143d0d03b694..d8c5ed4d98e3b3d06f751ba585d9865c2f282483 100644 (file)
@@ -186,9 +186,9 @@ hashvalidate(Oid opclassoid)
                {
                        ereport(INFO,
                                        (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
-                       errmsg("hash operator family \"%s\" lacks support function for operator %s",
-                                  opfamilyname,
-                                  format_operator(oprform->amopopr))));
+                                        errmsg("hash operator family \"%s\" lacks support function for operator %s",
+                                                       opfamilyname,
+                                                       format_operator(oprform->amopopr))));
                        result = false;
                }
        }
index 4acc62a550ea5e22de4dbbfe0f04d10546653c13..24bd9be5e17cefebf33c52823ceca3a17e0c8b5a 100644 (file)
@@ -4571,7 +4571,7 @@ heap_lock_tuple(Relation relation, HeapTuple tuple,
        ItemId          lp;
        Page            page;
        Buffer          vmbuffer = InvalidBuffer;
-       BlockNumber     block;
+       BlockNumber block;
        TransactionId xid,
                                xmax;
        uint16          old_infomask,
@@ -5643,7 +5643,7 @@ static HTSU_Result
 heap_lock_updated_tuple_rec(Relation rel, ItemPointer tid, TransactionId xid,
                                                        LockTupleMode mode)
 {
-       HTSU_Result     result;
+       HTSU_Result result;
        ItemPointerData tupid;
        HeapTupleData mytup;
        Buffer          buf;
@@ -6698,6 +6698,7 @@ heap_prepare_freeze_tuple(HeapTupleHeader tuple, TransactionId cutoff_xid,
        if (tuple->t_infomask & HEAP_MOVED)
        {
                xid = HeapTupleHeaderGetXvac(tuple);
+
                /*
                 * For Xvac, we ignore the cutoff_xid and just always perform the
                 * freeze operation.  The oldest release in which such a value can
@@ -8841,9 +8842,9 @@ heap_xlog_lock(XLogReaderState *record)
         */
        if (xlrec->flags & XLH_LOCK_ALL_FROZEN_CLEARED)
        {
-               RelFileNode     rnode;
+               RelFileNode rnode;
                Buffer          vmbuffer = InvalidBuffer;
-               BlockNumber     block;
+               BlockNumber block;
                Relation        reln;
 
                XLogRecGetBlockTag(record, 0, &rnode, NULL, &block);
@@ -8914,9 +8915,9 @@ heap_xlog_lock_updated(XLogReaderState *record)
         */
        if (xlrec->flags & XLH_LOCK_ALL_FROZEN_CLEARED)
        {
-               RelFileNode     rnode;
+               RelFileNode rnode;
                Buffer          vmbuffer = InvalidBuffer;
-               BlockNumber     block;
+               BlockNumber block;
                Relation        reln;
 
                XLogRecGetBlockTag(record, 0, &rnode, NULL, &block);
index 31b0132b4223cca4c7050dce43ea842d14f515ba..bbb26493717a5b50b3aeb084a587295ad8f49ca4 100644 (file)
@@ -1667,7 +1667,7 @@ toast_delete_datum(Relation rel, Datum value)
        HeapTuple       toasttup;
        int                     num_indexes;
        int                     validIndex;
-       SnapshotData    SnapshotToast;
+       SnapshotData SnapshotToast;
 
        if (!VARATT_IS_EXTERNAL_ONDISK(attr))
                return;
@@ -1734,7 +1734,7 @@ toastrel_valueid_exists(Relation toastrel, Oid valueid)
        int                     num_indexes;
        int                     validIndex;
        Relation   *toastidxs;
-       SnapshotData    SnapshotToast;
+       SnapshotData SnapshotToast;
 
        /* Fetch a valid index relation */
        validIndex = toast_open_indexes(toastrel,
@@ -1819,7 +1819,7 @@ toast_fetch_datum(struct varlena * attr)
        int32           chunksize;
        int                     num_indexes;
        int                     validIndex;
-       SnapshotData    SnapshotToast;
+       SnapshotData SnapshotToast;
 
        if (!VARATT_IS_EXTERNAL_ONDISK(attr))
                elog(ERROR, "toast_fetch_datum shouldn't be called for non-ondisk datums");
@@ -1998,7 +1998,7 @@ toast_fetch_datum_slice(struct varlena * attr, int32 sliceoffset, int32 length)
        int32           chcpyend;
        int                     num_indexes;
        int                     validIndex;
-       SnapshotData    SnapshotToast;
+       SnapshotData SnapshotToast;
 
        if (!VARATT_IS_EXTERNAL_ONDISK(attr))
                elog(ERROR, "toast_fetch_datum_slice shouldn't be called for non-ondisk datums");
@@ -2094,7 +2094,7 @@ toast_fetch_datum_slice(struct varlena * attr, int32 sliceoffset, int32 length)
        init_toast_snapshot(&SnapshotToast);
        nextidx = startchunk;
        toastscan = systable_beginscan_ordered(toastrel, toastidxs[validIndex],
-                                                                                &SnapshotToast, nscankeys, toastkey);
+                                                                               &SnapshotToast, nscankeys, toastkey);
        while ((ttup = systable_getnext_ordered(toastscan, ForwardScanDirection)) != NULL)
        {
                /*
index 4ad8266a51c576edb7c60cf74b6c8cf14008caf3..defafa54b2968169abf28df85b33dcd72edf6259 100644 (file)
@@ -776,7 +776,7 @@ assign_client_encoding(const char *newval, void *extra)
                 */
                ereport(ERROR,
                                (errcode(ERRCODE_INVALID_TRANSACTION_STATE),
-                                errmsg("cannot change client_encoding in a parallel worker")));
+                         errmsg("cannot change client_encoding in a parallel worker")));
        }
 
        /* We do not expect an error if PrepareClientEncoding succeeded */
index 45dccb3e0de749baf07961116aa0fd7457d5ede6..f1c843e868c08bedf5a520df3052564f9c7fae15 100644 (file)
@@ -129,7 +129,7 @@ libpqrcv_get_conninfo(void)
 {
        PQconninfoOption *conn_opts;
        PQconninfoOption *conn_opt;
-       PQExpBufferData buf;
+       PQExpBufferData buf;
        char       *retval;
 
        Assert(streamConn != NULL);
@@ -145,7 +145,7 @@ libpqrcv_get_conninfo(void)
        /* build a clean connection string from pieces */
        for (conn_opt = conn_opts; conn_opt->keyword != NULL; conn_opt++)
        {
-               bool    obfuscate;
+               bool            obfuscate;
 
                /* Skip debug and empty options */
                if (strchr(conn_opt->dispchar, 'D') ||
index 47a67cf4b14cfffc468d81db2e59f34a1b56618a..b442d061ec0c341158f51d782749b9ccba718825 100644 (file)
@@ -193,10 +193,11 @@ SyncRepWaitForLSN(XLogRecPtr lsn, bool commit)
                ResetLatch(MyLatch);
 
                /*
-                * Acquiring the lock is not needed, the latch ensures proper barriers.
-                * If it looks like we're done, we must really be done, because once
-                * walsender changes the state to SYNC_REP_WAIT_COMPLETE, it will never
-                * update it again, so we can't be seeing a stale value in that case.
+                * Acquiring the lock is not needed, the latch ensures proper
+                * barriers. If it looks like we're done, we must really be done,
+                * because once walsender changes the state to SYNC_REP_WAIT_COMPLETE,
+                * it will never update it again, so we can't be seeing a stale value
+                * in that case.
                 */
                if (MyProc->syncRepState == SYNC_REP_WAIT_COMPLETE)
                        break;
index c94268a78a99ecdd6d2bbe290e5979bf7811b7db..821f611e4f8ee8f719c2c2e5db6667d56674dff0 100644 (file)
@@ -433,7 +433,8 @@ getNextFlagFromString(IspellDict *Conf, char **sflagset, char *sflag)
        if (Conf->flagMode == FM_LONG && maxstep > 0)
                ereport(ERROR,
                                (errcode(ERRCODE_CONFIG_FILE_ERROR),
-                       errmsg("invalid affix flag \"%s\" with \"long\" flag value", sbuf)));
+                                errmsg("invalid affix flag \"%s\" with \"long\" flag value",
+                                               sbuf)));
 
        *sflag = '\0';
 }
@@ -1263,8 +1264,9 @@ NIImportOOAffixes(IspellDict *Conf, const char *filename)
                                else if (STRNCMP(s, "default") != 0)
                                        ereport(ERROR,
                                                        (errcode(ERRCODE_CONFIG_FILE_ERROR),
-                                                  errmsg("Ispell dictionary supports only \"default\", "
-                                                                 "\"long\", and \"num\" flag value")));
+                                                        errmsg("Ispell dictionary supports only "
+                                                                       "\"default\", \"long\", "
+                                                                       "and \"num\" flag values")));
                        }
                }
 
index ab46823e8185b4c7ec01dcbd61f979bb1dbbc6c1..987cfd18625008bcae84c2ce9d5bc3f1758a072f 100644 (file)
@@ -1849,9 +1849,9 @@ jsonb_object_agg_transfn(PG_FUNCTION_ARGS)
        single_scalar = false;
 
        /*
-        * values can be anything, including structured and null, so we treat
-        * them as in json_agg_transfn, except that single scalars are always
-        * pushed as WJB_VALUE items.
+        * values can be anything, including structured and null, so we treat them
+        * as in json_agg_transfn, except that single scalars are always pushed as
+        * WJB_VALUE items.
         */
 
        while ((type = JsonbIteratorNext(&it, &v, false)) != WJB_DONE)
index ab4aa7ca2d6f193bc1bb64683b8e6ab7284f1c44..c0a4a0606bce074a16e1218adb69aff4d349d5b5 100644 (file)
@@ -430,15 +430,16 @@ pushStop(TSQueryParserState state)
 
 #define STACKDEPTH     32
 
-typedef struct OperatorElement {
-       int8    op;
-       int16   distance;
+typedef struct OperatorElement
+{
+       int8            op;
+       int16           distance;
 } OperatorElement;
 
 static void
 pushOpStack(OperatorElement *stack, int *lenstack, int8 op, int16 distance)
 {
-       if (*lenstack == STACKDEPTH) /* internal error */
+       if (*lenstack == STACKDEPTH)    /* internal error */
                elog(ERROR, "tsquery stack too small");
 
        stack[*lenstack].op = op;
@@ -449,20 +450,20 @@ pushOpStack(OperatorElement *stack, int *lenstack, int8 op, int16 distance)
 
 static void
 cleanOpStack(TSQueryParserState state,
-                  OperatorElement *stack, int *lenstack, int8 op)
+                        OperatorElement *stack, int *lenstack, int8 op)
 {
-       int     opPriority = OP_PRIORITY(op);
+       int                     opPriority = OP_PRIORITY(op);
 
-       while(*lenstack)
+       while (*lenstack)
        {
                /* NOT is right associative unlike to others */
                if ((op != OP_NOT && opPriority > OP_PRIORITY(stack[*lenstack - 1].op)) ||
-                       (op == OP_NOT && opPriority >=  OP_PRIORITY(stack[*lenstack - 1].op)))
+               (op == OP_NOT && opPriority >= OP_PRIORITY(stack[*lenstack - 1].op)))
                        break;
 
                (*lenstack)--;
                pushOperator(state, stack[*lenstack].op,
-                                                       stack[*lenstack].distance);
+                                        stack[*lenstack].distance);
        }
 }
 
@@ -480,7 +481,7 @@ makepol(TSQueryParserState state,
        ts_tokentype type;
        int                     lenval = 0;
        char       *strval = NULL;
-       OperatorElement opstack[STACKDEPTH];
+       OperatorElement opstack[STACKDEPTH];
        int                     lenstack = 0;
        int16           weight = 0;
        bool            prefix;
@@ -503,7 +504,7 @@ makepol(TSQueryParserState state,
                                makepol(state, pushval, opaque);
                                break;
                        case PT_CLOSE:
-                               cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */);
+                               cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */ );
                                return;
                        case PT_ERR:
                        default:
@@ -514,7 +515,7 @@ makepol(TSQueryParserState state,
                }
        }
 
-       cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */);
+       cleanOpStack(state, opstack, &lenstack, OP_OR /* lowest */ );
 }
 
 static void
@@ -845,8 +846,8 @@ infix(INFIX *in, int parentPriority, bool rightPhraseOp)
 
                in->curpol++;
                if (priority < parentPriority ||
-                        /* phrase operator depends on order */
-                        (op == OP_PHRASE && rightPhraseOp))
+               /* phrase operator depends on order */
+                       (op == OP_PHRASE && rightPhraseOp))
                {
                        needParenthesis = true;
                        RESIZEBUF(in, 2);
@@ -916,7 +917,7 @@ tsqueryout(PG_FUNCTION_ARGS)
        nrm.cur = nrm.buf = (char *) palloc(sizeof(char) * nrm.buflen);
        *(nrm.cur) = '\0';
        nrm.op = GETOPERAND(query);
-       infix(&nrm, -1 /* lowest priority */, false);
+       infix(&nrm, -1 /* lowest priority */ , false);
 
        PG_FREE_IF_COPY(query, 0);
        PG_RETURN_CSTRING(nrm.buf);
index 8c2df73ea6b72351ba06055b97e3325b7566a581..d9ec8ed43149275efc2bb08c8dc0018651397656 100644 (file)
@@ -406,7 +406,7 @@ normalize_phrase_tree(NODE *node)
 
        if (node->valnode->qoperator.oper == OP_NOT)
        {
-               NODE *orignode = node;
+               NODE       *orignode = node;
 
                /* eliminate NOT sequence */
                while (node->valnode->type == QI_OPR &&
@@ -430,8 +430,8 @@ normalize_phrase_tree(NODE *node)
                node->right = normalize_phrase_tree(node->right);
 
                /*
-                * if subtree contains only nodes with higher "priority" then
-                * we are done. See comment near NODE_PRIORITY()
+                * if subtree contains only nodes with higher "priority" then we are
+                * done. See comment near NODE_PRIORITY()
                 */
                if (NODE_PRIORITY(node) <= NODE_PRIORITY(node->right) &&
                        NODE_PRIORITY(node) <= NODE_PRIORITY(node->left))
index 6756f2632d44c0a390959bb29bd5adac41a32a0e..510565c339e49272d71b7ec1ef6d7a2f2210d8bc 100644 (file)
@@ -299,9 +299,9 @@ struct Tuplesortstate
         * Function to read a stored tuple from tape back into memory. 'len' is
         * the already-read length of the stored tuple.  Create a palloc'd copy,
         * initialize tuple/datum1/isnull1 in the target SortTuple struct, and
-        * decrease state->availMem by the amount of memory space consumed.
-        * (See batchUsed notes for details on how memory is handled when
-        * incremental accounting is abandoned.)
+        * decrease state->availMem by the amount of memory space consumed. (See
+        * batchUsed notes for details on how memory is handled when incremental
+        * accounting is abandoned.)
         */
        void            (*readtup) (Tuplesortstate *state, SortTuple *stup,
                                                                                int tapenum, unsigned int len);
@@ -309,8 +309,8 @@ struct Tuplesortstate
        /*
         * Function to move a caller tuple.  This is usually implemented as a
         * memmove() shim, but function may also perform additional fix-up of
-        * caller tuple where needed.  Batch memory support requires the
-        * movement of caller tuples from one location in memory to another.
+        * caller tuple where needed.  Batch memory support requires the movement
+        * of caller tuples from one location in memory to another.
         */
        void            (*movetup) (void *dest, void *src, unsigned int len);
 
index 4ee10fcb39cc43a396f232d18fef2ea8ecdeea84..a5c2d09e2949c6f553b80ca6fada1479fd51c252 100644 (file)
@@ -6042,15 +6042,16 @@ getOwnedSeqs(Archive *fout, TableInfo tblinfo[], int numTables)
                 * We need to dump the components that are being dumped for the table
                 * and any components which the sequence is explicitly marked with.
                 *
-                * We can't simply use the set of components which are being dumped for
-                * the table as the table might be in an extension (and only the
+                * We can't simply use the set of components which are being dumped
+                * for the table as the table might be in an extension (and only the
                 * non-extension components, eg: ACLs if changed, security labels, and
-                * policies, are being dumped) while the sequence is not (and therefore
-                * the definition and other components should also be dumped).
+                * policies, are being dumped) while the sequence is not (and
+                * therefore the definition and other components should also be
+                * dumped).
                 *
                 * If the sequence is part of the extension then it should be properly
-                * marked by checkExtensionMembership() and this will be a no-op as the
-                * table will be equivalently marked.
+                * marked by checkExtensionMembership() and this will be a no-op as
+                * the table will be equivalently marked.
                 */
                seqinfo->dobj.dump = seqinfo->dobj.dump | owning_tab->dobj.dump;
 
index fdbaa9dffb55edcebc40435ae3d0c7e60451b240..54a9f482008489a7f3886fda8644a1624a731aa9 100644 (file)
@@ -1098,7 +1098,8 @@ dumpTablespaces(PGconn *conn)
        else if (server_version >= 90200)
                res = executeQuery(conn, "SELECT oid, spcname, "
                                                 "pg_catalog.pg_get_userbyid(spcowner) AS spcowner, "
-                                                  "pg_catalog.pg_tablespace_location(oid), spcacl, '' as rspcacl, "
+                                                  "pg_catalog.pg_tablespace_location(oid), "
+                                                  "spcacl, '' as rspcacl, "
                                                   "array_to_string(spcoptions, ', '),"
                                                "pg_catalog.shobj_description(oid, 'pg_tablespace') "
                                                   "FROM pg_catalog.pg_tablespace "
@@ -1329,10 +1330,10 @@ dumpCreateDB(PGconn *conn)
                                                   "datistemplate, "
                                                   "(SELECT pg_catalog.array_agg(acl) FROM (SELECT pg_catalog.unnest(coalesce(datacl,pg_catalog.acldefault('d',datdba))) AS acl "
                                                   "EXCEPT SELECT pg_catalog.unnest(pg_catalog.acldefault('d',datdba))) as foo)"
-                                                  "AS datacl,"
+                                                  "AS datacl, "
                                                   "(SELECT pg_catalog.array_agg(acl) FROM (SELECT pg_catalog.unnest(pg_catalog.acldefault('d',datdba)) AS acl "
                                                   "EXCEPT SELECT pg_catalog.unnest(coalesce(datacl,pg_catalog.acldefault('d',datdba)))) as foo)"
-                                                  "AS rdatacl,"
+                                                  "AS rdatacl, "
                                                   "datconnlimit, "
                                                   "(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
                          "FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
@@ -1343,7 +1344,8 @@ dumpCreateDB(PGconn *conn)
                                                   "coalesce(rolname, (select rolname from pg_authid where oid=(select datdba from pg_database where datname='template0'))), "
                                                   "pg_encoding_to_char(d.encoding), "
                                                   "datcollate, datctype, datfrozenxid, datminmxid, "
-                                                  "datistemplate, datacl, '' as rdatacl, datconnlimit, "
+                                                  "datistemplate, datacl, '' as rdatacl, "
+                                                  "datconnlimit, "
                                                   "(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
                          "FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
                                                   "WHERE datallowconn ORDER BY 1");
@@ -1353,7 +1355,8 @@ dumpCreateDB(PGconn *conn)
                                                   "coalesce(rolname, (select rolname from pg_authid where oid=(select datdba from pg_database where datname='template0'))), "
                                                   "pg_encoding_to_char(d.encoding), "
                                          "datcollate, datctype, datfrozenxid, 0 AS datminmxid, "
-                                                  "datistemplate, datacl, '' as rdatacl, datconnlimit, "
+                                                  "datistemplate, datacl, '' as rdatacl, "
+                                                  "datconnlimit, "
                                                   "(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
                          "FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
                                                   "WHERE datallowconn ORDER BY 1");
@@ -1363,7 +1366,8 @@ dumpCreateDB(PGconn *conn)
                                                   "coalesce(rolname, (select rolname from pg_authid where oid=(select datdba from pg_database where datname='template0'))), "
                                                   "pg_encoding_to_char(d.encoding), "
                                                   "null::text AS datcollate, null::text AS datctype, datfrozenxid, 0 AS datminmxid, "
-                                                  "datistemplate, datacl, '' as rdatacl, datconnlimit, "
+                                                  "datistemplate, datacl, '' as rdatacl, "
+                                                  "datconnlimit, "
                                                   "(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
                          "FROM pg_database d LEFT JOIN pg_authid u ON (datdba = u.oid) "
                                                   "WHERE datallowconn ORDER BY 1");
@@ -1373,7 +1377,8 @@ dumpCreateDB(PGconn *conn)
                                                   "coalesce(usename, (select usename from pg_shadow where usesysid=(select datdba from pg_database where datname='template0'))), "
                                                   "pg_encoding_to_char(d.encoding), "
                                                   "null::text AS datcollate, null::text AS datctype, datfrozenxid, 0 AS datminmxid, "
-                                                  "datistemplate, datacl, '' as rdatacl, -1 as datconnlimit, "
+                                                  "datistemplate, datacl, '' as rdatacl, "
+                                                  "-1 as datconnlimit, "
                                                   "(SELECT spcname FROM pg_tablespace t WHERE t.oid = d.dattablespace) AS dattablespace "
                   "FROM pg_database d LEFT JOIN pg_shadow u ON (datdba = usesysid) "
                                                   "WHERE datallowconn ORDER BY 1");
@@ -1383,7 +1388,8 @@ dumpCreateDB(PGconn *conn)
                                                   "coalesce(usename, (select usename from pg_shadow where usesysid=(select datdba from pg_database where datname='template0'))), "
                                                   "pg_encoding_to_char(d.encoding), "
                                                   "null::text AS datcollate, null::text AS datctype, datfrozenxid, 0 AS datminmxid, "
-                                                  "datistemplate, datacl, '' as rdatacl, -1 as datconnlimit, "
+                                                  "datistemplate, datacl, '' as rdatacl, "
+                                                  "-1 as datconnlimit, "
                                                   "'pg_default' AS dattablespace "
                   "FROM pg_database d LEFT JOIN pg_shadow u ON (datdba = usesysid) "
                                                   "WHERE datallowconn ORDER BY 1");
@@ -1395,7 +1401,8 @@ dumpCreateDB(PGconn *conn)
                                                   "(select usename from pg_shadow where usesysid=(select datdba from pg_database where datname='template0'))), "
                                                   "pg_encoding_to_char(d.encoding), "
                                                   "null::text AS datcollate, null::text AS datctype, 0 AS datfrozenxid, 0 AS datminmxid, "
-                                                  "datistemplate, '' as datacl, '' as rdatacl, -1 as datconnlimit, "
+                                                  "datistemplate, '' as datacl, '' as rdatacl, "
+                                                  "-1 as datconnlimit, "
                                                   "'pg_default' AS dattablespace "
                                                   "FROM pg_database d "
                                                   "WHERE datallowconn ORDER BY 1");
@@ -1522,7 +1529,8 @@ dumpCreateDB(PGconn *conn)
                }
 
                if (!skip_acls &&
-                       !buildACLCommands(fdbname, NULL, "DATABASE", dbacl, rdbacl, dbowner,
+                       !buildACLCommands(fdbname, NULL, "DATABASE",
+                                                         dbacl, rdbacl, dbowner,
                                                          "", server_version, buf))
                {
                        fprintf(stderr, _("%s: could not parse ACL list (%s) for database \"%s\"\n"),
index 5a042b85528173d970c17587432187b9d99be49a..37cbdcdaa5990e374baea8fee7902eab8d0da95e 100644 (file)
@@ -161,8 +161,8 @@ my %pgdump_runs = (
                        'postgres', ], },
        role => {
                dump_cmd => [
-                       'pg_dump',                          '-f',
-                       "$tempdir/role.sql",                '--role=regress_dump_test_role',
+                       'pg_dump', '-f', "$tempdir/role.sql",
+                       '--role=regress_dump_test_role',
                        '--schema=dump_test_second_schema', 'postgres', ], },
        schema_only => {
                dump_cmd =>
@@ -939,8 +939,8 @@ my %tests = (
                        test_schema_plus_blobs   => 1, }, },
        'CREATE DATABASE dump_test' => {
                create_order => 47,
-               create_sql => 'CREATE DATABASE dump_test;',
-               regexp => qr/^
+               create_sql   => 'CREATE DATABASE dump_test;',
+               regexp       => qr/^
                        \QCREATE DATABASE dump_test WITH TEMPLATE = template0 \E
                        .*;/xm,
                like   => { pg_dumpall_dbprivs => 1, },
@@ -2192,12 +2192,12 @@ my %tests = (
                        test_schema_plus_blobs => 1, }, },
        'GRANT CREATE ON DATABASE dump_test' => {
                create_order => 48,
-               create_sql   => 'GRANT CREATE ON DATABASE dump_test TO regress_dump_test_role;',
+               create_sql =>
+                 'GRANT CREATE ON DATABASE dump_test TO regress_dump_test_role;',
                regexp => qr/^
                        \QGRANT CREATE ON DATABASE dump_test TO regress_dump_test_role;\E
                        /xm,
-               like => {
-                       pg_dumpall_dbprivs       => 1, },
+               like   => { pg_dumpall_dbprivs => 1, },
                unlike => {
                        binary_upgrade           => 1,
                        clean                    => 1,
@@ -2212,14 +2212,15 @@ my %tests = (
                        only_dump_test_table     => 1,
                        pg_dumpall_globals       => 1,
                        schema_only              => 1,
-                       section_pre_data         => 1, ,
+                       section_pre_data         => 1,
                        test_schema_plus_blobs   => 1, }, },
        'GRANT SELECT ON TABLE test_table' => {
                create_order => 5,
                create_sql   => 'GRANT SELECT ON TABLE dump_test.test_table
                                                   TO regress_dump_test_role;',
-               regexp => qr/^GRANT SELECT ON TABLE test_table TO regress_dump_test_role;/m,
-               like   => {
+               regexp =>
+                 qr/^GRANT SELECT ON TABLE test_table TO regress_dump_test_role;/m,
+               like => {
                        binary_upgrade          => 1,
                        clean                   => 1,
                        clean_if_exists         => 1,
@@ -2242,8 +2243,9 @@ my %tests = (
                create_sql   => 'GRANT SELECT ON
                                                   TABLE dump_test_second_schema.test_third_table
                                                   TO regress_dump_test_role;',
-               regexp => qr/^GRANT SELECT ON TABLE test_third_table TO regress_dump_test_role;/m,
-               like   => {
+               regexp =>
+qr/^GRANT SELECT ON TABLE test_third_table TO regress_dump_test_role;/m,
+               like => {
                        binary_upgrade           => 1,
                        clean                    => 1,
                        clean_if_exists          => 1,
@@ -2480,12 +2482,11 @@ my %tests = (
        'REVOKE CONNECT ON DATABASE dump_test FROM public' => {
                create_order => 49,
                create_sql   => 'REVOKE CONNECT ON DATABASE dump_test FROM public;',
-               regexp => qr/^
+               regexp       => qr/^
                        \QREVOKE CONNECT,TEMPORARY ON DATABASE dump_test FROM PUBLIC;\E\n
                        \QGRANT TEMPORARY ON DATABASE dump_test TO PUBLIC;\E
                        /xm,
-               like => {
-                       pg_dumpall_dbprivs       => 1, },
+               like   => { pg_dumpall_dbprivs => 1, },
                unlike => {
                        binary_upgrade           => 1,
                        clean                    => 1,
index 95559df19fea933ce1fb85da4ec166430062963e..93cb92aea6bbde770520829876d4f38e55058d36 100644 (file)
@@ -29,8 +29,8 @@ extern Size AsyncShmemSize(void);
 extern void AsyncShmemInit(void);
 
 extern void NotifyMyFrontEnd(const char *channel,
-                                                        const char *payload,
-                                                        int32 srcPid);
+                                const char *payload,
+                                int32 srcPid);
 
 /* notify-related SQL statements */
 extern void Async_Notify(const char *channel, const char *payload);
index d1f4114d21fb5ffdd55ea044970285c61c6138be..5062aaafee9f3de88ed7732e0e67c17c7b47a4d2 100644 (file)
@@ -221,7 +221,7 @@ typedef struct
 #define OP_NOT                 1
 #define OP_AND                 2
 #define OP_OR                  3
-#define OP_PHRASE              4  /* highest code, tsquery_cleanup.c */
+#define OP_PHRASE              4               /* highest code, tsquery_cleanup.c */
 #define OP_COUNT               4
 
 extern const int tsearch_op_priority[OP_COUNT];
index 4dd7949a9c992ece129d4f0ba74c19e65832d655..f520e7725f308c35f49458129be126f480dd26e4 100644 (file)
@@ -472,7 +472,7 @@ PLy_output(volatile int level, PyObject *self, PyObject *args, PyObject *kw)
                        else
                        {
                                PLy_exception_set(PyExc_TypeError,
-                                                                 "'%s' is an invalid keyword argument for this function",
+                                        "'%s' is an invalid keyword argument for this function",
                                                                  keyword);
                                return NULL;
                        }
@@ -528,7 +528,7 @@ PLy_output(volatile int level, PyObject *self, PyObject *args, PyObject *kw)
                                 (column_name != NULL) ?
                                 err_generic_string(PG_DIAG_COLUMN_NAME, column_name) : 0,
                                 (constraint_name != NULL) ?
-                                err_generic_string(PG_DIAG_CONSTRAINT_NAME, constraint_name) : 0,
+                       err_generic_string(PG_DIAG_CONSTRAINT_NAME, constraint_name) : 0,
                                 (datatype_name != NULL) ?
                                 err_generic_string(PG_DIAG_DATATYPE_NAME, datatype_name) : 0,
                                 (table_name != NULL) ?
index fb4f5737e584817bd0b0d16f3b4b70135120de73..f02beb3d9c481cf27b295d885a98a18076ed6978 100644 (file)
@@ -223,8 +223,8 @@ my %tests = (
                        schema_only       => 1,
                        section_pre_data  => 1,
                        section_post_data => 1, }, },
-    'CREATE SEQUENCE regress_pg_dump_table_col1_seq' => {
-            regexp => qr/^
+       'CREATE SEQUENCE regress_pg_dump_table_col1_seq' => {
+               regexp => qr/^
                     \QCREATE SEQUENCE regress_pg_dump_table_col1_seq\E
                     \n\s+\QSTART WITH 1\E
                     \n\s+\QINCREMENT BY 1\E
@@ -232,20 +232,20 @@ my %tests = (
                     \n\s+\QNO MAXVALUE\E
                     \n\s+\QCACHE 1;\E
                     $/xm,
-            like   => { binary_upgrade => 1, },
-            unlike => {
-                    clean              => 1,
-                    clean_if_exists    => 1,
-                    createdb           => 1,
-                    defaults           => 1,
-                    no_privs           => 1,
-                    no_owner           => 1,
-                    pg_dumpall_globals => 1,
-                    schema_only        => 1,
-                    section_pre_data   => 1,
-                    section_post_data  => 1, }, },
-    'CREATE SEQUENCE regress_pg_dump_seq' => {
-            regexp => qr/^
+               like   => { binary_upgrade => 1, },
+               unlike => {
+                       clean              => 1,
+                       clean_if_exists    => 1,
+                       createdb           => 1,
+                       defaults           => 1,
+                       no_privs           => 1,
+                       no_owner           => 1,
+                       pg_dumpall_globals => 1,
+                       schema_only        => 1,
+                       section_pre_data   => 1,
+                       section_post_data  => 1, }, },
+       'CREATE SEQUENCE regress_pg_dump_seq' => {
+               regexp => qr/^
                     \QCREATE SEQUENCE regress_pg_dump_seq\E
                     \n\s+\QSTART WITH 1\E
                     \n\s+\QINCREMENT BY 1\E
@@ -253,18 +253,18 @@ my %tests = (
                     \n\s+\QNO MAXVALUE\E
                     \n\s+\QCACHE 1;\E
                     $/xm,
-            like   => { binary_upgrade => 1, },
-            unlike => {
-                    clean              => 1,
-                    clean_if_exists    => 1,
-                    createdb           => 1,
-                    defaults           => 1,
-                    no_privs           => 1,
-                    no_owner           => 1,
-                    pg_dumpall_globals => 1,
-                    schema_only        => 1,
-                    section_pre_data   => 1,
-                    section_post_data  => 1, }, },
+               like   => { binary_upgrade => 1, },
+               unlike => {
+                       clean              => 1,
+                       clean_if_exists    => 1,
+                       createdb           => 1,
+                       defaults           => 1,
+                       no_privs           => 1,
+                       no_owner           => 1,
+                       pg_dumpall_globals => 1,
+                       schema_only        => 1,
+                       section_pre_data   => 1,
+                       section_post_data  => 1, }, },
        'CREATE TABLE regress_pg_dump_table' => {
                regexp => qr/^
                        \QCREATE TABLE regress_pg_dump_table (\E
@@ -353,8 +353,8 @@ my %tests = (
                        no_privs           => 1,
                        pg_dumpall_globals => 1,
                        section_post_data  => 1, }, },
-       'GRANT SELECT(col2) ON regress_pg_dump_table TO regress_dump_test_role' => {
-               create_order => 4,
+       'GRANT SELECT(col2) ON regress_pg_dump_table TO regress_dump_test_role' =>
+         { create_order => 4,
                create_sql   => 'GRANT SELECT(col2) ON regress_pg_dump_table
                                                   TO regress_dump_test_role;',
                regexp => qr/^
@@ -373,9 +373,10 @@ my %tests = (
                        no_privs           => 1,
                        pg_dumpall_globals => 1,
                        section_post_data  => 1, }, },
-       'GRANT USAGE ON regress_pg_dump_table_col1_seq TO regress_dump_test_role' => {
+       'GRANT USAGE ON regress_pg_dump_table_col1_seq TO regress_dump_test_role'
+         => {
                create_order => 5,
-               create_sql   => 'GRANT USAGE ON SEQUENCE regress_pg_dump_table_col1_seq
+               create_sql => 'GRANT USAGE ON SEQUENCE regress_pg_dump_table_col1_seq
                                   TO regress_dump_test_role;',
                regexp => qr/^
                        \QGRANT USAGE ON SEQUENCE regress_pg_dump_table_col1_seq TO regress_dump_test_role;\E
@@ -397,8 +398,7 @@ my %tests = (
                regexp => qr/^
                        \QGRANT USAGE ON SEQUENCE regress_pg_dump_seq TO regress_dump_test_role;\E
                        $/xm,
-               like => {
-                       binary_upgrade   => 1, },
+               like   => { binary_upgrade => 1, },
                unlike => {
                        clean              => 1,
                        clean_if_exists    => 1,
index e629373ea2b9bd5e02118729d9dd9a04fa14331b..fede1e601b96d122a3594e01195c6744ed4f51ef 100644 (file)
@@ -475,8 +475,8 @@ sub backup
        my $name        = $self->name;
 
        print "# Taking pg_basebackup $backup_name from node \"$name\"\n";
-       TestLib::system_or_bail('pg_basebackup', '-D', $backup_path,
-                                                       '-p', $port, '-x');
+       TestLib::system_or_bail('pg_basebackup', '-D', $backup_path, '-p', $port,
+               '-x');
        print "# Backup finished\n";
 }
 
index 82371503cccc792357f8fffb508cac9eb1e08f25..d1f6d78388b4125d887e289dc8dfc80228f993ba 100644 (file)
@@ -66,15 +66,15 @@ $node_master->backup('my_backup');
 # target TXID.
 $node_master->safe_psql('postgres',
        "INSERT INTO tab_int VALUES (generate_series(1001,2000))");
-my $ret =
-  $node_master->safe_psql('postgres', "SELECT pg_current_xlog_location(), txid_current();");
+my $ret = $node_master->safe_psql('postgres',
+       "SELECT pg_current_xlog_location(), txid_current();");
 my ($lsn2, $recovery_txid) = split /\|/, $ret;
 
 # More data, with recovery target timestamp
 $node_master->safe_psql('postgres',
        "INSERT INTO tab_int VALUES (generate_series(2001,3000))");
-$ret =
-  $node_master->safe_psql('postgres', "SELECT pg_current_xlog_location(), now();");
+$ret = $node_master->safe_psql('postgres',
+       "SELECT pg_current_xlog_location(), now();");
 my ($lsn3, $recovery_time) = split /\|/, $ret;
 
 # Even more data, this time with a recovery target name
index e27e3f6476482e7a1d6fc6ba620349b21036dfe3..52dde65a0e2fe1ffd5d7d58f669ea3536e31b070 100644 (file)
@@ -386,13 +386,11 @@ sub GenerateTimezoneFiles
 
        print "Generating timezone files...";
 
-       my @args = ("$conf/zic/zic",
-                               '-d',
-                               "$target/share/timezone");
+       my @args = ("$conf/zic/zic", '-d', "$target/share/timezone");
        foreach (@tzfiles)
        {
                my $tzfile = $_;
-               push(@args, "src/timezone/data/$tzfile")
+               push(@args, "src/timezone/data/$tzfile");
        }
 
        system(@args);
@@ -643,9 +641,9 @@ sub CopyIncludeFiles
                next unless (-d "src/include/$d");
 
                EnsureDirectories("$target/include/server/$d");
-               my @args = ('xcopy', '/s', '/i', '/q', '/r', '/y',
-                                "src\\include\\$d\\*.h",
-                                "$ctarget\\include\\server\\$d\\");
+               my @args = (
+                       'xcopy', '/s', '/i', '/q', '/r', '/y', "src\\include\\$d\\*.h",
+                       "$ctarget\\include\\server\\$d\\");
                system(@args) && croak("Failed to copy include directory $d\n");
        }
        closedir($D);
@@ -699,10 +697,11 @@ sub GenerateNLSFiles
 
                        EnsureDirectories($target, "share/locale/$lang",
                                "share/locale/$lang/LC_MESSAGES");
-                       my @args = ("$nlspath\\bin\\msgfmt",
-                          '-o',
-                          "$target\\share\\locale\\$lang\\LC_MESSAGES\\$prgm-$majorver.mo",
-                          $_);
+                       my @args = (
+                               "$nlspath\\bin\\msgfmt",
+                               '-o',
+"$target\\share\\locale\\$lang\\LC_MESSAGES\\$prgm-$majorver.mo",
+                               $_);
                        system(@args) && croak("Could not run msgfmt on $dir\\$_");
                        print ".";
                }
index 35c0158830d3eb9af9c68dcd8c7b85fa1c93a7d9..b4f946474f509c2421cd386588d76a260fe35447 100644 (file)
@@ -469,8 +469,9 @@ sub upgradecheck
        print "\nSetting up new cluster\n\n";
        standard_initdb() or exit 1;
        print "\nRunning pg_upgrade\n\n";
-       @args = ('pg_upgrade', '-d', "$data.old", '-D', $data, '-b', $bindir,
-                        '-B', $bindir);
+       @args = (
+               'pg_upgrade', '-d', "$data.old", '-D', $data, '-b',
+               $bindir,      '-B', $bindir);
        system(@args) == 0 or exit 1;
        print "\nStarting new cluster\n\n";
        @args = ('pg_ctl', '-l', "$logdir/postmaster2.log", '-w', 'start');
index 0c61fc287ad2bc7f94cfe195a12e5b9250488999..dff6f65ef0fe65be303b4caa383d52666df60395 100644 (file)
@@ -40,6 +40,7 @@ AggClauseCosts
 AggHashEntry
 AggInfo
 AggPath
+AggSplit
 AggState
 AggStatePerAgg
 AggStatePerGroup
@@ -123,6 +124,7 @@ ArrayMetaState
 ArrayParseState
 ArrayRef
 ArrayRefExprState
+ArrayRemapInfo
 ArrayType
 AsyncQueueControl
 AsyncQueueEntry
@@ -612,6 +614,8 @@ FmgrBuiltin
 FmgrHookEventType
 FmgrInfo
 ForeignDataWrapper
+ForeignKeyCacheInfo
+ForeignKeyOptInfo
 ForeignPath
 ForeignScan
 ForeignScanState
@@ -917,6 +921,7 @@ ImportQual
 InclusionOpaque
 IncrementVarSublevelsUp_context
 Index
+IndexAMProperty
 IndexAmRoutine
 IndexArrayKeyInfo
 IndexAttrBitmapKind
@@ -1239,6 +1244,7 @@ OpFamilyMember
 OpFamilyOpFuncGroup
 OpclassInfo
 Operator
+OperatorElement
 OpfamilyInfo
 OprCacheEntry
 OprCacheKey
@@ -1456,7 +1462,6 @@ ParsedText
 ParsedWord
 ParserSetupHook
 ParserState
-PartialAggType
 Path
 PathClauseUsage
 PathCostComparison
@@ -1651,6 +1656,7 @@ RangeBox
 RangeFunction
 RangeIOData
 RangeQueryClause
+RangeRemapInfo
 RangeSubselect
 RangeTableSample
 RangeTblEntry
@@ -1671,7 +1677,8 @@ RecheckForeignScan_function
 RecordCacheEntry
 RecordCompareData
 RecordIOData
-RecordTypemodMap
+RecordRemapInfo
+RecordTypmodMap
 RecoveryTargetAction
 RecoveryTargetType
 RectBox
@@ -1708,8 +1715,6 @@ RelfilenodeMapEntry
 RelfilenodeMapKey
 Relids
 RelocationBufferInfo
-RemapClass
-RemapInfo
 RenameStmt
 ReopenPtr
 ReorderBuffer
@@ -2123,6 +2128,8 @@ TupleHashEntryData
 TupleHashIterator
 TupleHashTable
 TupleQueueReader
+TupleRemapClass
+TupleRemapInfo
 TupleTableSlot
 Tuplesortstate
 Tuplestorestate
@@ -2323,6 +2330,7 @@ amgettuple_function
 aminsert_function
 ammarkpos_function
 amoptions_function
+amproperty_function
 amrescan_function
 amrestrpos_function
 amvacuumcleanup_function
@@ -2353,6 +2361,7 @@ cashKEY
 celt
 cfp
 check_agg_arguments_context
+check_function_callback
 check_network_data
 check_object_relabel_type
 check_password_hook_type
@@ -2374,7 +2383,6 @@ core_yy_extra_type
 core_yyscan_t
 corrupt_items
 cost_qual_eval_context
-count_agg_clauses_context
 create_upper_paths_hook_type
 createdb_failure_params
 crosstab_HashEnt
@@ -2452,6 +2460,7 @@ generate_series_numeric_fctx
 generate_series_timestamp_fctx
 generate_series_timestamptz_fctx
 generate_subscripts_fctx
+get_agg_clause_costs_context
 get_attavgwidth_hook_type
 get_index_stats_hook_type
 get_relation_info_hook_type
@@ -2580,7 +2589,6 @@ pairingheap_comparator
 pairingheap_node
 parallel_worker_main_type
 parse_error_callback_arg
-partial_agg_context
 pcolor
 pendingPosition
 pgParameterStatus
@@ -2839,6 +2847,7 @@ vbits
 walrcv_connect_type
 walrcv_disconnect_type
 walrcv_endstreaming_type
+walrcv_get_conninfo_type
 walrcv_identify_system_type
 walrcv_readtimelinehistoryfile_type
 walrcv_receive_type