]> granicus.if.org Git - postgresql/commitdiff
tableam: Formatting and other minor cleanups.
authorAndres Freund <andres@anarazel.de>
Sun, 31 Mar 2019 03:13:56 +0000 (20:13 -0700)
committerAndres Freund <andres@anarazel.de>
Mon, 1 Apr 2019 01:16:53 +0000 (18:16 -0700)
The superflous heapam_xlog.h includes were reported by Peter
Geoghegan.

src/backend/access/hash/hash_xlog.c
src/backend/access/heap/heapam_handler.c
src/backend/access/nbtree/nbtxlog.c
src/backend/access/table/tableam.c
src/backend/access/table/tableamapi.c
src/backend/optimizer/util/plancat.c
src/include/access/tableam.h

index 20441e307a81d1c563acfb262a035db157427d8c..d7b70981101430d8c9d60bb8778256c6e1a0bc87 100644 (file)
@@ -14,7 +14,6 @@
  */
 #include "postgres.h"
 
-#include "access/heapam_xlog.h"
 #include "access/bufmask.h"
 #include "access/hash.h"
 #include "access/hash_xlog.h"
index cf6e35670b0e24e81d63a0fa1c763dad36448e1c..314806bdcd7da70de5a4b96437049c310cc4f303 100644 (file)
@@ -48,8 +48,7 @@
 #include "utils/rel.h"
 
 
-static void
-reform_and_rewrite_tuple(HeapTuple tuple,
+static void reform_and_rewrite_tuple(HeapTuple tuple,
                                                 Relation OldHeap, Relation NewHeap,
                                                 Datum *values, bool *isnull, RewriteState rwstate);
 
@@ -247,8 +246,9 @@ heapam_tuple_insert(Relation relation, TupleTableSlot *slot, CommandId cid,
 }
 
 static void
-heapam_tuple_insert_speculative(Relation relation, TupleTableSlot *slot, CommandId cid,
-                                                               int options, BulkInsertState bistate, uint32 specToken)
+heapam_tuple_insert_speculative(Relation relation, TupleTableSlot *slot,
+                                                               CommandId cid, int options,
+                                                               BulkInsertState bistate, uint32 specToken)
 {
        bool            shouldFree = true;
        HeapTuple       tuple = ExecFetchSlotHeapTuple(slot, true, &shouldFree);
@@ -269,8 +269,8 @@ heapam_tuple_insert_speculative(Relation relation, TupleTableSlot *slot, Command
 }
 
 static void
-heapam_tuple_complete_speculative(Relation relation, TupleTableSlot *slot, uint32 spekToken,
-                                                                 bool succeeded)
+heapam_tuple_complete_speculative(Relation relation, TupleTableSlot *slot,
+                                                                 uint32 spekToken, bool succeeded)
 {
        bool            shouldFree = true;
        HeapTuple       tuple = ExecFetchSlotHeapTuple(slot, true, &shouldFree);
@@ -703,11 +703,11 @@ heapam_relation_copy_for_cluster(Relation OldHeap, Relation NewHeap,
         */
        if (OldIndex != NULL && !use_sort)
        {
-               const int   ci_index[] = {
+               const int       ci_index[] = {
                        PROGRESS_CLUSTER_PHASE,
                        PROGRESS_CLUSTER_INDEX_RELID
                };
-               int64       ci_val[2];
+               int64           ci_val[2];
 
                /* Set phase and OIDOldIndex to columns */
                ci_val[0] = PROGRESS_CLUSTER_PHASE_INDEX_SCAN_HEAP;
@@ -765,7 +765,10 @@ heapam_relation_copy_for_cluster(Relation OldHeap, Relation NewHeap,
                        if (!table_scan_getnextslot(tableScan, ForwardScanDirection, slot))
                                break;
 
-                       /* In scan-and-sort mode and also VACUUM FULL, set heap blocks scanned */
+                       /*
+                        * In scan-and-sort mode and also VACUUM FULL, set heap blocks
+                        * scanned
+                        */
                        pgstat_progress_update_param(PROGRESS_CLUSTER_HEAP_BLKS_SCANNED,
                                                                                 heapScan->rs_cblock + 1);
                }
@@ -844,17 +847,20 @@ heapam_relation_copy_for_cluster(Relation OldHeap, Relation NewHeap,
                {
                        tuplesort_putheaptuple(tuplesort, tuple);
 
-                       /* In scan-and-sort mode, report increase in number of tuples scanned */
+                       /*
+                        * In scan-and-sort mode, report increase in number of tuples
+                        * scanned
+                        */
                        pgstat_progress_update_param(PROGRESS_CLUSTER_HEAP_TUPLES_SCANNED,
                                                                                 *num_tuples);
                }
                else
                {
-                       const int   ct_index[] = {
+                       const int       ct_index[] = {
                                PROGRESS_CLUSTER_HEAP_TUPLES_SCANNED,
                                PROGRESS_CLUSTER_HEAP_TUPLES_WRITTEN
                        };
-                       int64       ct_val[2];
+                       int64           ct_val[2];
 
                        reform_and_rewrite_tuple(tuple, OldHeap, NewHeap,
                                                                         values, isnull, rwstate);
@@ -882,7 +888,8 @@ heapam_relation_copy_for_cluster(Relation OldHeap, Relation NewHeap,
         */
        if (tuplesort != NULL)
        {
-               double n_tuples = 0;
+               double          n_tuples = 0;
+
                /* Report that we are now sorting tuples */
                pgstat_progress_update_param(PROGRESS_CLUSTER_PHASE,
                                                                         PROGRESS_CLUSTER_PHASE_SORT_TUPLES);
@@ -925,10 +932,10 @@ heapam_relation_copy_for_cluster(Relation OldHeap, Relation NewHeap,
 }
 
 static bool
-heapam_scan_analyze_next_block(TableScanDesc sscan, BlockNumber blockno,
+heapam_scan_analyze_next_block(TableScanDesc scan, BlockNumber blockno,
                                                           BufferAccessStrategy bstrategy)
 {
-       HeapScanDesc scan = (HeapScanDesc) sscan;
+       HeapScanDesc hscan = (HeapScanDesc) scan;
 
        /*
         * We must maintain a pin on the target page's buffer to ensure that
@@ -939,22 +946,22 @@ heapam_scan_analyze_next_block(TableScanDesc sscan, BlockNumber blockno,
         * doing much work per tuple, the extra lock traffic is probably better
         * avoided.
         */
-       scan->rs_cblock = blockno;
-       scan->rs_cindex = FirstOffsetNumber;
-       scan->rs_cbuf = ReadBufferExtended(scan->rs_base.rs_rd, MAIN_FORKNUM,
-                                                                          blockno, RBM_NORMAL, bstrategy);
-       LockBuffer(scan->rs_cbuf, BUFFER_LOCK_SHARE);
+       hscan->rs_cblock = blockno;
+       hscan->rs_cindex = FirstOffsetNumber;
+       hscan->rs_cbuf = ReadBufferExtended(scan->rs_rd, MAIN_FORKNUM,
+                                                                               blockno, RBM_NORMAL, bstrategy);
+       LockBuffer(hscan->rs_cbuf, BUFFER_LOCK_SHARE);
 
        /* in heap all blocks can contain tuples, so always return true */
        return true;
 }
 
 static bool
-heapam_scan_analyze_next_tuple(TableScanDesc sscan, TransactionId OldestXmin,
+heapam_scan_analyze_next_tuple(TableScanDesc scan, TransactionId OldestXmin,
                                                           double *liverows, double *deadrows,
                                                           TupleTableSlot *slot)
 {
-       HeapScanDesc scan = (HeapScanDesc) sscan;
+       HeapScanDesc hscan = (HeapScanDesc) scan;
        Page            targpage;
        OffsetNumber maxoffset;
        BufferHeapTupleTableSlot *hslot;
@@ -962,17 +969,17 @@ heapam_scan_analyze_next_tuple(TableScanDesc sscan, TransactionId OldestXmin,
        Assert(TTS_IS_BUFFERTUPLE(slot));
 
        hslot = (BufferHeapTupleTableSlot *) slot;
-       targpage = BufferGetPage(scan->rs_cbuf);
+       targpage = BufferGetPage(hscan->rs_cbuf);
        maxoffset = PageGetMaxOffsetNumber(targpage);
 
        /* Inner loop over all tuples on the selected page */
-       for (; scan->rs_cindex <= maxoffset; scan->rs_cindex++)
+       for (; hscan->rs_cindex <= maxoffset; hscan->rs_cindex++)
        {
                ItemId          itemid;
                HeapTuple       targtuple = &hslot->base.tupdata;
                bool            sample_it = false;
 
-               itemid = PageGetItemId(targpage, scan->rs_cindex);
+               itemid = PageGetItemId(targpage, hscan->rs_cindex);
 
                /*
                 * We ignore unused and redirect line pointers.  DEAD line pointers
@@ -987,13 +994,14 @@ heapam_scan_analyze_next_tuple(TableScanDesc sscan, TransactionId OldestXmin,
                        continue;
                }
 
-               ItemPointerSet(&targtuple->t_self, scan->rs_cblock, scan->rs_cindex);
+               ItemPointerSet(&targtuple->t_self, hscan->rs_cblock, hscan->rs_cindex);
 
-               targtuple->t_tableOid = RelationGetRelid(scan->rs_base.rs_rd);
+               targtuple->t_tableOid = RelationGetRelid(scan->rs_rd);
                targtuple->t_data = (HeapTupleHeader) PageGetItem(targpage, itemid);
                targtuple->t_len = ItemIdGetLength(itemid);
 
-               switch (HeapTupleSatisfiesVacuum(targtuple, OldestXmin, scan->rs_cbuf))
+               switch (HeapTupleSatisfiesVacuum(targtuple, OldestXmin,
+                                                                                hscan->rs_cbuf))
                {
                        case HEAPTUPLE_LIVE:
                                sample_it = true;
@@ -1073,8 +1081,8 @@ heapam_scan_analyze_next_tuple(TableScanDesc sscan, TransactionId OldestXmin,
 
                if (sample_it)
                {
-                       ExecStoreBufferHeapTuple(targtuple, slot, scan->rs_cbuf);
-                       scan->rs_cindex++;
+                       ExecStoreBufferHeapTuple(targtuple, slot, hscan->rs_cbuf);
+                       hscan->rs_cindex++;
 
                        /* note that we leave the buffer locked here! */
                        return true;
@@ -1082,8 +1090,8 @@ heapam_scan_analyze_next_tuple(TableScanDesc sscan, TransactionId OldestXmin,
        }
 
        /* Now release the lock and pin on the page */
-       UnlockReleaseBuffer(scan->rs_cbuf);
-       scan->rs_cbuf = InvalidBuffer;
+       UnlockReleaseBuffer(hscan->rs_cbuf);
+       hscan->rs_cbuf = InvalidBuffer;
 
        /* also prevent old slot contents from having pin on page */
        ExecClearTuple(slot);
@@ -1290,9 +1298,10 @@ heapam_index_build_range_scan(Relation heapRelation,
 
                        /*
                         * The criteria for counting a tuple as live in this block need to
-                        * match what analyze.c's acquire_sample_rows() does, otherwise
-                        * CREATE INDEX and ANALYZE may produce wildly different reltuples
-                        * values, e.g. when there are many recently-dead tuples.
+                        * match what analyze.c's heapam_scan_analyze_next_tuple() does,
+                        * otherwise CREATE INDEX and ANALYZE may produce wildly different
+                        * reltuples values, e.g. when there are many recently-dead
+                        * tuples.
                         */
                        switch (HeapTupleSatisfiesVacuum(heapTuple, OldestXmin,
                                                                                         hscan->rs_cbuf))
@@ -1323,7 +1332,7 @@ heapam_index_build_range_scan(Relation heapRelation,
                                         * index as unusable for them.
                                         *
                                         * We don't count recently-dead tuples in reltuples, even
-                                        * if we index them; see acquire_sample_rows().
+                                        * if we index them; see heapam_scan_analyze_next_tuple().
                                         */
                                        if (HeapTupleIsHotUpdated(heapTuple))
                                        {
@@ -1387,7 +1396,8 @@ heapam_index_build_range_scan(Relation heapRelation,
                                        else
                                        {
                                                /*
-                                                * For consistency with acquire_sample_rows(), count
+                                                * For consistency with
+                                                * heapam_scan_analyze_next_tuple(), count
                                                 * HEAPTUPLE_INSERT_IN_PROGRESS tuples as live only
                                                 * when inserted by our own transaction.
                                                 */
@@ -1460,8 +1470,9 @@ heapam_index_build_range_scan(Relation heapRelation,
                                                /*
                                                 * Count HEAPTUPLE_DELETE_IN_PROGRESS tuples as live,
                                                 * if they were not deleted by the current
-                                                * transaction.  That's what acquire_sample_rows()
-                                                * does, and we want the behavior to be consistent.
+                                                * transaction.  That's what
+                                                * heapam_scan_analyze_next_tuple() does, and we want
+                                                * the behavior to be consistent.
                                                 */
                                                reltuples += 1;
                                        }
@@ -1596,7 +1607,7 @@ heapam_index_validate_scan(Relation heapRelation,
                                                   Relation indexRelation,
                                                   IndexInfo *indexInfo,
                                                   Snapshot snapshot,
-                                                  ValidateIndexState * state)
+                                                  ValidateIndexState *state)
 {
        TableScanDesc scan;
        HeapScanDesc hscan;
@@ -1824,55 +1835,6 @@ heapam_index_validate_scan(Relation heapRelation,
 }
 
 
-/* ----------------------------------------------------------------------------
- *  Helper functions for the above.
- * ----------------------------------------------------------------------------
- */
-
-/*
- * Reconstruct and rewrite the given tuple
- *
- * We cannot simply copy the tuple as-is, for several reasons:
- *
- * 1. We'd like to squeeze out the values of any dropped columns, both
- * to save space and to ensure we have no corner-case failures. (It's
- * possible for example that the new table hasn't got a TOAST table
- * and so is unable to store any large values of dropped cols.)
- *
- * 2. The tuple might not even be legal for the new table; this is
- * currently only known to happen as an after-effect of ALTER TABLE
- * SET WITHOUT OIDS.
- *
- * So, we must reconstruct the tuple from component Datums.
- */
-static void
-reform_and_rewrite_tuple(HeapTuple tuple,
-                                                Relation OldHeap, Relation NewHeap,
-                                                Datum *values, bool *isnull, RewriteState rwstate)
-{
-       TupleDesc       oldTupDesc = RelationGetDescr(OldHeap);
-       TupleDesc       newTupDesc = RelationGetDescr(NewHeap);
-       HeapTuple       copiedTuple;
-       int                     i;
-
-       heap_deform_tuple(tuple, oldTupDesc, values, isnull);
-
-       /* Be sure to null out any dropped columns */
-       for (i = 0; i < newTupDesc->natts; i++)
-       {
-               if (TupleDescAttr(newTupDesc, i)->attisdropped)
-                       isnull[i] = true;
-       }
-
-       copiedTuple = heap_form_tuple(newTupDesc, values, isnull);
-
-       /* The heap rewrite module does the rest */
-       rewrite_heap_tuple(rwstate, tuple, copiedTuple);
-
-       heap_freetuple(copiedTuple);
-}
-
-
 /* ------------------------------------------------------------------------
  * Planner related callbacks for the heap AM
  * ------------------------------------------------------------------------
@@ -1981,6 +1943,55 @@ heapam_estimate_rel_size(Relation rel, int32 *attr_widths,
 }
 
 
+/* ----------------------------------------------------------------------------
+ *  Helper functions for the above.
+ * ----------------------------------------------------------------------------
+ */
+
+/*
+ * Reconstruct and rewrite the given tuple
+ *
+ * We cannot simply copy the tuple as-is, for several reasons:
+ *
+ * 1. We'd like to squeeze out the values of any dropped columns, both
+ * to save space and to ensure we have no corner-case failures. (It's
+ * possible for example that the new table hasn't got a TOAST table
+ * and so is unable to store any large values of dropped cols.)
+ *
+ * 2. The tuple might not even be legal for the new table; this is
+ * currently only known to happen as an after-effect of ALTER TABLE
+ * SET WITHOUT OIDS.
+ *
+ * So, we must reconstruct the tuple from component Datums.
+ */
+static void
+reform_and_rewrite_tuple(HeapTuple tuple,
+                                                Relation OldHeap, Relation NewHeap,
+                                                Datum *values, bool *isnull, RewriteState rwstate)
+{
+       TupleDesc       oldTupDesc = RelationGetDescr(OldHeap);
+       TupleDesc       newTupDesc = RelationGetDescr(NewHeap);
+       HeapTuple       copiedTuple;
+       int                     i;
+
+       heap_deform_tuple(tuple, oldTupDesc, values, isnull);
+
+       /* Be sure to null out any dropped columns */
+       for (i = 0; i < newTupDesc->natts; i++)
+       {
+               if (TupleDescAttr(newTupDesc, i)->attisdropped)
+                       isnull[i] = true;
+       }
+
+       copiedTuple = heap_form_tuple(newTupDesc, values, isnull);
+
+       /* The heap rewrite module does the rest */
+       rewrite_heap_tuple(rwstate, tuple, copiedTuple);
+
+       heap_freetuple(copiedTuple);
+}
+
+
 /* ------------------------------------------------------------------------
  * Definition of the heap table access method.
  * ------------------------------------------------------------------------
index dd5f54eb2d27c2da752048729fb5ea81374131b8..0a85d8b535a52dae96ddf387c5fd75c61ad0a206 100644 (file)
@@ -15,7 +15,6 @@
 #include "postgres.h"
 
 #include "access/bufmask.h"
-#include "access/heapam_xlog.h"
 #include "access/nbtree.h"
 #include "access/nbtxlog.h"
 #include "access/transam.h"
index 8ad4c6294324a0df2e9ae0781b039965f286c3f4..baba1ea699bfca830901763f5cf4dc851aef7b7f 100644 (file)
@@ -96,8 +96,9 @@ table_beginscan_catalog(Relation relation, int nkeys, struct ScanKeyData *key)
        Oid                     relid = RelationGetRelid(relation);
        Snapshot        snapshot = RegisterSnapshot(GetCatalogSnapshot(relid));
 
-       return relation->rd_tableam->scan_begin(relation, snapshot, nkeys, key, NULL,
-                                                                                       true, true, true, false, false, true);
+       return relation->rd_tableam->scan_begin(relation, snapshot, nkeys, key,
+                                                                                       NULL, true, true, true, false,
+                                                                                       false, true);
 }
 
 void
@@ -171,8 +172,10 @@ table_beginscan_parallel(Relation relation, ParallelTableScanDesc parallel_scan)
                snapshot = SnapshotAny;
        }
 
-       return relation->rd_tableam->scan_begin(relation, snapshot, 0, NULL, parallel_scan,
-                                                                                       true, true, true, false, false, !parallel_scan->phs_snapshot_any);
+       return relation->rd_tableam->scan_begin(relation, snapshot, 0, NULL,
+                                                                                       parallel_scan, true, true, true,
+                                                                                       false, false,
+                                                                                       !parallel_scan->phs_snapshot_any);
 }
 
 
index e1817a612f051f45aec310a41940896f32f173ab..96660bf907f979a99a33639dfeb9d000ba279009 100644 (file)
@@ -78,6 +78,17 @@ GetTableAmRoutine(Oid amhandler)
        Assert(routine->tuple_update != NULL);
        Assert(routine->tuple_lock != NULL);
 
+       Assert(routine->relation_set_new_filenode != NULL);
+       Assert(routine->relation_nontransactional_truncate != NULL);
+       Assert(routine->relation_copy_data != NULL);
+       Assert(routine->relation_copy_for_cluster != NULL);
+       Assert(routine->relation_vacuum != NULL);
+       Assert(routine->scan_analyze_next_block != NULL);
+       Assert(routine->scan_analyze_next_tuple != NULL);
+       Assert(routine->index_build_range_scan != NULL);
+       Assert(routine->index_validate_scan != NULL);
+       Assert(routine->relation_estimate_size != NULL);
+
        return routine;
 }
 
index 89f909bceb36df9fb235119eeb37908f7d06913e..74867203c7ef9ee976c2d667c6ea374081bc0adb 100644 (file)
@@ -2068,9 +2068,9 @@ bool
 has_stored_generated_columns(PlannerInfo *root, Index rti)
 {
        RangeTblEntry *rte = planner_rt_fetch(rti, root);
-       Relation    relation;
+       Relation        relation;
        TupleDesc       tupdesc;
-       bool        result = false;
+       bool            result = false;
 
        /* Assume we already have adequate lock */
        relation = heap_open(rte->relid, NoLock);
index 873ad1531341d8df2e4496eea91f0dc63865da5a..0151f17af42e544f2133ffe9a8b84fd17603c9cd 100644 (file)
@@ -192,14 +192,16 @@ typedef struct TableAmRoutine
         * Restart relation scan.  If set_params is set to true, allow{strat,
         * sync, pagemode} (see scan_begin) changes should be taken into account.
         */
-       void            (*scan_rescan) (TableScanDesc scan, struct ScanKeyData *key, bool set_params,
-                                                               bool allow_strat, bool allow_sync, bool allow_pagemode);
+       void            (*scan_rescan) (TableScanDesc scan, struct ScanKeyData *key,
+                                                               bool set_params, bool allow_strat,
+                                                               bool allow_sync, bool allow_pagemode);
 
        /*
         * Return next tuple from `scan`, store in slot.
         */
        bool            (*scan_getnextslot) (TableScanDesc scan,
-                                                                        ScanDirection direction, TupleTableSlot *slot);
+                                                                        ScanDirection direction,
+                                                                        TupleTableSlot *slot);
 
 
        /* ------------------------------------------------------------------------
@@ -218,13 +220,15 @@ typedef struct TableAmRoutine
         * pscan will be sized according to parallelscan_estimate() for the same
         * relation.
         */
-       Size            (*parallelscan_initialize) (Relation rel, ParallelTableScanDesc pscan);
+       Size            (*parallelscan_initialize) (Relation rel,
+                                                                                       ParallelTableScanDesc pscan);
 
        /*
         * Reinitialize `pscan` for a new scan. `rel` will be the same relation as
         * when `pscan` was initialized by parallelscan_initialize.
         */
-       void            (*parallelscan_reinitialize) (Relation rel, ParallelTableScanDesc pscan);
+       void            (*parallelscan_reinitialize) (Relation rel,
+                                                                                         ParallelTableScanDesc pscan);
 
 
        /* ------------------------------------------------------------------------
@@ -323,8 +327,9 @@ typedef struct TableAmRoutine
         */
 
        /* see table_insert() for reference about parameters */
-       void            (*tuple_insert) (Relation rel, TupleTableSlot *slot, CommandId cid,
-                                                                int options, struct BulkInsertStateData *bistate);
+       void            (*tuple_insert) (Relation rel, TupleTableSlot *slot,
+                                                                CommandId cid, int options,
+                                                                struct BulkInsertStateData *bistate);
 
        /* see table_insert_speculative() for reference about parameters */
        void            (*tuple_insert_speculative) (Relation rel,
@@ -414,10 +419,16 @@ typedef struct TableAmRoutine
        void            (*relation_copy_data) (Relation rel, RelFileNode newrnode);
 
        /* See table_relation_copy_for_cluster() */
-       void            (*relation_copy_for_cluster) (Relation NewHeap, Relation OldHeap, Relation OldIndex,
+       void            (*relation_copy_for_cluster) (Relation NewHeap,
+                                                                                         Relation OldHeap,
+                                                                                         Relation OldIndex,
                                                                                          bool use_sort,
-                                                                                         TransactionId OldestXmin, TransactionId FreezeXid, MultiXactId MultiXactCutoff,
-                                                                                         double *num_tuples, double *tups_vacuumed, double *tups_recently_dead);
+                                                                                         TransactionId OldestXmin,
+                                                                                         TransactionId FreezeXid,
+                                                                                         MultiXactId MultiXactCutoff,
+                                                                                         double *num_tuples,
+                                                                                         double *tups_vacuumed,
+                                                                                         double *tups_recently_dead);
 
        /*
         * React to VACUUM command on the relation. The VACUUM might be user
@@ -434,7 +445,8 @@ typedef struct TableAmRoutine
         * There probably, in the future, needs to be a separate callback to
         * integrate with autovacuum's scheduling.
         */
-       void            (*relation_vacuum) (Relation onerel, struct VacuumParams *params,
+       void            (*relation_vacuum) (Relation onerel,
+                                                                       struct VacuumParams *params,
                                                                        BufferAccessStrategy bstrategy);
 
        /*
@@ -595,7 +607,8 @@ table_beginscan_bm(Relation rel, Snapshot snapshot,
 static inline TableScanDesc
 table_beginscan_sampling(Relation rel, Snapshot snapshot,
                                                 int nkeys, struct ScanKeyData *key,
-                                                bool allow_strat, bool allow_sync, bool allow_pagemode)
+                                                bool allow_strat, bool allow_sync,
+                                                bool allow_pagemode)
 {
        return rel->rd_tableam->scan_begin(rel, snapshot, nkeys, key, NULL,
                                                                           allow_strat, allow_sync, allow_pagemode,
@@ -686,7 +699,9 @@ extern Size table_parallelscan_estimate(Relation rel, Snapshot snapshot);
  * for the same relation.  Call this just once in the leader process; then,
  * individual workers attach via table_beginscan_parallel.
  */
-extern void table_parallelscan_initialize(Relation rel, ParallelTableScanDesc pscan, Snapshot snapshot);
+extern void table_parallelscan_initialize(Relation rel,
+                                                         ParallelTableScanDesc pscan,
+                                                         Snapshot snapshot);
 
 /*
  * Begin a parallel scan. `pscan` needs to have been initialized with
@@ -695,7 +710,8 @@ extern void table_parallelscan_initialize(Relation rel, ParallelTableScanDesc ps
  *
  * Caller must hold a suitable lock on the correct relation.
  */
-extern TableScanDesc table_beginscan_parallel(Relation rel, ParallelTableScanDesc pscan);
+extern TableScanDesc table_beginscan_parallel(Relation rel,
+                                                ParallelTableScanDesc pscan);
 
 /*
  * Restart a parallel scan.  Call this in the leader process.  Caller is
@@ -836,7 +852,8 @@ table_get_latest_tid(Relation rel, Snapshot snapshot, ItemPointer tid)
  * they ought to mark the relevant buffer dirty.
  */
 static inline bool
-table_tuple_satisfies_snapshot(Relation rel, TupleTableSlot *slot, Snapshot snapshot)
+table_tuple_satisfies_snapshot(Relation rel, TupleTableSlot *slot,
+                                                          Snapshot snapshot)
 {
        return rel->rd_tableam->tuple_satisfies_snapshot(rel, slot, snapshot);
 }
@@ -919,7 +936,8 @@ table_insert(Relation rel, TupleTableSlot *slot, CommandId cid,
  */
 static inline void
 table_insert_speculative(Relation rel, TupleTableSlot *slot, CommandId cid,
-                                                int options, struct BulkInsertStateData *bistate, uint32 specToken)
+                                                int options, struct BulkInsertStateData *bistate,
+                                                uint32 specToken)
 {
        rel->rd_tableam->tuple_insert_speculative(rel, slot, cid, options,
                                                                                          bistate, specToken);
@@ -930,8 +948,8 @@ table_insert_speculative(Relation rel, TupleTableSlot *slot, CommandId cid,
  * succeeded is true, the tuple is fully inserted, if false, it's removed.
  */
 static inline void
-table_complete_speculative(Relation rel, TupleTableSlot *slot, uint32 specToken,
-                                                  bool succeeded)
+table_complete_speculative(Relation rel, TupleTableSlot *slot,
+                                                  uint32 specToken, bool succeeded)
 {
        rel->rd_tableam->tuple_complete_speculative(rel, slot, specToken,
                                                                                                succeeded);
@@ -1342,9 +1360,12 @@ extern void simple_table_update(Relation rel, ItemPointer otid,
 extern Size table_block_parallelscan_estimate(Relation rel);
 extern Size table_block_parallelscan_initialize(Relation rel,
                                                                        ParallelTableScanDesc pscan);
-extern void table_block_parallelscan_reinitialize(Relation rel, ParallelTableScanDesc pscan);
-extern BlockNumber table_block_parallelscan_nextpage(Relation rel, ParallelBlockTableScanDesc pbscan);
-extern void table_block_parallelscan_startblock_init(Relation rel, ParallelBlockTableScanDesc pbscan);
+extern void table_block_parallelscan_reinitialize(Relation rel,
+                                                                         ParallelTableScanDesc pscan);
+extern BlockNumber table_block_parallelscan_nextpage(Relation rel,
+                                                                 ParallelBlockTableScanDesc pbscan);
+extern void table_block_parallelscan_startblock_init(Relation rel,
+                                                                                ParallelBlockTableScanDesc pbscan);
 
 
 /* ----------------------------------------------------------------------------