From c7b124071fb0a2fc0369d4bbec363473fbc6e13a Mon Sep 17 00:00:00 2001
From: "Peter W. Draper" <p.w.draper@durham.ac.uk>
Date: Thu, 17 Jan 2019 21:01:00 +0000
Subject: [PATCH] Formatting

---
 examples/main.c |   5 +-
 src/engine.c    |  10 ++--
 src/partition.c | 150 ++++++++++++++++++++++++------------------------
 src/scheduler.c |   2 +-
 src/task.c      |   9 ++-
 src/task.h      |   3 +-
 6 files changed, 92 insertions(+), 87 deletions(-)

diff --git a/examples/main.c b/examples/main.c
index 9c71a6f6b4..0a96ca0f55 100644
--- a/examples/main.c
+++ b/examples/main.c
@@ -291,8 +291,9 @@ int main(int argc, char *argv[]) {
 #ifndef SWIFT_DEBUG_TASKS
   if (dump_tasks) {
     if (myrank == 0) {
-      message("WARNING: complete task dumps are only created when "
-              "configured with --enable-task-debugging.");
+      message(
+          "WARNING: complete task dumps are only created when "
+          "configured with --enable-task-debugging.");
       message("         Basic task statistics will be output.");
     }
   }
diff --git a/src/engine.c b/src/engine.c
index ea699ec419..ab2742b197 100644
--- a/src/engine.c
+++ b/src/engine.c
@@ -995,7 +995,8 @@ void engine_repartition(struct engine *e) {
 
   /* Generate the fixed costs include file. */
   if (e->step > 3 && e->reparttype->trigger <= 1.f) {
-    task_dump_stats("partition_fixed_costs.h", e, /* header = */ 1, /* allranks = */ 1);
+    task_dump_stats("partition_fixed_costs.h", e, /* header = */ 1,
+                    /* allranks = */ 1);
   }
 
   /* Do the repartitioning. */
@@ -1056,7 +1057,8 @@ void engine_repartition_trigger(struct engine *e) {
   /* Do nothing if there have not been enough steps since the last repartition
    * as we don't want to repeat this too often or immediately after a
    * repartition step. Also nothing to do when requested. */
-  if (e->step - e->last_repartition >= 2 && e->reparttype->type != REPART_NONE) {
+  if (e->step - e->last_repartition >= 2 &&
+      e->reparttype->type != REPART_NONE) {
 
     /* If we have fixed costs available and this is step 2 or we are forcing
      * repartitioning then we do a fixed costs one now. */
@@ -1072,7 +1074,6 @@ void engine_repartition_trigger(struct engine *e) {
 
     } else {
 
-
       /* It is only worth checking the CPU loads when we have processed a
        * significant number of all particles as we require all tasks to have
        * timings. */
@@ -1089,7 +1090,8 @@ void engine_repartition_trigger(struct engine *e) {
         }
 
         /* Get CPU time used since the last call to this function. */
-        double elapsed_cputime = clocks_get_cputime_used() - e->cputime_last_step;
+        double elapsed_cputime =
+            clocks_get_cputime_used() - e->cputime_last_step;
 
         /* Gather the elapsed CPU times from all ranks for the last step. */
         double elapsed_cputimes[e->nr_nodes];
diff --git a/src/partition.c b/src/partition.c
index 25ab31e5be..d05057649a 100644
--- a/src/partition.c
+++ b/src/partition.c
@@ -1215,7 +1215,8 @@ static void partition_gather_weights(void *map_data, int num_elements,
 
     /* Skip un-interesting tasks. */
     if (t->type == task_type_send || t->type == task_type_recv ||
-        t->type == task_type_logger || t->implicit || t->ci == NULL) continue;
+        t->type == task_type_logger || t->implicit || t->ci == NULL)
+      continue;
 
     /* Get weight for this task. Either based on fixed costs or task timings. */
     double w = 0.0;
@@ -1545,91 +1546,90 @@ static void repart_edge_metis(int vweights, int eweights, int timebins,
 static void repart_memory_metis(struct repartition *repartition, int nodeID,
                                 int nr_nodes, struct space *s) {
 
-    /* Space for counts of particle memory use per cell. */
-    double *weights = NULL;
-    if ((weights = (double *)malloc(sizeof(double) * s->nr_cells)) == NULL)
-        error("Failed to allocate cell weights buffer.");
-    bzero(weights, sizeof(double) * s->nr_cells);
+  /* Space for counts of particle memory use per cell. */
+  double *weights = NULL;
+  if ((weights = (double *)malloc(sizeof(double) * s->nr_cells)) == NULL)
+    error("Failed to allocate cell weights buffer.");
+  bzero(weights, sizeof(double) * s->nr_cells);
 
-    /* Check each particle and accumulate the sizes per cell. */
-    accumulate_sizes(s, weights);
+  /* Check each particle and accumulate the sizes per cell. */
+  accumulate_sizes(s, weights);
 
-    /* Get all the counts from all the nodes. */
-    if (MPI_Allreduce(MPI_IN_PLACE, weights, s->nr_cells, MPI_DOUBLE, MPI_SUM,
-                      MPI_COMM_WORLD) != MPI_SUCCESS)
-        error("Failed to allreduce particle cell weights.");
+  /* Get all the counts from all the nodes. */
+  if (MPI_Allreduce(MPI_IN_PLACE, weights, s->nr_cells, MPI_DOUBLE, MPI_SUM,
+                    MPI_COMM_WORLD) != MPI_SUCCESS)
+    error("Failed to allreduce particle cell weights.");
 
     /* Allocate cell list for the partition. If not already done. */
 #ifdef HAVE_PARMETIS
-    int refine = 1;
+  int refine = 1;
 #endif
-    if (repartition->ncelllist != s->nr_cells) {
+  if (repartition->ncelllist != s->nr_cells) {
 #ifdef HAVE_PARMETIS
-        refine = 0;
+    refine = 0;
 #endif
-        free(repartition->celllist);
-        repartition->ncelllist = 0;
-        if ((repartition->celllist = (int *)malloc(sizeof(int) * s->nr_cells)) == NULL)
-            error("Failed to allocate celllist");
-        repartition->ncelllist = s->nr_cells;
-    }
+    free(repartition->celllist);
+    repartition->ncelllist = 0;
+    if ((repartition->celllist = (int *)malloc(sizeof(int) * s->nr_cells)) ==
+        NULL)
+      error("Failed to allocate celllist");
+    repartition->ncelllist = s->nr_cells;
+  }
 
-    /* We need to rescale the sum of the weights so that the sum is
-     * less than IDX_MAX, that is the range of idx_t. */
-    double sum = 0.0;
-    for (int k = 0; k < s->nr_cells; k++) sum += weights[k];
-    if (sum > (double)IDX_MAX) {
-        double scale = (double)(IDX_MAX - 1000) / sum;
-        for (int k = 0; k < s->nr_cells; k++) weights[k] *= scale;
-    }
+  /* We need to rescale the sum of the weights so that the sum is
+   * less than IDX_MAX, that is the range of idx_t. */
+  double sum = 0.0;
+  for (int k = 0; k < s->nr_cells; k++) sum += weights[k];
+  if (sum > (double)IDX_MAX) {
+    double scale = (double)(IDX_MAX - 1000) / sum;
+    for (int k = 0; k < s->nr_cells; k++) weights[k] *= scale;
+  }
 
     /* And repartition. */
 #ifdef HAVE_PARMETIS
-    if (repartition_partition->usemetis) {
-        pick_metis(nodeID, s, nr_nodes, weights, NULL, reparition->celllist);
-    } else {
-        pick_parmetis(nodeID, s, nr_nodes, weights, NULL, refine,
-                      repartition->adaptive, repartition->itr,
-                      repartition->celllist);
-    }
+  if (repartition_partition->usemetis) {
+    pick_metis(nodeID, s, nr_nodes, weights, NULL, reparition->celllist);
+  } else {
+    pick_parmetis(nodeID, s, nr_nodes, weights, NULL, refine,
+                  repartition->adaptive, repartition->itr,
+                  repartition->celllist);
+  }
 #else
-    pick_metis(nodeID, s, nr_nodes, weights, NULL, repartition->celllist);
+  pick_metis(nodeID, s, nr_nodes, weights, NULL, repartition->celllist);
 #endif
 
-    /* Check that all cells have good values. All nodes have same copy, so just
-     * check on one. */
-    if (nodeID == 0) {
-        for (int k = 0; k < s->nr_cells; k++)
-            if (repartition->celllist[k] < 0 || repartition->celllist[k] >= nr_nodes)
-                error("Got bad nodeID %d for cell %i.", repartition->celllist[k], k);
-    }
-
-
-    /* Check that the partition is complete and all nodes have some cells. */
-    int present[nr_nodes];
-    int failed = 0;
-    for (int i = 0; i < nr_nodes; i++) present[i] = 0;
-    for (int i = 0; i < s->nr_cells; i++) present[repartition->celllist[i]]++;
-    for (int i = 0; i < nr_nodes; i++) {
-        if (!present[i]) {
-            failed = 1;
-            if (nodeID == 0) message("Node %d is not present after repartition", i);
-        }
-    }
+  /* Check that all cells have good values. All nodes have same copy, so just
+   * check on one. */
+  if (nodeID == 0) {
+    for (int k = 0; k < s->nr_cells; k++)
+      if (repartition->celllist[k] < 0 || repartition->celllist[k] >= nr_nodes)
+        error("Got bad nodeID %d for cell %i.", repartition->celllist[k], k);
+  }
 
-    /* If partition failed continue with the current one, but make this clear. */
-    if (failed) {
-        if (nodeID == 0)
-            message(
-                    "WARNING: repartition has failed, continuing with the current"
-                    " partition, load balance will not be optimal");
-        for (int k = 0; k < s->nr_cells; k++)
-            repartition->celllist[k] = s->cells_top[k].nodeID;
+  /* Check that the partition is complete and all nodes have some cells. */
+  int present[nr_nodes];
+  int failed = 0;
+  for (int i = 0; i < nr_nodes; i++) present[i] = 0;
+  for (int i = 0; i < s->nr_cells; i++) present[repartition->celllist[i]]++;
+  for (int i = 0; i < nr_nodes; i++) {
+    if (!present[i]) {
+      failed = 1;
+      if (nodeID == 0) message("Node %d is not present after repartition", i);
     }
+  }
 
-    /* And apply to our cells */
-    split_metis(s, nr_nodes, repartition->celllist);
+  /* If partition failed continue with the current one, but make this clear. */
+  if (failed) {
+    if (nodeID == 0)
+      message(
+          "WARNING: repartition has failed, continuing with the current"
+          " partition, load balance will not be optimal");
+    for (int k = 0; k < s->nr_cells; k++)
+      repartition->celllist[k] = s->cells_top[k].nodeID;
+  }
 
+  /* And apply to our cells */
+  split_metis(s, nr_nodes, repartition->celllist);
 }
 #endif /* WITH_MPI && (HAVE_METIS || HAVE_PARMETIS) */
 
@@ -1667,7 +1667,7 @@ void partition_repartition(struct repartition *reparttype, int nodeID,
                       nr_tasks);
 
   } else if (reparttype->type == REPART_METIS_VERTEX_COUNTS) {
-      repart_memory_metis(reparttype, nodeID, nr_nodes, s);
+    repart_memory_metis(reparttype, nodeID, nr_nodes, s);
 
   } else if (reparttype->type == REPART_NONE) {
     /* Doing nothing. */
@@ -1965,8 +1965,8 @@ void partition_init(struct partition *partition,
 
   /* Do we have fixed costs available? These can be used to force
    * repartitioning at any time. Not required if not repartitioning.*/
-  repartition->use_fixed_costs =
-      parser_get_opt_param_int(params, "DomainDecomposition:use_fixed_costs", 0);
+  repartition->use_fixed_costs = parser_get_opt_param_int(
+      params, "DomainDecomposition:use_fixed_costs", 0);
   if (repartition->type == REPART_NONE) repartition->use_fixed_costs = 0;
 
   /* Check if this is true or required and initialise them. */
@@ -1974,12 +1974,14 @@ void partition_init(struct partition *partition,
     if (!repart_init_fixed_costs()) {
       if (repartition->trigger <= 1) {
         if (engine_rank == 0)
-          message("WARNING: fixed cost repartitioning was requested but is"
-                  " not available.");
+          message(
+              "WARNING: fixed cost repartitioning was requested but is"
+              " not available.");
         repartition->use_fixed_costs = 0;
       } else {
-        error("Forced fixed cost repartitioning was requested but is"
-              " not available.");
+        error(
+            "Forced fixed cost repartitioning was requested but is"
+            " not available.");
       }
     }
   }
diff --git a/src/scheduler.c b/src/scheduler.c
index 06e07d4178..1083be1996 100644
--- a/src/scheduler.c
+++ b/src/scheduler.c
@@ -2171,7 +2171,7 @@ void scheduler_enqueue_mapper(void *map_data, int num_elements,
  */
 void scheduler_start(struct scheduler *s) {
 
-/* Reset all task timers. */
+  /* Reset all task timers. */
   for (int i = 0; i < s->nr_tasks; ++i) {
     s->tasks[i].tic = 0;
     s->tasks[i].toc = 0;
diff --git a/src/task.c b/src/task.c
index e7fbb6e339..d5e0a7d8dc 100644
--- a/src/task.c
+++ b/src/task.c
@@ -838,7 +838,8 @@ void task_dump_all(struct engine *e, int step) {
  * @param allranks do the statistics over all ranks, if not just the current
  *                 one, only used if header is false.
  */
-void task_dump_stats(const char *dumpfile, struct engine *e, int header, int allranks) {
+void task_dump_stats(const char *dumpfile, struct engine *e, int header,
+                     int allranks) {
 
   /* Need arrays for sum, min and max across all types and subtypes. */
   double sum[task_type_count][task_subtype_count];
@@ -878,7 +879,6 @@ void task_dump_stats(const char *dumpfile, struct engine *e, int header, int all
     }
   }
 
-
 #ifdef WITH_MPI
   if (allranks || header) {
     /* Get these from all ranks for output from rank 0. Could wrap these into a
@@ -934,8 +934,8 @@ void task_dump_stats(const char *dumpfile, struct engine *e, int header, int all
                     "%15s/%-10s %10d %14.4f %14.4f %14.4f %14.4f %14.4f %10d\n",
                     taskID, subtaskID_names[k], count[j][k],
                     clocks_from_ticks(min[j][k]), clocks_from_ticks(max[j][k]),
-                    clocks_from_ticks(sum[j][k]), clocks_from_ticks(mean),
-                    perc, fixed_cost);
+                    clocks_from_ticks(sum[j][k]), clocks_from_ticks(mean), perc,
+                    fixed_cost);
           }
         }
       }
@@ -944,5 +944,4 @@ void task_dump_stats(const char *dumpfile, struct engine *e, int header, int all
 #ifdef WITH_MPI
   }
 #endif
-
 }
diff --git a/src/task.h b/src/task.h
index d83e46069d..35b46bd383 100644
--- a/src/task.h
+++ b/src/task.h
@@ -201,7 +201,8 @@ int task_lock(struct task *t);
 void task_do_rewait(struct task *t);
 void task_print(const struct task *t);
 void task_dump_all(struct engine *e, int step);
-void task_dump_stats(const char *dumpfile, struct engine *e, int header, int allranks);
+void task_dump_stats(const char *dumpfile, struct engine *e, int header,
+                     int allranks);
 void task_get_full_name(int type, int subtype, char *name);
 void task_get_group_name(int type, int subtype, char *cluster);
 
-- 
GitLab