diff --git a/src/runner.c b/src/runner.c
index c08c425a7bed76d0cdc77593079af0d184b7d265..c50c5363a4c46b21ba0f68d2cd9dfcd9d39f4274 100644
--- a/src/runner.c
+++ b/src/runner.c
@@ -1889,62 +1889,6 @@ void *runner_main(void *data) {
 /* Check that we haven't scheduled an inactive task */
 #ifdef SWIFT_DEBUG_CHECKS
       t->ti_run = e->ti_current;
-/* #ifndef WITH_MPI */
-/*       if (t->type == task_type_grav_top_level) { */
-/*         if (ci != NULL || cj != NULL) */
-/*           error("Top-level gravity task associated with a cell"); */
-/*       } else if (ci == NULL && cj == NULL) { */
-
-/*         error("Task not associated with cells!"); */
-/*       } else if (cj == NULL) { /\* self *\/ */
-
-/*         if (!cell_is_active_hydro(ci, e) && t->type != task_type_sort && */
-/*             t->type != task_type_send && t->type != task_type_recv && */
-/*             t->type != task_type_kick1 && t->type != task_type_drift_part &&
- */
-/*             t->type != task_type_drift_gpart) */
-/*           error( */
-/*               "Task (type='%s/%s') should have been skipped ti_current=%lld "
- */
-/*               "c->ti_end_min=%lld", */
-/*               taskID_names[t->type], subtaskID_names[t->subtype],
- * e->ti_current, */
-/*               ci->ti_end_min); */
-
-/*         /\* Special case for sorts *\/ */
-/*         if (!cell_is_active_hydro(ci, e) && t->type == task_type_sort && */
-/*             !(ci->do_sort || ci->do_sub_sort)) */
-/*           error( */
-/*               "Task (type='%s/%s') should have been skipped ti_current=%lld "
- */
-/*               "c->ti_end_min=%lld t->flags=%d", */
-/*               taskID_names[t->type], subtaskID_names[t->subtype],
- * e->ti_current, */
-/*               ci->ti_end_min, t->flags); */
-
-/*         /\* Special case for kick1 *\/ */
-/*         if (!cell_is_starting(ci, e) && t->type == task_type_kick1 && */
-/*             t->flags == 0) */
-/*           error( */
-/*               "Task (type='%s/%s') should have been skipped ti_current=%lld "
- */
-/*               "c->ti_end_min=%lld t->flags=%d", */
-/*               taskID_names[t->type], subtaskID_names[t->subtype],
- * e->ti_current, */
-/*               ci->ti_end_min, t->flags); */
-
-/*       } else { /\* pair *\/ */
-/*         if (!cell_is_active_hydro(ci, e) && !cell_is_active_hydro(cj, e)) */
-
-/*           if (t->type != task_type_send && t->type != task_type_recv) */
-/*             error( */
-/*                 "Task (type='%s/%s') should have been skipped ti_current=%lld
- * " */
-/*                 "ci->ti_end_min=%lld cj->ti_end_min=%lld", */
-/*                 taskID_names[t->type], subtaskID_names[t->subtype], */
-/*                 e->ti_current, ci->ti_end_min, cj->ti_end_min); */
-/*       } */
-/* #endif */
 #endif
 
       /* Different types of tasks... */
diff --git a/src/scheduler.c b/src/scheduler.c
index fd283b0f25bf1837ddd52a9b1c49815b2b339d4c..0032196b437f533df35286a83076bedb09712952 100644
--- a/src/scheduler.c
+++ b/src/scheduler.c
@@ -1374,83 +1374,6 @@ void scheduler_start(struct scheduler *s) {
     scheduler_rewait_mapper(s->tid_active, s->active_count, s);
   }
 
-  /* Check we have not missed an active task */
-  /* #ifdef SWIFT_DEBUG_CHECKS */
-
-  /*   const integertime_t ti_current = s->space->e->ti_current; */
-
-  /*   if (ti_current > 0) { */
-
-  /*     for (int k = 0; k < s->nr_tasks; k++) { */
-
-  /*       struct task *t = &s->tasks[k]; */
-  /*       struct cell *ci = t->ci; */
-  /*       struct cell *cj = t->cj; */
-
-  /*       if (t->type == task_type_none) continue; */
-
-  /*       /\* Don't check MPI stuff *\/ */
-  /*       if (t->type == task_type_send || t->type == task_type_recv) continue;
-   */
-
-  /*       /\* Don't check the FFT task *\/ */
-  /*       if (t->type == task_type_grav_top_level || */
-  /*           t->type == task_type_grav_ghost_in || */
-  /*           t->type == task_type_grav_ghost_out) */
-  /*         continue; */
-
-  /*       if (ci == NULL && cj == NULL) { */
-
-  /*         error("Task not associated with cells!"); */
-
-  /*       } else if (cj == NULL) { /\* self *\/ */
-
-  /*         if (ci->ti_hydro_end_min == ti_current && t->skip && */
-  /*             t->type != task_type_sort && t->type != task_type_drift_part &&
-   */
-  /*             t->type != task_type_drift_gpart) */
-  /*           error( */
-  /*               "Task (type='%s/%s') should not have been skipped " */
-  /*               "ti_current=%lld " */
-  /*               "c->ti_hydro_end_min=%lld", */
-  /*               taskID_names[t->type], subtaskID_names[t->subtype],
-   * ti_current, */
-  /*               ci->ti_hydro_end_min); */
-
-  /*         /\* Special treatment for sort tasks *\/ */
-  /*         /\* if (ci->ti_end_min == ti_current && t->skip && */
-  /*             t->type == task_type_sort && t->flags == 0) */
-  /*           error( */
-  /*               "Task (type='%s/%s') should not have been skipped " */
-  /*               "ti_current=%lld " */
-  /*               "c->ti_end_min=%lld t->flags=%d", */
-  /*               taskID_names[t->type], subtaskID_names[t->subtype],
-   * ti_current, */
-  /*               ci->ti_end_min, t->flags); *\/ */
-
-  /*       } else { /\* pair *\/ */
-
-  /*         if (t->skip) { */
-
-  /*           /\* Check that the pair is active if the local cell is active *\/
-   */
-  /*           if ((ci->ti_hydro_end_min == ti_current && ci->nodeID ==
-   * engine_rank) || */
-  /*               (cj->ti_hydro_end_min == ti_current && cj->nodeID ==
-   * engine_rank)) */
-  /*             error( */
-  /*                 "Task (type='%s/%s') should not have been skipped " */
-  /*                 "ti_current=%lld " */
-  /*                 "ci->ti_hydro_end_min=%lld cj->ti_hydro_end_min=%lld", */
-  /*                 taskID_names[t->type], subtaskID_names[t->subtype],
-   * ti_current, */
-  /*                 ci->ti_hydro_end_min, cj->ti_hydro_end_min); */
-  /*         } */
-  /*       } */
-  /*     } */
-  /*   } */
-  /* #endif */
-
   /* Loop over the tasks and enqueue whoever is ready. */
   if (s->active_count > 1000) {
     threadpool_map(s->threadpool, scheduler_enqueue_mapper, s->tid_active,