diff options
author | Tom Lane <tgl@sss.pgh.pa.us> | 2017-09-01 17:38:54 -0400 |
---|---|---|
committer | Tom Lane <tgl@sss.pgh.pa.us> | 2017-09-01 17:38:54 -0400 |
commit | 01edb5c7fc3bcf6aea15f2b3be36189b52ad9d1a (patch) | |
tree | ecdc6beba03ff3b1e032bdbcf16c63eff081c544 /src/include/executor/execParallel.h | |
parent | f2fe1cbef11c5fc962e338c8523667314faa6d89 (diff) |
Improve division of labor between execParallel.c and nodeGather[Merge].c.
Move the responsibility for creating/destroying TupleQueueReaders into
execParallel.c, to avoid duplicative coding in nodeGather.c and
nodeGatherMerge.c. Also, instead of having DestroyTupleQueueReader do
shm_mq_detach, do it in the caller (which is now only ExecParallelFinish).
This means execParallel.c does both the attaching and detaching of the
tuple-queue-reader shm_mqs, which seems less weird than the previous
arrangement.
These changes also eliminate a vestigial memory leak (of the pei->tqueue
array). It's now demonstrable that rescans of Gather or GatherMerge don't
leak memory.
Discussion: https://postgr.es/m/8670.1504192177@sss.pgh.pa.us
Diffstat (limited to 'src/include/executor/execParallel.h')
-rw-r--r-- | src/include/executor/execParallel.h | 18 |
1 files changed, 11 insertions, 7 deletions
diff --git a/src/include/executor/execParallel.h b/src/include/executor/execParallel.h index a6512245b15..8b714193c57 100644 --- a/src/include/executor/execParallel.h +++ b/src/include/executor/execParallel.h @@ -23,17 +23,21 @@ typedef struct SharedExecutorInstrumentation SharedExecutorInstrumentation; typedef struct ParallelExecutorInfo { - PlanState *planstate; - ParallelContext *pcxt; - BufferUsage *buffer_usage; - SharedExecutorInstrumentation *instrumentation; - shm_mq_handle **tqueue; - dsa_area *area; - bool finished; + PlanState *planstate; /* plan subtree we're running in parallel */ + ParallelContext *pcxt; /* parallel context we're using */ + BufferUsage *buffer_usage; /* points to bufusage area in DSM */ + SharedExecutorInstrumentation *instrumentation; /* optional */ + dsa_area *area; /* points to DSA area in DSM */ + bool finished; /* set true by ExecParallelFinish */ + /* These two arrays have pcxt->nworkers_launched entries: */ + shm_mq_handle **tqueue; /* tuple queues for worker output */ + struct TupleQueueReader **reader; /* tuple reader/writer support */ } ParallelExecutorInfo; extern ParallelExecutorInfo *ExecInitParallelPlan(PlanState *planstate, EState *estate, int nworkers); +extern void ExecParallelCreateReaders(ParallelExecutorInfo *pei, + TupleDesc tupDesc); extern void ExecParallelFinish(ParallelExecutorInfo *pei); extern void ExecParallelCleanup(ParallelExecutorInfo *pei); extern void ExecParallelReinitialize(PlanState *planstate, |