summaryrefslogtreecommitdiff
path: root/src/backend/optimizer/util/pathnode.c
diff options
context:
space:
mode:
Diffstat (limited to 'src/backend/optimizer/util/pathnode.c')
-rw-r--r--src/backend/optimizer/util/pathnode.c30
1 files changed, 24 insertions, 6 deletions
diff --git a/src/backend/optimizer/util/pathnode.c b/src/backend/optimizer/util/pathnode.c
index 89cae793ca3..6182b362a3a 100644
--- a/src/backend/optimizer/util/pathnode.c
+++ b/src/backend/optimizer/util/pathnode.c
@@ -394,8 +394,14 @@ set_cheapest(RelOptInfo *parent_rel)
* but just recycling discarded Path nodes is a very useful savings in
* a large join tree. We can recycle the List nodes of pathlist, too.
*
- * BUT: we do not pfree IndexPath objects, since they may be referenced as
- * children of BitmapHeapPaths as well as being paths in their own right.
+ * As noted in optimizer/README, deleting a previously-accepted Path is
+ * safe because we know that Paths of this rel cannot yet be referenced
+ * from any other rel, such as a higher-level join. However, in some cases
+ * it is possible that a Path is referenced by another Path for its own
+ * rel; we must not delete such a Path, even if it is dominated by the new
+ * Path. Currently this occurs only for IndexPath objects, which may be
+ * referenced as children of BitmapHeapPaths as well as being paths in
+ * their own right. Hence, we don't pfree IndexPaths when rejecting them.
*
* 'parent_rel' is the relation entry to which the path corresponds.
* 'new_path' is a potential path for parent_rel.
@@ -711,6 +717,10 @@ add_path_precheck(RelOptInfo *parent_rel,
* parallel such that each worker will generate a subset of the path's
* overall result.
*
+ * As in add_path, the partial_pathlist is kept sorted with the cheapest
+ * total path in front. This is depended on by multiple places, which
+ * just take the front entry as the cheapest path without searching.
+ *
* We don't generate parameterized partial paths for several reasons. Most
* importantly, they're not safe to execute, because there's nothing to
* make sure that a parallel scan within the parameterized portion of the
@@ -721,8 +731,8 @@ add_path_precheck(RelOptInfo *parent_rel,
* side of the plan will be small anyway. There could be rare cases where
* this wins big - e.g. if join order constraints put a 1-row relation on
* the outer side of the topmost join with a parameterized plan on the inner
- * side - but we'll have to be content not to handle such cases until somebody
- * builds an executor infrastructure that can cope with them.
+ * side - but we'll have to be content not to handle such cases until
+ * somebody builds an executor infrastructure that can cope with them.
*
* Because we don't consider parameterized paths here, we also don't
* need to consider the row counts as a measure of quality: every path will
@@ -730,6 +740,14 @@ add_path_precheck(RelOptInfo *parent_rel,
* costs: parallelism is only used for plans that will be run to completion.
* Therefore, this routine is much simpler than add_path: it needs to
* consider only pathkeys and total cost.
+ *
+ * As with add_path, we pfree paths that are found to be dominated by
+ * another partial path; this requires that there be no other references to
+ * such paths yet. Hence, GatherPaths must not be created for a rel until
+ * we're done creating all partial paths for it. We do not currently build
+ * partial indexscan paths, so there is no need for an exception for
+ * IndexPaths here; for safety, we instead Assert that a path to be freed
+ * isn't an IndexPath.
*/
void
add_partial_path(RelOptInfo *parent_rel, Path *new_path)
@@ -808,7 +826,7 @@ add_partial_path(RelOptInfo *parent_rel, Path *new_path)
{
parent_rel->partial_pathlist =
list_delete_cell(parent_rel->partial_pathlist, p1, p1_prev);
- /* add_path has a special case for IndexPath; we don't need it */
+ /* we should not see IndexPaths here, so always safe to delete */
Assert(!IsA(old_path, IndexPath));
pfree(old_path);
/* p1_prev does not advance */
@@ -842,7 +860,7 @@ add_partial_path(RelOptInfo *parent_rel, Path *new_path)
}
else
{
- /* add_path has a special case for IndexPath; we don't need it */
+ /* we should not see IndexPaths here, so always safe to delete */
Assert(!IsA(new_path, IndexPath));
/* Reject and recycle the new path */
pfree(new_path);