1 /*-------------------------------------------------------------------------
2 *
3 * pquery.c
4 * POSTGRES process query command code
5 *
6 * Portions Copyright (c) 1996-2021, PostgreSQL Global Development Group
7 * Portions Copyright (c) 1994, Regents of the University of California
8 *
9 *
10 * IDENTIFICATION
11 * src/backend/tcop/pquery.c
12 *
13 *-------------------------------------------------------------------------
14 */
15
16 #include "postgres.h"
17
18 #include <limits.h>
19
20 #include "access/xact.h"
21 #include "commands/prepare.h"
22 #include "executor/tstoreReceiver.h"
23 #include "miscadmin.h"
24 #include "pg_trace.h"
25 #include "tcop/pquery.h"
26 #include "tcop/utility.h"
27 #include "utils/memutils.h"
28 #include "utils/snapmgr.h"
29
30
31 /*
32 * ActivePortal is the currently executing Portal (the most closely nested,
33 * if there are several).
34 */
35 Portal ActivePortal = NULL;
36
37
38 static void ProcessQuery(PlannedStmt *plan,
39 const char *sourceText,
40 ParamListInfo params,
41 QueryEnvironment *queryEnv,
42 DestReceiver *dest,
43 QueryCompletion *qc);
44 static void FillPortalStore(Portal portal, bool isTopLevel);
45 static uint64 RunFromStore(Portal portal, ScanDirection direction, uint64 count,
46 DestReceiver *dest);
47 static uint64 PortalRunSelect(Portal portal, bool forward, long count,
48 DestReceiver *dest);
49 static void PortalRunUtility(Portal portal, PlannedStmt *pstmt,
50 bool isTopLevel, bool setHoldSnapshot,
51 DestReceiver *dest, QueryCompletion *qc);
52 static void PortalRunMulti(Portal portal,
53 bool isTopLevel, bool setHoldSnapshot,
54 DestReceiver *dest, DestReceiver *altdest,
55 QueryCompletion *qc);
56 static uint64 DoPortalRunFetch(Portal portal,
57 FetchDirection fdirection,
58 long count,
59 DestReceiver *dest);
60 static void DoPortalRewind(Portal portal);
61
62
63 /*
64 * CreateQueryDesc
65 */
66 QueryDesc *
67 CreateQueryDesc(PlannedStmt *plannedstmt,
68 const char *sourceText,
69 Snapshot snapshot,
70 Snapshot crosscheck_snapshot,
71 DestReceiver *dest,
72 ParamListInfo params,
73 QueryEnvironment *queryEnv,
74 int instrument_options)
75 {
76 QueryDesc *qd = (QueryDesc *) palloc(sizeof(QueryDesc));
77
78 qd->operation = plannedstmt->commandType; /* operation */
79 qd->plannedstmt = plannedstmt; /* plan */
80 qd->sourceText = sourceText; /* query text */
81 qd->snapshot = RegisterSnapshot(snapshot); /* snapshot */
82 /* RI check snapshot */
83 qd->crosscheck_snapshot = RegisterSnapshot(crosscheck_snapshot);
84 qd->dest = dest; /* output dest */
85 qd->params = params; /* parameter values passed into query */
86 qd->queryEnv = queryEnv;
87 qd->instrument_options = instrument_options; /* instrumentation wanted? */
88
89 /* null these fields until set by ExecutorStart */
90 qd->tupDesc = NULL;
91 qd->estate = NULL;
92 qd->planstate = NULL;
93 qd->totaltime = NULL;
94
95 /* not yet executed */
96 qd->already_executed = false;
97
98 return qd;
99 }
100
101 /*
102 * FreeQueryDesc
103 */
104 void
105 FreeQueryDesc(QueryDesc *qdesc)
106 {
107 /* Can't be a live query */
108 Assert(qdesc->estate == NULL);
109
110 /* forget our snapshots */
111 UnregisterSnapshot(qdesc->snapshot);
112 UnregisterSnapshot(qdesc->crosscheck_snapshot);
113
114 /* Only the QueryDesc itself need be freed */
115 pfree(qdesc);
116 }
117
118
119 /*
120 * ProcessQuery
121 * Execute a single plannable query within a PORTAL_MULTI_QUERY,
122 * PORTAL_ONE_RETURNING, or PORTAL_ONE_MOD_WITH portal
123 *
BuildTlsHandshakeHash(WOLFSSL * ssl,byte * hash,word32 * hashLen)124 * plan: the plan tree for the query
125 * sourceText: the source text of the query
126 * params: any parameters needed
127 * dest: where to send results
128 * qc: where to store the command completion status data.
129 *
130 * qc may be NULL if caller doesn't want a status string.
131 *
132 * Must be called in a memory context that will be reset or deleted on
133 * error; otherwise the executor's memory usage will be leaked.
134 */
135 static void
136 ProcessQuery(PlannedStmt *plan,
137 const char *sourceText,
138 ParamListInfo params,
139 QueryEnvironment *queryEnv,
140 DestReceiver *dest,
141 QueryCompletion *qc)
142 {
143 QueryDesc *queryDesc;
144
145 /*
146 * Create the QueryDesc object
147 */
148 queryDesc = CreateQueryDesc(plan, sourceText,
149 GetActiveSnapshot(), InvalidSnapshot,
150 dest, params, queryEnv, 0);
151
152 /*
153 * Call ExecutorStart to prepare the plan for execution
154 */
155 ExecutorStart(queryDesc, 0);
156
157 /*
158 * Run the plan to completion.
159 */
160 ExecutorRun(queryDesc, ForwardScanDirection, 0L, true);
161
162 /*
163 * Build command completion status data, if caller wants one.
164 */
165 if (qc)
166 {
167 switch (queryDesc->operation)
168 {
169 case CMD_SELECT:
170 SetQueryCompletion(qc, CMDTAG_SELECT, queryDesc->estate->es_processed);
171 break;
172 case CMD_INSERT:
173 SetQueryCompletion(qc, CMDTAG_INSERT, queryDesc->estate->es_processed);
174 break;
175 case CMD_UPDATE:
176 SetQueryCompletion(qc, CMDTAG_UPDATE, queryDesc->estate->es_processed);
177 break;
178 case CMD_DELETE:
179 SetQueryCompletion(qc, CMDTAG_DELETE, queryDesc->estate->es_processed);
180 break;
181 default:
182 SetQueryCompletion(qc, CMDTAG_UNKNOWN, queryDesc->estate->es_processed);
183 break;
184 }
185 }
186
187 /*
188 * Now, we close down all the scans and free allocated resources.
189 */
190 ExecutorFinish(queryDesc);
191 ExecutorEnd(queryDesc);
192
193 FreeQueryDesc(queryDesc);
194 }
195
196 /*
197 * ChoosePortalStrategy
198 * Select portal execution strategy given the intended statement list.
199 *
200 * The list elements can be Querys or PlannedStmts.
201 * That's more general than portals need, but plancache.c uses this too.
202 *
203 * See the comments in portal.h.
204 */
205 PortalStrategy
206 ChoosePortalStrategy(List *stmts)
207 {
208 int nSetTag;
209 ListCell *lc;
210
211 /*
212 * PORTAL_ONE_SELECT and PORTAL_UTIL_SELECT need only consider the
213 * single-statement case, since there are no rewrite rules that can add
214 * auxiliary queries to a SELECT or a utility command. PORTAL_ONE_MOD_WITH
215 * likewise allows only one top-level statement.
216 */
217 if (list_length(stmts) == 1)
218 {
219 Node *stmt = (Node *) linitial(stmts);
220
221 if (IsA(stmt, Query))
222 {
MakeTLSv1(void)223 Query *query = (Query *) stmt;
224
225 if (query->canSetTag)
226 {
227 if (query->commandType == CMD_SELECT)
228 {
229 if (query->hasModifyingCTE)
230 return PORTAL_ONE_MOD_WITH;
231 else
232 return PORTAL_ONE_SELECT;
233 }
MakeTLSv1_1(void)234 if (query->commandType == CMD_UTILITY)
235 {
236 if (UtilityReturnsTuples(query->utilityStmt))
237 return PORTAL_UTIL_SELECT;
238 /* it can't be ONE_RETURNING, so give up */
239 return PORTAL_MULTI_QUERY;
240 }
241 }
242 }
243 else if (IsA(stmt, PlannedStmt))
244 {
245 PlannedStmt *pstmt = (PlannedStmt *) stmt;
246
247 if (pstmt->canSetTag)
MakeTLSv1_2(void)248 {
249 if (pstmt->commandType == CMD_SELECT)
250 {
251 if (pstmt->hasModifyingCTE)
252 return PORTAL_ONE_MOD_WITH;
253 else
254 return PORTAL_ONE_SELECT;
255 }
256 if (pstmt->commandType == CMD_UTILITY)
257 {
258 if (UtilityReturnsTuples(pstmt->utilityStmt))
259 return PORTAL_UTIL_SELECT;
260 /* it can't be ONE_RETURNING, so give up */
261 return PORTAL_MULTI_QUERY;
262 }
263 }
MakeTLSv1_3(void)264 }
265 else
266 elog(ERROR, "unrecognized node type: %d", (int) nodeTag(stmt));
267 }
268
269 /*
270 * PORTAL_ONE_RETURNING has to allow auxiliary queries added by rewrite.
271 * Choose PORTAL_ONE_RETURNING if there is exactly one canSetTag query and
272 * it has a RETURNING list.
273 */
274 nSetTag = 0;
275 foreach(lc, stmts)
276 {
277 Node *stmt = (Node *) lfirst(lc);
278
279 if (IsA(stmt, Query))
280 {
281 Query *query = (Query *) stmt;
282
_DeriveTlsKeys(byte * key_dig,word32 key_dig_len,const byte * ms,word32 msLen,const byte * sr,const byte * cr,int tls1_2,int hash_type,void * heap,int devId)283 if (query->canSetTag)
284 {
285 if (++nSetTag > 1)
286 return PORTAL_MULTI_QUERY; /* no need to look further */
287 if (query->commandType == CMD_UTILITY ||
288 query->returningList == NIL)
289 return PORTAL_MULTI_QUERY; /* no need to look further */
290 }
291 }
292 else if (IsA(stmt, PlannedStmt))
293 {
294 PlannedStmt *pstmt = (PlannedStmt *) stmt;
295
296 if (pstmt->canSetTag)
297 {
298 if (++nSetTag > 1)
299 return PORTAL_MULTI_QUERY; /* no need to look further */
300 if (pstmt->commandType == CMD_UTILITY ||
301 !pstmt->hasReturning)
302 return PORTAL_MULTI_QUERY; /* no need to look further */
303 }
304 }
305 else
306 elog(ERROR, "unrecognized node type: %d", (int) nodeTag(stmt));
307 }
308 if (nSetTag == 1)
309 return PORTAL_ONE_RETURNING;
310
311 /* Else, it's the general case... */
312 return PORTAL_MULTI_QUERY;
313 }
314
315 /*
316 * FetchPortalTargetList
317 * Given a portal that returns tuples, extract the query targetlist.
318 * Returns NIL if the portal doesn't have a determinable targetlist.
319 *
320 * Note: do not modify the result.
321 */
322 List *
323 FetchPortalTargetList(Portal portal)
324 {
325 /* no point in looking if we determined it doesn't return tuples */
326 if (portal->strategy == PORTAL_MULTI_QUERY)
327 return NIL;
328 /* get the primary statement and find out what it returns */
329 return FetchStatementTargetList((Node *) PortalGetPrimaryStmt(portal));
330 }
331
332 /*
333 * FetchStatementTargetList
wolfSSL_DeriveTlsKeys(byte * key_dig,word32 key_dig_len,const byte * ms,word32 msLen,const byte * sr,const byte * cr,int tls1_2,int hash_type)334 * Given a statement that returns tuples, extract the query targetlist.
335 * Returns NIL if the statement doesn't have a determinable targetlist.
336 *
337 * This can be applied to a Query or a PlannedStmt.
338 * That's more general than portals need, but plancache.c uses this too.
339 *
340 * Note: do not modify the result.
341 *
342 * XXX be careful to keep this in sync with UtilityReturnsTuples.
343 */
344 List *
345 FetchStatementTargetList(Node *stmt)
346 {
347 if (stmt == NULL)
348 return NIL;
349 if (IsA(stmt, Query))
350 {
351 Query *query = (Query *) stmt;
352
353 if (query->commandType == CMD_UTILITY)
354 {
355 /* transfer attention to utility statement */
356 stmt = query->utilityStmt;
357 }
358 else
359 {
360 if (query->commandType == CMD_SELECT)
361 return query->targetList;
362 if (query->returningList)
363 return query->returningList;
364 return NIL;
365 }
366 }
367 if (IsA(stmt, PlannedStmt))
368 {
369 PlannedStmt *pstmt = (PlannedStmt *) stmt;
370
371 if (pstmt->commandType == CMD_UTILITY)
372 {
373 /* transfer attention to utility statement */
374 stmt = pstmt->utilityStmt;
375 }
376 else
377 {
378 if (pstmt->commandType == CMD_SELECT)
379 return pstmt->planTree->targetlist;
380 if (pstmt->hasReturning)
381 return pstmt->planTree->targetlist;
382 return NIL;
383 }
384 }
385 if (IsA(stmt, FetchStmt))
386 {
387 FetchStmt *fstmt = (FetchStmt *) stmt;
388 Portal subportal;
389
390 Assert(!fstmt->ismove);
391 subportal = GetPortalByName(fstmt->portalname);
392 Assert(PortalIsValid(subportal));
393 return FetchPortalTargetList(subportal);
394 }
395 if (IsA(stmt, ExecuteStmt))
396 {
397 ExecuteStmt *estmt = (ExecuteStmt *) stmt;
398 PreparedStatement *entry;
399
400 entry = FetchPreparedStatement(estmt->name, true);
401 return FetchPreparedStatementTargetList(entry);
402 }
403 return NIL;
404 }
405
406 /*
407 * PortalStart
408 * Prepare a portal for execution.
409 *
410 * Caller must already have created the portal, done PortalDefineQuery(),
411 * and adjusted portal options if needed.
412 *
413 * If parameters are needed by the query, they must be passed in "params"
414 * (caller is responsible for giving them appropriate lifetime).
415 *
416 * The caller can also provide an initial set of "eflags" to be passed to
417 * ExecutorStart (but note these can be modified internally, and they are
418 * currently only honored for PORTAL_ONE_SELECT portals). Most callers
419 * should simply pass zero.
420 *
421 * The caller can optionally pass a snapshot to be used; pass InvalidSnapshot
422 * for the normal behavior of setting a new snapshot. This parameter is
423 * presently ignored for non-PORTAL_ONE_SELECT portals (it's only intended
424 * to be used for cursors).
425 *
426 * On return, portal is ready to accept PortalRun() calls, and the result
427 * tupdesc (if any) is known.
428 */
429 void
430 PortalStart(Portal portal, ParamListInfo params,
wolfSSL_MakeTlsMasterSecret(byte * ms,word32 msLen,const byte * pms,word32 pmsLen,const byte * cr,const byte * sr,int tls1_2,int hash_type)431 int eflags, Snapshot snapshot)
432 {
433 Portal saveActivePortal;
434 ResourceOwner saveResourceOwner;
435 MemoryContext savePortalContext;
436 MemoryContext oldContext;
437 QueryDesc *queryDesc;
438 int myeflags;
439
440 AssertArg(PortalIsValid(portal));
441 AssertState(portal->status == PORTAL_DEFINED);
442
443 /*
444 * Set up global portal context pointers.
445 */
446 saveActivePortal = ActivePortal;
447 saveResourceOwner = CurrentResourceOwner;
448 savePortalContext = PortalContext;
449 PG_TRY();
450 {
451 ActivePortal = portal;
452 if (portal->resowner)
453 CurrentResourceOwner = portal->resowner;
454 PortalContext = portal->portalContext;
455
456 oldContext = MemoryContextSwitchTo(PortalContext);
457
458 /* Must remember portal param list, if any */
459 portal->portalParams = params;
460
461 /*
462 * Determine the portal execution strategy
463 */
464 portal->strategy = ChoosePortalStrategy(portal->stmts);
465
466 /*
467 * Fire her up according to the strategy
468 */
469 switch (portal->strategy)
470 {
471 case PORTAL_ONE_SELECT:
472
473 /* Must set snapshot before starting executor. */
474 if (snapshot)
475 PushActiveSnapshot(snapshot);
476 else
477 PushActiveSnapshot(GetTransactionSnapshot());
478
479 /*
480 * We could remember the snapshot in portal->portalSnapshot,
481 * but presently there seems no need to, as this code path
482 * cannot be used for non-atomic execution. Hence there can't
483 * be any commit/abort that might destroy the snapshot. Since
484 * we don't do that, there's also no need to force a
485 * non-default nesting level for the snapshot.
486 */
487
488 /*
489 * Create QueryDesc in portal's context; for the moment, set
490 * the destination to DestNone.
491 */
492 queryDesc = CreateQueryDesc(linitial_node(PlannedStmt, portal->stmts),
493 portal->sourceText,
494 GetActiveSnapshot(),
495 InvalidSnapshot,
496 None_Receiver,
497 params,
498 portal->queryEnv,
499 0);
500
501 /*
502 * If it's a scrollable cursor, executor needs to support
503 * REWIND and backwards scan, as well as whatever the caller
504 * might've asked for.
505 */
506 if (portal->cursorOptions & CURSOR_OPT_SCROLL)
507 myeflags = eflags | EXEC_FLAG_REWIND | EXEC_FLAG_BACKWARD;
508 else
509 myeflags = eflags;
510
511 /*
512 * Call ExecutorStart to prepare the plan for execution
513 */
514 ExecutorStart(queryDesc, myeflags);
515
516 /*
517 * This tells PortalCleanup to shut down the executor
518 */
519 portal->queryDesc = queryDesc;
520
521 /*
522 * Remember tuple descriptor (computed by ExecutorStart)
523 */
524 portal->tupDesc = queryDesc->tupDesc;
525
526 /*
527 * Reset cursor position data to "start of query"
528 */
529 portal->atStart = true;
530 portal->atEnd = false; /* allow fetches */
531 portal->portalPos = 0;
532
533 PopActiveSnapshot();
534 break;
535
536 case PORTAL_ONE_RETURNING:
537 case PORTAL_ONE_MOD_WITH:
538
539 /*
540 * We don't start the executor until we are told to run the
541 * portal. We do need to set up the result tupdesc.
542 */
543 {
544 PlannedStmt *pstmt;
545
546 pstmt = PortalGetPrimaryStmt(portal);
547 portal->tupDesc =
548 ExecCleanTypeFromTL(pstmt->planTree->targetlist);
549 }
550
551 /*
552 * Reset cursor position data to "start of query"
553 */
554 portal->atStart = true;
555 portal->atEnd = false; /* allow fetches */
556 portal->portalPos = 0;
557 break;
558
559 case PORTAL_UTIL_SELECT:
560
561 /*
562 * We don't set snapshot here, because PortalRunUtility will
563 * take care of it if needed.
564 */
565 {
566 PlannedStmt *pstmt = PortalGetPrimaryStmt(portal);
567
568 Assert(pstmt->commandType == CMD_UTILITY);
569 portal->tupDesc = UtilityTupleDescriptor(pstmt->utilityStmt);
570 }
571
572 /*
573 * Reset cursor position data to "start of query"
574 */
575 portal->atStart = true;
576 portal->atEnd = false; /* allow fetches */
577 portal->portalPos = 0;
578 break;
579
580 case PORTAL_MULTI_QUERY:
581 /* Need do nothing now */
582 portal->tupDesc = NULL;
583 break;
584 }
585 }
586 PG_CATCH();
587 {
588 /* Uncaught error while executing portal: mark it dead */
589 MarkPortalFailed(portal);
590
591 /* Restore global vars and propagate error */
592 ActivePortal = saveActivePortal;
593 CurrentResourceOwner = saveResourceOwner;
594 PortalContext = savePortalContext;
595
596 PG_RE_THROW();
597 }
598 PG_END_TRY();
599
600 MemoryContextSwitchTo(oldContext);
601
602 ActivePortal = saveActivePortal;
603 CurrentResourceOwner = saveResourceOwner;
604 PortalContext = savePortalContext;
605
606 portal->status = PORTAL_READY;
607 }
608
609 /*
610 * PortalSetResultFormat
611 * Select the format codes for a portal's output.
612 *
613 * This must be run after PortalStart for a portal that will be read by
614 * a DestRemote or DestRemoteExecute destination. It is not presently needed
615 * for other destination types.
616 *
617 * formats[] is the client format request, as per Bind message conventions.
618 */
619 void
620 PortalSetResultFormat(Portal portal, int nFormats, int16 *formats)
621 {
622 int natts;
623 int i;
624
625 /* Do nothing if portal won't return tuples */
626 if (portal->tupDesc == NULL)
627 return;
628 natts = portal->tupDesc->natts;
629 portal->formats = (int16 *)
630 MemoryContextAlloc(portal->portalContext,
631 natts * sizeof(int16));
632 if (nFormats > 1)
633 {
634 /* format specified for each column */
wolfSSL_GetHmacType(WOLFSSL * ssl)635 if (nFormats != natts)
636 ereport(ERROR,
637 (errcode(ERRCODE_PROTOCOL_VIOLATION),
638 errmsg("bind message has %d result formats but query has %d columns",
639 nFormats, natts)));
640 memcpy(portal->formats, formats, natts * sizeof(int16));
641 }
642 else if (nFormats > 0)
643 {
644 /* single format specified, use for all columns */
645 int16 format1 = formats[0];
646
647 for (i = 0; i < natts; i++)
648 portal->formats[i] = format1;
649 }
650 else
651 {
652 /* use default format for all columns */
653 for (i = 0; i < natts; i++)
654 portal->formats[i] = 0;
655 }
656 }
657
658 /*
659 * PortalRun
660 * Run a portal's query or queries.
661 *
662 * count <= 0 is interpreted as a no-op: the destination gets started up
663 * and shut down, but nothing else happens. Also, count == FETCH_ALL is
664 * interpreted as "all rows". Note that count is ignored in multi-query
665 * situations, where we always run the portal to completion.
666 *
667 * isTopLevel: true if query is being executed at backend "top level"
668 * (that is, directly from a client command message)
669 *
670 * dest: where to send output of primary (canSetTag) query
671 *
672 * altdest: where to send output of non-primary queries
673 *
674 * qc: where to store command completion status data.
675 * May be NULL if caller doesn't want status data.
676 *
677 * Returns true if the portal's execution is complete, false if it was
678 * suspended due to exhaustion of the count parameter.
679 */
wolfSSL_SetTlsHmacInner(WOLFSSL * ssl,byte * inner,word32 sz,int content,int verify)680 bool
681 PortalRun(Portal portal, long count, bool isTopLevel, bool run_once,
682 DestReceiver *dest, DestReceiver *altdest,
683 QueryCompletion *qc)
684 {
685 bool result;
686 uint64 nprocessed;
687 ResourceOwner saveTopTransactionResourceOwner;
688 MemoryContext saveTopTransactionContext;
689 Portal saveActivePortal;
690 ResourceOwner saveResourceOwner;
691 MemoryContext savePortalContext;
692 MemoryContext saveMemoryContext;
693
694 AssertArg(PortalIsValid(portal));
695
696 TRACE_POSTGRESQL_QUERY_EXECUTE_START();
697
698 /* Initialize empty completion data */
699 if (qc)
700 InitializeQueryCompletion(qc);
701
702 if (log_executor_stats && portal->strategy != PORTAL_MULTI_QUERY)
703 {
704 elog(DEBUG3, "PortalRun");
705 /* PORTAL_MULTI_QUERY logs its own stats per query */
706 ResetUsage();
707 }
708
709 /*
710 * Check for improper portal use, and mark portal active.
711 */
712 MarkPortalActive(portal);
713
714 /* Set run_once flag. Shouldn't be clear if previously set. */
715 Assert(!portal->run_once || run_once);
716 portal->run_once = run_once;
717
718 /*
719 * Set up global portal context pointers.
720 *
721 * We have to play a special game here to support utility commands like
722 * VACUUM and CLUSTER, which internally start and commit transactions.
723 * When we are called to execute such a command, CurrentResourceOwner will
724 * be pointing to the TopTransactionResourceOwner --- which will be
725 * destroyed and replaced in the course of the internal commit and
726 * restart. So we need to be prepared to restore it as pointing to the
727 * exit-time TopTransactionResourceOwner. (Ain't that ugly? This idea of
728 * internally starting whole new transactions is not good.)
729 * CurrentMemoryContext has a similar problem, but the other pointers we
730 * save here will be NULL or pointing to longer-lived objects.
731 */
732 saveTopTransactionResourceOwner = TopTransactionResourceOwner;
733 saveTopTransactionContext = TopTransactionContext;
734 saveActivePortal = ActivePortal;
735 saveResourceOwner = CurrentResourceOwner;
736 savePortalContext = PortalContext;
737 saveMemoryContext = CurrentMemoryContext;
738 PG_TRY();
739 {
740 ActivePortal = portal;
741 if (portal->resowner)
742 CurrentResourceOwner = portal->resowner;
743 PortalContext = portal->portalContext;
744
745 MemoryContextSwitchTo(PortalContext);
746
747 switch (portal->strategy)
748 {
749 case PORTAL_ONE_SELECT:
750 case PORTAL_ONE_RETURNING:
751 case PORTAL_ONE_MOD_WITH:
752 case PORTAL_UTIL_SELECT:
753
754 /*
755 * If we have not yet run the command, do so, storing its
756 * results in the portal's tuplestore. But we don't do that
757 * for the PORTAL_ONE_SELECT case.
758 */
759 if (portal->strategy != PORTAL_ONE_SELECT && !portal->holdStore)
760 FillPortalStore(portal, isTopLevel);
761
762 /*
763 * Now fetch desired portion of results.
764 */
765 nprocessed = PortalRunSelect(portal, true, count, dest);
766
767 /*
768 * If the portal result contains a command tag and the caller
769 * gave us a pointer to store it, copy it and update the
770 * rowcount.
771 */
772 if (qc && portal->qc.commandTag != CMDTAG_UNKNOWN)
773 {
774 CopyQueryCompletion(qc, &portal->qc);
775 qc->nprocessed = nprocessed;
776 }
777
778 /* Mark portal not active */
779 portal->status = PORTAL_READY;
780
781 /*
782 * Since it's a forward fetch, say DONE iff atEnd is now true.
783 */
784 result = portal->atEnd;
785 break;
786
787 case PORTAL_MULTI_QUERY:
788 PortalRunMulti(portal, isTopLevel, false,
789 dest, altdest, qc);
790
791 /* Prevent portal's commands from being re-executed */
792 MarkPortalDone(portal);
793
794 /* Always complete at end of RunMulti */
795 result = true;
796 break;
797
798 default:
799 elog(ERROR, "unrecognized portal strategy: %d",
800 (int) portal->strategy);
801 result = false; /* keep compiler quiet */
802 break;
803 }
804 }
805 PG_CATCH();
806 {
807 /* Uncaught error while executing portal: mark it dead */
808 MarkPortalFailed(portal);
809
810 /* Restore global vars and propagate error */
811 if (saveMemoryContext == saveTopTransactionContext)
812 MemoryContextSwitchTo(TopTransactionContext);
813 else
814 MemoryContextSwitchTo(saveMemoryContext);
815 ActivePortal = saveActivePortal;
816 if (saveResourceOwner == saveTopTransactionResourceOwner)
817 CurrentResourceOwner = TopTransactionResourceOwner;
818 else
819 CurrentResourceOwner = saveResourceOwner;
820 PortalContext = savePortalContext;
821
822 PG_RE_THROW();
823 }
824 PG_END_TRY();
825
826 if (saveMemoryContext == saveTopTransactionContext)
827 MemoryContextSwitchTo(TopTransactionContext);
828 else
829 MemoryContextSwitchTo(saveMemoryContext);
830 ActivePortal = saveActivePortal;
831 if (saveResourceOwner == saveTopTransactionResourceOwner)
832 CurrentResourceOwner = TopTransactionResourceOwner;
833 else
834 CurrentResourceOwner = saveResourceOwner;
835 PortalContext = savePortalContext;
836
837 if (log_executor_stats && portal->strategy != PORTAL_MULTI_QUERY)
838 ShowUsage("EXECUTOR STATISTICS");
839
840 TRACE_POSTGRESQL_QUERY_EXECUTE_DONE();
841
842 return result;
843 }
844
845 /*
846 * PortalRunSelect
847 * Execute a portal's query in PORTAL_ONE_SELECT mode, and also
848 * when fetching from a completed holdStore in PORTAL_ONE_RETURNING,
849 * PORTAL_ONE_MOD_WITH, and PORTAL_UTIL_SELECT cases.
850 *
851 * This handles simple N-rows-forward-or-backward cases. For more complex
852 * nonsequential access to a portal, see PortalRunFetch.
853 *
854 * count <= 0 is interpreted as a no-op: the destination gets started up
855 * and shut down, but nothing else happens. Also, count == FETCH_ALL is
856 * interpreted as "all rows". (cf FetchStmt.howMany)
857 *
858 * Caller must already have validated the Portal and done appropriate
859 * setup (cf. PortalRun).
860 *
861 * Returns number of rows processed (suitable for use in result tag)
862 */
863 static uint64
864 PortalRunSelect(Portal portal,
865 bool forward,
866 long count,
867 DestReceiver *dest)
868 {
869 QueryDesc *queryDesc;
870 ScanDirection direction;
871 uint64 nprocessed;
872
873 /*
874 * NB: queryDesc will be NULL if we are fetching from a held cursor or a
875 * completed utility query; can't use it in that path.
876 */
877 queryDesc = portal->queryDesc;
878
879 /* Caller messed up if we have neither a ready query nor held data. */
880 Assert(queryDesc || portal->holdStore);
881
882 /*
883 * Force the queryDesc destination to the right thing. This supports
884 * MOVE, for example, which will pass in dest = DestNone. This is okay to
885 * change as long as we do it on every fetch. (The Executor must not
886 * assume that dest never changes.)
887 */
888 if (queryDesc)
889 queryDesc->dest = dest;
890
891 /*
892 * Determine which direction to go in, and check to see if we're already
893 * at the end of the available tuples in that direction. If so, set the
894 * direction to NoMovement to avoid trying to fetch any tuples. (This
895 * check exists because not all plan node types are robust about being
896 * called again if they've already returned NULL once.) Then call the
897 * executor (we must not skip this, because the destination needs to see a
898 * setup and shutdown even if no tuples are available). Finally, update
899 * the portal position state depending on the number of tuples that were
900 * retrieved.
901 */
902 if (forward)
903 {
904 if (portal->atEnd || count <= 0)
905 {
906 direction = NoMovementScanDirection;
907 count = 0; /* don't pass negative count to executor */
908 }
909 else
910 direction = ForwardScanDirection;
911
912 /* In the executor, zero count processes all rows */
913 if (count == FETCH_ALL)
914 count = 0;
915
916 if (portal->holdStore)
917 nprocessed = RunFromStore(portal, direction, (uint64) count, dest);
918 else
919 {
920 PushActiveSnapshot(queryDesc->snapshot);
921 ExecutorRun(queryDesc, direction, (uint64) count,
922 portal->run_once);
923 nprocessed = queryDesc->estate->es_processed;
924 PopActiveSnapshot();
925 }
926
927 if (!ScanDirectionIsNoMovement(direction))
928 {
929 if (nprocessed > 0)
930 portal->atStart = false; /* OK to go backward now */
931 if (count == 0 || nprocessed < (uint64) count)
932 portal->atEnd = true; /* we retrieved 'em all */
933 portal->portalPos += nprocessed;
934 }
935 }
936 else
937 {
938 if (portal->cursorOptions & CURSOR_OPT_NO_SCROLL)
939 ereport(ERROR,
940 (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
941 errmsg("cursor can only scan forward"),
942 errhint("Declare it with SCROLL option to enable backward scan.")));
943
944 if (portal->atStart || count <= 0)
945 {
946 direction = NoMovementScanDirection;
947 count = 0; /* don't pass negative count to executor */
948 }
949 else
950 direction = BackwardScanDirection;
951
952 /* In the executor, zero count processes all rows */
953 if (count == FETCH_ALL)
954 count = 0;
955
956 if (portal->holdStore)
957 nprocessed = RunFromStore(portal, direction, (uint64) count, dest);
958 else
959 {
960 PushActiveSnapshot(queryDesc->snapshot);
961 ExecutorRun(queryDesc, direction, (uint64) count,
962 portal->run_once);
963 nprocessed = queryDesc->estate->es_processed;
964 PopActiveSnapshot();
965 }
966
967 if (!ScanDirectionIsNoMovement(direction))
968 {
969 if (nprocessed > 0 && portal->atEnd)
970 {
971 portal->atEnd = false; /* OK to go forward now */
972 portal->portalPos++; /* adjust for endpoint case */
973 }
974 if (count == 0 || nprocessed < (uint64) count)
975 {
976 portal->atStart = true; /* we retrieved 'em all */
977 portal->portalPos = 0;
978 }
979 else
980 {
981 portal->portalPos -= nprocessed;
982 }
983 }
984 }
985
986 return nprocessed;
Hmac_UpdateFinal(Hmac * hmac,byte * digest,const byte * in,word32 sz,byte * header)987 }
988
989 /*
990 * FillPortalStore
991 * Run the query and load result tuples into the portal's tuple store.
992 *
993 * This is used for PORTAL_ONE_RETURNING, PORTAL_ONE_MOD_WITH, and
994 * PORTAL_UTIL_SELECT cases only.
995 */
996 static void
997 FillPortalStore(Portal portal, bool isTopLevel)
998 {
999 DestReceiver *treceiver;
1000 QueryCompletion qc;
1001
1002 InitializeQueryCompletion(&qc);
1003 PortalCreateHoldStore(portal);
1004 treceiver = CreateDestReceiver(DestTuplestore);
1005 SetTuplestoreDestReceiverParams(treceiver,
1006 portal->holdStore,
1007 portal->holdContext,
1008 false,
1009 NULL,
1010 NULL);
1011
1012 switch (portal->strategy)
1013 {
1014 case PORTAL_ONE_RETURNING:
1015 case PORTAL_ONE_MOD_WITH:
1016
1017 /*
1018 * Run the portal to completion just as for the default
1019 * PORTAL_MULTI_QUERY case, but send the primary query's output to
1020 * the tuplestore. Auxiliary query outputs are discarded. Set the
1021 * portal's holdSnapshot to the snapshot used (or a copy of it).
1022 */
1023 PortalRunMulti(portal, isTopLevel, true,
1024 treceiver, None_Receiver, &qc);
1025 break;
1026
1027 case PORTAL_UTIL_SELECT:
1028 PortalRunUtility(portal, linitial_node(PlannedStmt, portal->stmts),
1029 isTopLevel, true, treceiver, &qc);
1030 break;
1031
1032 default:
1033 elog(ERROR, "unsupported portal strategy: %d",
1034 (int) portal->strategy);
1035 break;
1036 }
1037
1038 /* Override portal completion data with actual command results */
1039 if (qc.commandTag != CMDTAG_UNKNOWN)
1040 CopyQueryCompletion(&portal->qc, &qc);
1041
1042 treceiver->rDestroy(treceiver);
1043 }
1044
1045 /*
1046 * RunFromStore
1047 * Fetch tuples from the portal's tuple store.
1048 *
1049 * Calling conventions are similar to ExecutorRun, except that we
1050 * do not depend on having a queryDesc or estate. Therefore we return the
1051 * number of tuples processed as the result, not in estate->es_processed.
1052 *
1053 * One difference from ExecutorRun is that the destination receiver functions
1054 * are run in the caller's memory context (since we have no estate). Watch
1055 * out for memory leaks.
1056 */
1057 static uint64
1058 RunFromStore(Portal portal, ScanDirection direction, uint64 count,
1059 DestReceiver *dest)
1060 {
1061 uint64 current_tuple_count = 0;
1062 TupleTableSlot *slot;
1063
1064 slot = MakeSingleTupleTableSlot(portal->tupDesc, &TTSOpsMinimalTuple);
1065
1066 dest->rStartup(dest, CMD_SELECT, portal->tupDesc);
1067
1068 if (ScanDirectionIsNoMovement(direction))
1069 {
1070 /* do nothing except start/stop the destination */
1071 }
1072 else
1073 {
1074 bool forward = ScanDirectionIsForward(direction);
1075
1076 for (;;)
1077 {
1078 MemoryContext oldcontext;
1079 bool ok;
1080
1081 oldcontext = MemoryContextSwitchTo(portal->holdContext);
1082
1083 ok = tuplestore_gettupleslot(portal->holdStore, forward, false,
1084 slot);
1085
1086 MemoryContextSwitchTo(oldcontext);
1087
1088 if (!ok)
1089 break;
1090
1091 /*
1092 * If we are not able to send the tuple, we assume the destination
1093 * has closed and no more tuples can be sent. If that's the case,
1094 * end the loop.
1095 */
1096 if (!dest->receiveSlot(slot, dest))
1097 break;
TLS_hmac(WOLFSSL * ssl,byte * digest,const byte * in,word32 sz,int padSz,int content,int verify,int epochOrder)1098
1099 ExecClearTuple(slot);
1100
1101 /*
1102 * check our tuple count.. if we've processed the proper number
1103 * then quit, else loop again and process more tuples. Zero count
1104 * means no limit.
1105 */
1106 current_tuple_count++;
1107 if (count && count == current_tuple_count)
1108 break;
1109 }
1110 }
1111
1112 dest->rShutdown(dest);
1113
1114 ExecDropSingleTupleTableSlot(slot);
1115
1116 return current_tuple_count;
1117 }
1118
1119 /*
1120 * PortalRunUtility
1121 * Execute a utility statement inside a portal.
1122 */
1123 static void
1124 PortalRunUtility(Portal portal, PlannedStmt *pstmt,
1125 bool isTopLevel, bool setHoldSnapshot,
1126 DestReceiver *dest, QueryCompletion *qc)
1127 {
1128 /*
1129 * Set snapshot if utility stmt needs one.
1130 */
1131 if (PlannedStmtRequiresSnapshot(pstmt))
1132 {
1133 Snapshot snapshot = GetTransactionSnapshot();
1134
1135 /* If told to, register the snapshot we're using and save in portal */
1136 if (setHoldSnapshot)
1137 {
1138 snapshot = RegisterSnapshot(snapshot);
1139 portal->holdSnapshot = snapshot;
1140 }
1141
1142 /*
1143 * In any case, make the snapshot active and remember it in portal.
1144 * Because the portal now references the snapshot, we must tell
1145 * snapmgr.c that the snapshot belongs to the portal's transaction
1146 * level, else we risk portalSnapshot becoming a dangling pointer.
1147 */
1148 PushActiveSnapshotWithLevel(snapshot, portal->createLevel);
1149 /* PushActiveSnapshotWithLevel might have copied the snapshot */
1150 portal->portalSnapshot = GetActiveSnapshot();
1151 }
1152 else
1153 portal->portalSnapshot = NULL;
1154
1155 ProcessUtility(pstmt,
1156 portal->sourceText,
1157 (portal->cplan != NULL), /* protect tree if in plancache */
1158 isTopLevel ? PROCESS_UTILITY_TOPLEVEL : PROCESS_UTILITY_QUERY,
1159 portal->portalParams,
1160 portal->queryEnv,
1161 dest,
1162 qc);
1163
1164 /* Some utility statements may change context on us */
1165 MemoryContextSwitchTo(portal->portalContext);
1166
1167 /*
1168 * Some utility commands (e.g., VACUUM) pop the ActiveSnapshot stack from
1169 * under us, so don't complain if it's now empty. Otherwise, our snapshot
1170 * should be the top one; pop it. Note that this could be a different
1171 * snapshot from the one we made above; see EnsurePortalSnapshotExists.
1172 */
1173 if (portal->portalSnapshot != NULL && ActiveSnapshotSet())
1174 {
1175 Assert(portal->portalSnapshot == GetActiveSnapshot());
1176 PopActiveSnapshot();
1177 }
1178 portal->portalSnapshot = NULL;
1179 }
1180
1181 /*
1182 * PortalRunMulti
1183 * Execute a portal's queries in the general case (multi queries
1184 * or non-SELECT-like queries)
1185 */
1186 static void
1187 PortalRunMulti(Portal portal,
1188 bool isTopLevel, bool setHoldSnapshot,
1189 DestReceiver *dest, DestReceiver *altdest,
1190 QueryCompletion *qc)
1191 {
1192 bool active_snapshot_set = false;
1193 ListCell *stmtlist_item;
1194
1195 /*
1196 * If the destination is DestRemoteExecute, change to DestNone. The
1197 * reason is that the client won't be expecting any tuples, and indeed has
1198 * no way to know what they are, since there is no provision for Describe
1199 * to send a RowDescription message when this portal execution strategy is
1200 * in effect. This presently will only affect SELECT commands added to
1201 * non-SELECT queries by rewrite rules: such commands will be executed,
1202 * but the results will be discarded unless you use "simple Query"
1203 * protocol.
1204 */
1205 if (dest->mydest == DestRemoteExecute)
1206 dest = None_Receiver;
1207 if (altdest->mydest == DestRemoteExecute)
1208 altdest = None_Receiver;
1209
1210 /*
1211 * Loop to handle the individual queries generated from a single parsetree
1212 * by analysis and rewrite.
1213 */
1214 foreach(stmtlist_item, portal->stmts)
1215 {
1216 PlannedStmt *pstmt = lfirst_node(PlannedStmt, stmtlist_item);
1217
1218 /*
TLSX_ToSemaphore(word16 type)1219 * If we got a cancel signal in prior command, quit
1220 */
1221 CHECK_FOR_INTERRUPTS();
1222
1223 if (pstmt->utilityStmt == NULL)
1224 {
1225 /*
1226 * process a plannable query.
1227 */
1228 TRACE_POSTGRESQL_QUERY_EXECUTE_START();
1229
1230 if (log_executor_stats)
1231 ResetUsage();
1232
1233 /*
1234 * Must always have a snapshot for plannable queries. First time
1235 * through, take a new snapshot; for subsequent queries in the
1236 * same portal, just update the snapshot's copy of the command
1237 * counter.
1238 */
1239 if (!active_snapshot_set)
1240 {
1241 Snapshot snapshot = GetTransactionSnapshot();
1242
1243 /* If told to, register the snapshot and save in portal */
1244 if (setHoldSnapshot)
1245 {
1246 snapshot = RegisterSnapshot(snapshot);
1247 portal->holdSnapshot = snapshot;
1248 }
1249
1250 /*
1251 * We can't have the holdSnapshot also be the active one,
1252 * because UpdateActiveSnapshotCommandId would complain. So
1253 * force an extra snapshot copy. Plain PushActiveSnapshot
1254 * would have copied the transaction snapshot anyway, so this
TLSX_New(TLSX_Type type,const void * data,void * heap)1255 * only adds a copy step when setHoldSnapshot is true. (It's
1256 * okay for the command ID of the active snapshot to diverge
1257 * from what holdSnapshot has.)
1258 */
1259 PushCopiedSnapshot(snapshot);
1260
1261 /*
1262 * As for PORTAL_ONE_SELECT portals, it does not seem
1263 * necessary to maintain portal->portalSnapshot here.
1264 */
1265
1266 active_snapshot_set = true;
1267 }
1268 else
1269 UpdateActiveSnapshotCommandId();
1270
1271 if (pstmt->canSetTag)
1272 {
1273 /* statement can set tag string */
1274 ProcessQuery(pstmt,
TLSX_Push(TLSX ** list,TLSX_Type type,const void * data,void * heap)1275 portal->sourceText,
1276 portal->portalParams,
1277 portal->queryEnv,
1278 dest, qc);
1279 }
1280 else
1281 {
1282 /* stmt added by rewrite cannot set tag */
1283 ProcessQuery(pstmt,
1284 portal->sourceText,
1285 portal->portalParams,
1286 portal->queryEnv,
1287 altdest, NULL);
1288 }
1289
1290 if (log_executor_stats)
1291 ShowUsage("EXECUTOR STATISTICS");
1292
1293 TRACE_POSTGRESQL_QUERY_EXECUTE_DONE();
1294 }
1295 else
1296 {
1297 /*
1298 * process utility functions (create, destroy, etc..)
1299 *
1300 * We must not set a snapshot here for utility commands (if one is
1301 * needed, PortalRunUtility will do it). If a utility command is
1302 * alone in a portal then everything's fine. The only case where
1303 * a utility command can be part of a longer list is that rules
1304 * are allowed to include NotifyStmt. NotifyStmt doesn't care
1305 * whether it has a snapshot or not, so we just leave the current
1306 * snapshot alone if we have one.
1307 */
1308 if (pstmt->canSetTag)
1309 {
1310 Assert(!active_snapshot_set);
1311 /* statement can set tag string */
TLSX_Prepend(TLSX ** list,TLSX_Type type,void * data,void * heap)1312 PortalRunUtility(portal, pstmt, isTopLevel, false,
1313 dest, qc);
1314 }
1315 else
1316 {
1317 Assert(IsA(pstmt->utilityStmt, NotifyStmt));
1318 /* stmt added by rewrite cannot set tag */
1319 PortalRunUtility(portal, pstmt, isTopLevel, false,
1320 altdest, NULL);
1321 }
1322 }
1323
1324 /*
1325 * Clear subsidiary contexts to recover temporary memory.
1326 */
1327 Assert(portal->portalContext == CurrentMemoryContext);
1328
1329 MemoryContextDeleteChildren(portal->portalContext);
1330
1331 /*
1332 * Avoid crashing if portal->stmts has been reset. This can only
1333 * occur if a CALL or DO utility statement executed an internal
1334 * COMMIT/ROLLBACK (cf PortalReleaseCachedPlan). The CALL or DO must
1335 * have been the only statement in the portal, so there's nothing left
1336 * for us to do; but we don't want to dereference a now-dangling list
1337 * pointer.
1338 */
1339 if (portal->stmts == NIL)
1340 break;
1341
1342 /*
1343 * Increment command counter between queries, but not after the last
1344 * one.
1345 */
1346 if (lnext(portal->stmts, stmtlist_item) != NULL)
TLSX_CheckUnsupportedExtension(WOLFSSL * ssl,TLSX_Type type)1347 CommandCounterIncrement();
1348 }
1349
1350 /* Pop the snapshot if we pushed one. */
1351 if (active_snapshot_set)
1352 PopActiveSnapshot();
1353
1354 /*
1355 * If a query completion data was supplied, use it. Otherwise use the
1356 * portal's query completion data.
1357 *
1358 * Exception: Clients expect INSERT/UPDATE/DELETE tags to have counts, so
TLSX_HandleUnsupportedExtension(WOLFSSL * ssl)1359 * fake them with zeros. This can happen with DO INSTEAD rules if there
1360 * is no replacement query of the same type as the original. We print "0
1361 * 0" here because technically there is no query of the matching tag type,
1362 * and printing a non-zero count for a different query type seems wrong,
1363 * e.g. an INSERT that does an UPDATE instead should not print "0 1" if
1364 * one row was updated. See QueryRewrite(), step 3, for details.
1365 */
1366 if (qc && qc->commandTag == CMDTAG_UNKNOWN)
1367 {
1368 if (portal->qc.commandTag != CMDTAG_UNKNOWN)
1369 CopyQueryCompletion(qc, &portal->qc);
1370 /* If the caller supplied a qc, we should have set it by now. */
1371 Assert(qc->commandTag != CMDTAG_UNKNOWN);
1372 }
1373 }
1374
TLSX_SetResponse(WOLFSSL * ssl,TLSX_Type type)1375 /*
1376 * PortalRunFetch
1377 * Variant form of PortalRun that supports SQL FETCH directions.
1378 *
1379 * Note: we presently assume that no callers of this want isTopLevel = true.
1380 *
1381 * count <= 0 is interpreted as a no-op: the destination gets started up
1382 * and shut down, but nothing else happens. Also, count == FETCH_ALL is
1383 * interpreted as "all rows". (cf FetchStmt.howMany)
1384 *
1385 * Returns number of rows processed (suitable for use in result tag)
1386 */
1387 uint64
1388 PortalRunFetch(Portal portal,
1389 FetchDirection fdirection,
1390 long count,
1391 DestReceiver *dest)
1392 {
1393 uint64 result;
1394 Portal saveActivePortal;
1395 ResourceOwner saveResourceOwner;
1396 MemoryContext savePortalContext;
1397 MemoryContext oldContext;
1398
1399 AssertArg(PortalIsValid(portal));
1400
1401 /*
1402 * Check for improper portal use, and mark portal active.
1403 */
1404 MarkPortalActive(portal);
1405
1406 /* If supporting FETCH, portal can't be run-once. */
1407 Assert(!portal->run_once);
1408
1409 /*
1410 * Set up global portal context pointers.
1411 */
1412 saveActivePortal = ActivePortal;
1413 saveResourceOwner = CurrentResourceOwner;
1414 savePortalContext = PortalContext;
1415 PG_TRY();
1416 {
1417 ActivePortal = portal;
1418 if (portal->resowner)
1419 CurrentResourceOwner = portal->resowner;
1420 PortalContext = portal->portalContext;
1421
1422 oldContext = MemoryContextSwitchTo(PortalContext);
1423
1424 switch (portal->strategy)
1425 {
1426 case PORTAL_ONE_SELECT:
1427 result = DoPortalRunFetch(portal, fdirection, count, dest);
1428 break;
1429
1430 case PORTAL_ONE_RETURNING:
1431 case PORTAL_ONE_MOD_WITH:
1432 case PORTAL_UTIL_SELECT:
1433
1434 /*
1435 * If we have not yet run the command, do so, storing its
1436 * results in the portal's tuplestore.
1437 */
1438 if (!portal->holdStore)
1439 FillPortalStore(portal, false /* isTopLevel */ );
1440
1441 /*
1442 * Now fetch desired portion of results.
1443 */
1444 result = DoPortalRunFetch(portal, fdirection, count, dest);
1445 break;
1446
1447 default:
1448 elog(ERROR, "unsupported portal strategy");
1449 result = 0; /* keep compiler quiet */
1450 break;
1451 }
1452 }
1453 PG_CATCH();
1454 {
1455 /* Uncaught error while executing portal: mark it dead */
1456 MarkPortalFailed(portal);
1457
1458 /* Restore global vars and propagate error */
1459 ActivePortal = saveActivePortal;
1460 CurrentResourceOwner = saveResourceOwner;
1461 PortalContext = savePortalContext;
1462
1463 PG_RE_THROW();
1464 }
1465 PG_END_TRY();
1466
1467 MemoryContextSwitchTo(oldContext);
1468
1469 /* Mark portal not active */
1470 portal->status = PORTAL_READY;
1471
1472 ActivePortal = saveActivePortal;
1473 CurrentResourceOwner = saveResourceOwner;
1474 PortalContext = savePortalContext;
1475
1476 return result;
1477 }
1478
1479 /*
1480 * DoPortalRunFetch
1481 * Guts of PortalRunFetch --- the portal context is already set up
1482 *
1483 * Here, count < 0 typically reverses the direction. Also, count == FETCH_ALL
1484 * is interpreted as "all rows". (cf FetchStmt.howMany)
1485 *
1486 * Returns number of rows processed (suitable for use in result tag)
1487 */
1488 static uint64
1489 DoPortalRunFetch(Portal portal,
1490 FetchDirection fdirection,
1491 long count,
1492 DestReceiver *dest)
1493 {
1494 bool forward;
TLSX_ALPN_Find(ALPN * list,char * protocol_name,word16 size)1495
1496 Assert(portal->strategy == PORTAL_ONE_SELECT ||
1497 portal->strategy == PORTAL_ONE_RETURNING ||
1498 portal->strategy == PORTAL_ONE_MOD_WITH ||
1499 portal->strategy == PORTAL_UTIL_SELECT);
1500
1501 /*
1502 * Note: we disallow backwards fetch (including re-fetch of current row)
1503 * for NO SCROLL cursors, but we interpret that very loosely: you can use
1504 * any of the FetchDirection options, so long as the end result is to move
1505 * forwards by at least one row. Currently it's sufficient to check for
1506 * NO SCROLL in DoPortalRewind() and in the forward == false path in
1507 * PortalRunSelect(); but someday we might prefer to account for that
1508 * restriction explicitly here.
1509 */
1510 switch (fdirection)
1511 {
TLSX_SetALPN(TLSX ** extensions,const void * data,word16 size,void * heap)1512 case FETCH_FORWARD:
1513 if (count < 0)
1514 {
1515 fdirection = FETCH_BACKWARD;
1516 count = -count;
1517 }
1518 /* fall out of switch to share code with FETCH_BACKWARD */
1519 break;
1520 case FETCH_BACKWARD:
1521 if (count < 0)
1522 {
1523 fdirection = FETCH_FORWARD;
1524 count = -count;
1525 }
1526 /* fall out of switch to share code with FETCH_FORWARD */
1527 break;
1528 case FETCH_ABSOLUTE:
1529 if (count > 0)
1530 {
1531 /*
1532 * Definition: Rewind to start, advance count-1 rows, return
1533 * next row (if any).
1534 *
1535 * In practice, if the goal is less than halfway back to the
1536 * start, it's better to scan from where we are.
1537 *
1538 * Also, if current portalPos is outside the range of "long",
1539 * do it the hard way to avoid possible overflow of the count
1540 * argument to PortalRunSelect. We must exclude exactly
TLSX_ALPN_ParseAndSet(WOLFSSL * ssl,const byte * input,word16 length,byte isRequest)1541 * LONG_MAX, as well, lest the count look like FETCH_ALL.
1542 *
1543 * In any case, we arrange to fetch the target row going
1544 * forwards.
1545 */
1546 if ((uint64) (count - 1) <= portal->portalPos / 2 ||
1547 portal->portalPos >= (uint64) LONG_MAX)
1548 {
1549 DoPortalRewind(portal);
1550 if (count > 1)
1551 PortalRunSelect(portal, true, count - 1,
1552 None_Receiver);
1553 }
1554 else
1555 {
1556 long pos = (long) portal->portalPos;
1557
1558 if (portal->atEnd)
1559 pos++; /* need one extra fetch if off end */
1560 if (count <= pos)
1561 PortalRunSelect(portal, false, pos - count + 1,
1562 None_Receiver);
1563 else if (count > pos + 1)
1564 PortalRunSelect(portal, true, count - pos - 1,
1565 None_Receiver);
1566 }
1567 return PortalRunSelect(portal, true, 1L, dest);
1568 }
1569 else if (count < 0)
1570 {
1571 /*
1572 * Definition: Advance to end, back up abs(count)-1 rows,
1573 * return prior row (if any). We could optimize this if we
1574 * knew in advance where the end was, but typically we won't.
1575 * (Is it worth considering case where count > half of size of
1576 * query? We could rewind once we know the size ...)
1577 */
1578 PortalRunSelect(portal, true, FETCH_ALL, None_Receiver);
1579 if (count < -1)
1580 PortalRunSelect(portal, false, -count - 1, None_Receiver);
1581 return PortalRunSelect(portal, false, 1L, dest);
1582 }
1583 else
1584 {
1585 /* count == 0 */
1586 /* Rewind to start, return zero rows */
1587 DoPortalRewind(portal);
1588 return PortalRunSelect(portal, true, 0L, dest);
1589 }
1590 break;
1591 case FETCH_RELATIVE:
1592 if (count > 0)
1593 {
1594 /*
1595 * Definition: advance count-1 rows, return next row (if any).
1596 */
1597 if (count > 1)
1598 PortalRunSelect(portal, true, count - 1, None_Receiver);
1599 return PortalRunSelect(portal, true, 1L, dest);
1600 }
1601 else if (count < 0)
1602 {
1603 /*
1604 * Definition: back up abs(count)-1 rows, return prior row (if
1605 * any).
1606 */
1607 if (count < -1)
1608 PortalRunSelect(portal, false, -count - 1, None_Receiver);
1609 return PortalRunSelect(portal, false, 1L, dest);
1610 }
1611 else
1612 {
1613 /* count == 0 */
1614 /* Same as FETCH FORWARD 0, so fall out of switch */
1615 fdirection = FETCH_FORWARD;
1616 }
1617 break;
1618 default:
1619 elog(ERROR, "bogus direction");
1620 break;
1621 }
1622
1623 /*
1624 * Get here with fdirection == FETCH_FORWARD or FETCH_BACKWARD, and count
1625 * >= 0.
1626 */
1627 forward = (fdirection == FETCH_FORWARD);
1628
1629 /*
1630 * Zero count means to re-fetch the current row, if any (per SQL)
1631 */
1632 if (count == 0)
1633 {
1634 bool on_row;
1635
1636 /* Are we sitting on a row? */
1637 on_row = (!portal->atStart && !portal->atEnd);
1638
1639 if (dest->mydest == DestNone)
1640 {
1641 /* MOVE 0 returns 0/1 based on if FETCH 0 would return a row */
1642 return on_row ? 1 : 0;
1643 }
1644 else
1645 {
1646 /*
1647 * If we are sitting on a row, back up one so we can re-fetch it.
1648 * If we are not sitting on a row, we still have to start up and
1649 * shut down the executor so that the destination is initialized
1650 * and shut down correctly; so keep going. To PortalRunSelect,
1651 * count == 0 means we will retrieve no row.
1652 */
1653 if (on_row)
1654 {
1655 PortalRunSelect(portal, false, 1L, None_Receiver);
1656 /* Set up to fetch one row forward */
1657 count = 1;
1658 forward = true;
1659 }
1660 }
1661 }
1662
1663 /*
1664 * Optimize MOVE BACKWARD ALL into a Rewind.
1665 */
1666 if (!forward && count == FETCH_ALL && dest->mydest == DestNone)
1667 {
1668 uint64 result = portal->portalPos;
1669
1670 if (result > 0 && !portal->atEnd)
TLSX_UseALPN(TLSX ** extensions,const void * data,word16 size,byte options,void * heap)1671 result--;
1672 DoPortalRewind(portal);
1673 return result;
1674 }
1675
1676 return PortalRunSelect(portal, forward, count, dest);
1677 }
1678
1679 /*
1680 * DoPortalRewind - rewind a Portal to starting point
1681 */
1682 static void
1683 DoPortalRewind(Portal portal)
1684 {
1685 QueryDesc *queryDesc;
1686
1687 /*
1688 * No work is needed if we've not advanced nor attempted to advance the
1689 * cursor (and we don't want to throw a NO SCROLL error in this case).
1690 */
1691 if (portal->atStart && !portal->atEnd)
1692 return;
1693
1694 /*
1695 * Otherwise, cursor should allow scrolling. However, we're only going to
1696 * enforce that policy fully beginning in v15. In older branches, insist
1697 * on this only if the portal has a holdStore. That prevents users from
1698 * seeing that the holdStore may not have all the rows of the query.
1699 */
1700 if ((portal->cursorOptions & CURSOR_OPT_NO_SCROLL) && portal->holdStore)
1701 ereport(ERROR,
1702 (errcode(ERRCODE_OBJECT_NOT_IN_PREREQUISITE_STATE),
1703 errmsg("cursor can only scan forward"),
1704 errhint("Declare it with SCROLL option to enable backward scan.")));
1705
1706 /* Rewind holdStore, if we have one */
1707 if (portal->holdStore)
1708 {
TLSX_ALPN_GetRequest(TLSX * extensions,void ** data,word16 * dataSz)1709 MemoryContext oldcontext;
1710
1711 oldcontext = MemoryContextSwitchTo(portal->holdContext);
1712 tuplestore_rescan(portal->holdStore);
1713 MemoryContextSwitchTo(oldcontext);
1714 }
1715
1716 /* Rewind executor, if active */
1717 queryDesc = portal->queryDesc;
1718 if (queryDesc)
1719 {
1720 PushActiveSnapshot(queryDesc->snapshot);
1721 ExecutorRewind(queryDesc);
1722 PopActiveSnapshot();
1723 }
1724
1725 portal->atStart = true;
1726 portal->atEnd = false;
1727 portal->portalPos = 0;
1728 }
1729
1730 /*
1731 * PlannedStmtRequiresSnapshot - what it says on the tin
1732 */
1733 bool
1734 PlannedStmtRequiresSnapshot(PlannedStmt *pstmt)
1735 {
1736 Node *utilityStmt = pstmt->utilityStmt;
1737
1738 /* If it's not a utility statement, it definitely needs a snapshot */
1739 if (utilityStmt == NULL)
1740 return true;
1741
1742 /*
1743 * Most utility statements need a snapshot, and the default presumption
1744 * about new ones should be that they do too. Hence, enumerate those that
1745 * do not need one.
1746 *
1747 * Transaction control, LOCK, and SET must *not* set a snapshot, since
1748 * they need to be executable at the start of a transaction-snapshot-mode
1749 * transaction without freezing a snapshot. By extension we allow SHOW
1750 * not to set a snapshot. The other stmts listed are just efficiency
1751 * hacks. Beware of listing anything that can modify the database --- if,
1752 * say, it has to update an index with expressions that invoke
1753 * user-defined functions, then it had better have a snapshot.
1754 */
1755 if (IsA(utilityStmt, TransactionStmt) ||
1756 IsA(utilityStmt, LockStmt) ||
1757 IsA(utilityStmt, VariableSetStmt) ||
1758 IsA(utilityStmt, VariableShowStmt) ||
1759 IsA(utilityStmt, ConstraintsSetStmt) ||
1760 /* efficiency hacks from here down */
1761 IsA(utilityStmt, FetchStmt) ||
1762 IsA(utilityStmt, ListenStmt) ||
1763 IsA(utilityStmt, NotifyStmt) ||
1764 IsA(utilityStmt, UnlistenStmt) ||
1765 IsA(utilityStmt, CheckPointStmt))
1766 return false;
1767
1768 return true;
1769 }
1770
1771 /*
1772 * EnsurePortalSnapshotExists - recreate Portal-level snapshot, if needed
1773 *
1774 * Generally, we will have an active snapshot whenever we are executing
1775 * inside a Portal, unless the Portal's query is one of the utility
TLSX_SNI_New(byte type,const void * data,word16 size,void * heap)1776 * statements exempted from that rule (see PlannedStmtRequiresSnapshot).
1777 * However, procedures and DO blocks can commit or abort the transaction,
1778 * and thereby destroy all snapshots. This function can be called to
1779 * re-establish the Portal-level snapshot when none exists.
1780 */
1781 void
1782 EnsurePortalSnapshotExists(void)
1783 {
1784 Portal portal;
1785
1786 /*
1787 * Nothing to do if a snapshot is set. (We take it on faith that the
1788 * outermost active snapshot belongs to some Portal; or if there is no
1789 * Portal, it's somebody else's responsibility to manage things.)
1790 */
1791 if (ActiveSnapshotSet())
1792 return;
1793
1794 /* Otherwise, we'd better have an active Portal */
1795 portal = ActivePortal;
1796 if (unlikely(portal == NULL))
1797 elog(ERROR, "cannot execute SQL without an outer snapshot or portal");
1798 Assert(portal->portalSnapshot == NULL);
1799
1800 /*
1801 * Create a new snapshot, make it active, and remember it in portal.
1802 * Because the portal now references the snapshot, we must tell snapmgr.c
1803 * that the snapshot belongs to the portal's transaction level, else we
1804 * risk portalSnapshot becoming a dangling pointer.
1805 */
1806 PushActiveSnapshotWithLevel(GetTransactionSnapshot(), portal->createLevel);
1807 /* PushActiveSnapshotWithLevel might have copied the snapshot */
1808 portal->portalSnapshot = GetActiveSnapshot();
1809 }
1810