Line data Source code
1 : /*-------------------------------------------------------------------------
2 : *
3 : * trigger.c
4 : * PostgreSQL TRIGGERs support code.
5 : *
6 : * Portions Copyright (c) 1996-2026, PostgreSQL Global Development Group
7 : * Portions Copyright (c) 1994, Regents of the University of California
8 : *
9 : * IDENTIFICATION
10 : * src/backend/commands/trigger.c
11 : *
12 : *-------------------------------------------------------------------------
13 : */
14 : #include "postgres.h"
15 :
16 : #include "access/genam.h"
17 : #include "access/htup_details.h"
18 : #include "access/relation.h"
19 : #include "access/sysattr.h"
20 : #include "access/table.h"
21 : #include "access/tableam.h"
22 : #include "access/tupconvert.h"
23 : #include "access/xact.h"
24 : #include "catalog/catalog.h"
25 : #include "catalog/dependency.h"
26 : #include "catalog/indexing.h"
27 : #include "catalog/objectaccess.h"
28 : #include "catalog/partition.h"
29 : #include "catalog/pg_constraint.h"
30 : #include "catalog/pg_inherits.h"
31 : #include "catalog/pg_proc.h"
32 : #include "catalog/pg_trigger.h"
33 : #include "catalog/pg_type.h"
34 : #include "commands/trigger.h"
35 : #include "executor/executor.h"
36 : #include "executor/instrument.h"
37 : #include "miscadmin.h"
38 : #include "nodes/bitmapset.h"
39 : #include "nodes/makefuncs.h"
40 : #include "optimizer/optimizer.h"
41 : #include "parser/parse_clause.h"
42 : #include "parser/parse_collate.h"
43 : #include "parser/parse_func.h"
44 : #include "parser/parse_relation.h"
45 : #include "partitioning/partdesc.h"
46 : #include "pgstat.h"
47 : #include "rewrite/rewriteHandler.h"
48 : #include "rewrite/rewriteManip.h"
49 : #include "storage/lmgr.h"
50 : #include "utils/acl.h"
51 : #include "utils/builtins.h"
52 : #include "utils/fmgroids.h"
53 : #include "utils/guc_hooks.h"
54 : #include "utils/inval.h"
55 : #include "utils/lsyscache.h"
56 : #include "utils/memutils.h"
57 : #include "utils/plancache.h"
58 : #include "utils/rel.h"
59 : #include "utils/snapmgr.h"
60 : #include "utils/syscache.h"
61 : #include "utils/tuplestore.h"
62 :
63 :
64 : /* GUC variables */
65 : int SessionReplicationRole = SESSION_REPLICATION_ROLE_ORIGIN;
66 :
67 : /* How many levels deep into trigger execution are we? */
68 : static int MyTriggerDepth = 0;
69 :
70 : /* Local function prototypes */
71 : static void renametrig_internal(Relation tgrel, Relation targetrel,
72 : HeapTuple trigtup, const char *newname,
73 : const char *expected_name);
74 : static void renametrig_partition(Relation tgrel, Oid partitionId,
75 : Oid parentTriggerOid, const char *newname,
76 : const char *expected_name);
77 : static void SetTriggerFlags(TriggerDesc *trigdesc, Trigger *trigger);
78 : static bool GetTupleForTrigger(EState *estate,
79 : EPQState *epqstate,
80 : ResultRelInfo *relinfo,
81 : ItemPointer tid,
82 : LockTupleMode lockmode,
83 : TupleTableSlot *oldslot,
84 : bool do_epq_recheck,
85 : TupleTableSlot **epqslot,
86 : TM_Result *tmresultp,
87 : TM_FailureData *tmfdp);
88 : static bool TriggerEnabled(EState *estate, ResultRelInfo *relinfo,
89 : Trigger *trigger, TriggerEvent event,
90 : Bitmapset *modifiedCols,
91 : TupleTableSlot *oldslot, TupleTableSlot *newslot);
92 : static HeapTuple ExecCallTriggerFunc(TriggerData *trigdata,
93 : int tgindx,
94 : FmgrInfo *finfo,
95 : TriggerInstrumentation *instr,
96 : MemoryContext per_tuple_context);
97 : static void AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
98 : ResultRelInfo *src_partinfo,
99 : ResultRelInfo *dst_partinfo,
100 : int event, bool row_trigger,
101 : TupleTableSlot *oldslot, TupleTableSlot *newslot,
102 : List *recheckIndexes, Bitmapset *modifiedCols,
103 : TransitionCaptureState *transition_capture,
104 : bool is_crosspart_update);
105 : static void AfterTriggerEnlargeQueryState(void);
106 : static bool before_stmt_triggers_fired(Oid relid, CmdType cmdType);
107 : static HeapTuple check_modified_virtual_generated(TupleDesc tupdesc, HeapTuple tuple);
108 :
109 :
110 : /*
111 : * Create a trigger. Returns the address of the created trigger.
112 : *
113 : * queryString is the source text of the CREATE TRIGGER command.
114 : * This must be supplied if a whenClause is specified, else it can be NULL.
115 : *
116 : * relOid, if nonzero, is the relation on which the trigger should be
117 : * created. If zero, the name provided in the statement will be looked up.
118 : *
119 : * refRelOid, if nonzero, is the relation to which the constraint trigger
120 : * refers. If zero, the constraint relation name provided in the statement
121 : * will be looked up as needed.
122 : *
123 : * constraintOid, if nonzero, says that this trigger is being created
124 : * internally to implement that constraint. A suitable pg_depend entry will
125 : * be made to link the trigger to that constraint. constraintOid is zero when
126 : * executing a user-entered CREATE TRIGGER command. (For CREATE CONSTRAINT
127 : * TRIGGER, we build a pg_constraint entry internally.)
128 : *
129 : * indexOid, if nonzero, is the OID of an index associated with the constraint.
130 : * We do nothing with this except store it into pg_trigger.tgconstrindid;
131 : * but when creating a trigger for a deferrable unique constraint on a
132 : * partitioned table, its children are looked up. Note we don't cope with
133 : * invalid indexes in that case.
134 : *
135 : * funcoid, if nonzero, is the OID of the function to invoke. When this is
136 : * given, stmt->funcname is ignored.
137 : *
138 : * parentTriggerOid, if nonzero, is a trigger that begets this one; so that
139 : * if that trigger is dropped, this one should be too. There are two cases
140 : * when a nonzero value is passed for this: 1) when this function recurses to
141 : * create the trigger on partitions, 2) when creating child foreign key
142 : * triggers; see CreateFKCheckTrigger() and createForeignKeyActionTriggers().
143 : *
144 : * If whenClause is passed, it is an already-transformed expression for
145 : * WHEN. In this case, we ignore any that may come in stmt->whenClause.
146 : *
147 : * If isInternal is true then this is an internally-generated trigger.
148 : * This argument sets the tgisinternal field of the pg_trigger entry, and
149 : * if true causes us to modify the given trigger name to ensure uniqueness.
150 : *
151 : * When isInternal is not true we require ACL_TRIGGER permissions on the
152 : * relation, as well as ACL_EXECUTE on the trigger function. For internal
153 : * triggers the caller must apply any required permission checks.
154 : *
155 : * When called on partitioned tables, this function recurses to create the
156 : * trigger on all the partitions, except if isInternal is true, in which
157 : * case caller is expected to execute recursion on its own. in_partition
158 : * indicates such a recursive call; outside callers should pass "false"
159 : * (but see CloneRowTriggersToPartition).
160 : */
161 : ObjectAddress
162 10966 : CreateTrigger(const CreateTrigStmt *stmt, const char *queryString,
163 : Oid relOid, Oid refRelOid, Oid constraintOid, Oid indexOid,
164 : Oid funcoid, Oid parentTriggerOid, Node *whenClause,
165 : bool isInternal, bool in_partition)
166 : {
167 : return
168 10966 : CreateTriggerFiringOn(stmt, queryString, relOid, refRelOid,
169 : constraintOid, indexOid, funcoid,
170 : parentTriggerOid, whenClause, isInternal,
171 : in_partition, TRIGGER_FIRES_ON_ORIGIN);
172 : }
173 :
174 : /*
175 : * Like the above; additionally the firing condition
176 : * (always/origin/replica/disabled) can be specified.
177 : */
178 : ObjectAddress
179 11569 : CreateTriggerFiringOn(const CreateTrigStmt *stmt, const char *queryString,
180 : Oid relOid, Oid refRelOid, Oid constraintOid,
181 : Oid indexOid, Oid funcoid, Oid parentTriggerOid,
182 : Node *whenClause, bool isInternal, bool in_partition,
183 : char trigger_fires_when)
184 : {
185 : int16 tgtype;
186 : int ncolumns;
187 : int16 *columns;
188 : int2vector *tgattr;
189 : List *whenRtable;
190 : char *qual;
191 : Datum values[Natts_pg_trigger];
192 : bool nulls[Natts_pg_trigger];
193 : Relation rel;
194 : AclResult aclresult;
195 : Relation tgrel;
196 : Relation pgrel;
197 11569 : HeapTuple tuple = NULL;
198 : Oid funcrettype;
199 11569 : Oid trigoid = InvalidOid;
200 : char internaltrigname[NAMEDATALEN];
201 : char *trigname;
202 11569 : Oid constrrelid = InvalidOid;
203 : ObjectAddress myself,
204 : referenced;
205 11569 : char *oldtablename = NULL;
206 11569 : char *newtablename = NULL;
207 : bool partition_recurse;
208 11569 : bool trigger_exists = false;
209 11569 : Oid existing_constraint_oid = InvalidOid;
210 11569 : bool existing_isInternal = false;
211 11569 : bool existing_isClone = false;
212 :
213 11569 : if (OidIsValid(relOid))
214 9451 : rel = table_open(relOid, ShareRowExclusiveLock);
215 : else
216 2118 : rel = table_openrv(stmt->relation, ShareRowExclusiveLock);
217 :
218 : /*
219 : * Triggers must be on tables or views, and there are additional
220 : * relation-type-specific restrictions.
221 : */
222 11569 : if (rel->rd_rel->relkind == RELKIND_RELATION)
223 : {
224 : /* Tables can't have INSTEAD OF triggers */
225 9501 : if (stmt->timing != TRIGGER_TYPE_BEFORE &&
226 8638 : stmt->timing != TRIGGER_TYPE_AFTER)
227 12 : ereport(ERROR,
228 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
229 : errmsg("\"%s\" is a table",
230 : RelationGetRelationName(rel)),
231 : errdetail("Tables cannot have INSTEAD OF triggers.")));
232 : }
233 2068 : else if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
234 : {
235 : /* Partitioned tables can't have INSTEAD OF triggers */
236 1871 : if (stmt->timing != TRIGGER_TYPE_BEFORE &&
237 1795 : stmt->timing != TRIGGER_TYPE_AFTER)
238 4 : ereport(ERROR,
239 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
240 : errmsg("\"%s\" is a table",
241 : RelationGetRelationName(rel)),
242 : errdetail("Tables cannot have INSTEAD OF triggers.")));
243 :
244 : /*
245 : * FOR EACH ROW triggers have further restrictions
246 : */
247 1867 : if (stmt->row)
248 : {
249 : /*
250 : * Disallow use of transition tables.
251 : *
252 : * Note that we have another restriction about transition tables
253 : * in partitions; search for 'has_superclass' below for an
254 : * explanation. The check here is just to protect from the fact
255 : * that if we allowed it here, the creation would succeed for a
256 : * partitioned table with no partitions, but would be blocked by
257 : * the other restriction when the first partition was created,
258 : * which is very unfriendly behavior.
259 : */
260 1703 : if (stmt->transitionRels != NIL)
261 4 : ereport(ERROR,
262 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
263 : errmsg("\"%s\" is a partitioned table",
264 : RelationGetRelationName(rel)),
265 : errdetail("ROW triggers with transition tables are not supported on partitioned tables.")));
266 : }
267 : }
268 197 : else if (rel->rd_rel->relkind == RELKIND_VIEW)
269 : {
270 : /*
271 : * Views can have INSTEAD OF triggers (which we check below are
272 : * row-level), or statement-level BEFORE/AFTER triggers.
273 : */
274 139 : if (stmt->timing != TRIGGER_TYPE_INSTEAD && stmt->row)
275 24 : ereport(ERROR,
276 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
277 : errmsg("\"%s\" is a view",
278 : RelationGetRelationName(rel)),
279 : errdetail("Views cannot have row-level BEFORE or AFTER triggers.")));
280 : /* Disallow TRUNCATE triggers on VIEWs */
281 115 : if (TRIGGER_FOR_TRUNCATE(stmt->events))
282 8 : ereport(ERROR,
283 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
284 : errmsg("\"%s\" is a view",
285 : RelationGetRelationName(rel)),
286 : errdetail("Views cannot have TRUNCATE triggers.")));
287 : }
288 58 : else if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
289 : {
290 58 : if (stmt->timing != TRIGGER_TYPE_BEFORE &&
291 31 : stmt->timing != TRIGGER_TYPE_AFTER)
292 0 : ereport(ERROR,
293 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
294 : errmsg("\"%s\" is a foreign table",
295 : RelationGetRelationName(rel)),
296 : errdetail("Foreign tables cannot have INSTEAD OF triggers.")));
297 :
298 : /*
299 : * We disallow constraint triggers to protect the assumption that
300 : * triggers on FKs can't be deferred. See notes with AfterTriggers
301 : * data structures, below.
302 : */
303 58 : if (stmt->isconstraint)
304 4 : ereport(ERROR,
305 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
306 : errmsg("\"%s\" is a foreign table",
307 : RelationGetRelationName(rel)),
308 : errdetail("Foreign tables cannot have constraint triggers.")));
309 : }
310 : else
311 0 : ereport(ERROR,
312 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
313 : errmsg("relation \"%s\" cannot have triggers",
314 : RelationGetRelationName(rel)),
315 : errdetail_relkind_not_supported(rel->rd_rel->relkind)));
316 :
317 11513 : if (!allowSystemTableMods && IsSystemRelation(rel))
318 1 : ereport(ERROR,
319 : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
320 : errmsg("permission denied: \"%s\" is a system catalog",
321 : RelationGetRelationName(rel))));
322 :
323 11512 : if (stmt->isconstraint)
324 : {
325 : /*
326 : * We must take a lock on the target relation to protect against
327 : * concurrent drop. It's not clear that AccessShareLock is strong
328 : * enough, but we certainly need at least that much... otherwise, we
329 : * might end up creating a pg_constraint entry referencing a
330 : * nonexistent table.
331 : */
332 8965 : if (OidIsValid(refRelOid))
333 : {
334 8774 : LockRelationOid(refRelOid, AccessShareLock);
335 8774 : constrrelid = refRelOid;
336 : }
337 191 : else if (stmt->constrrel != NULL)
338 16 : constrrelid = RangeVarGetRelid(stmt->constrrel, AccessShareLock,
339 : false);
340 : }
341 :
342 : /* permission checks */
343 11512 : if (!isInternal)
344 : {
345 2664 : aclresult = pg_class_aclcheck(RelationGetRelid(rel), GetUserId(),
346 : ACL_TRIGGER);
347 2664 : if (aclresult != ACLCHECK_OK)
348 0 : aclcheck_error(aclresult, get_relkind_objtype(rel->rd_rel->relkind),
349 0 : RelationGetRelationName(rel));
350 :
351 2664 : if (OidIsValid(constrrelid))
352 : {
353 28 : aclresult = pg_class_aclcheck(constrrelid, GetUserId(),
354 : ACL_TRIGGER);
355 28 : if (aclresult != ACLCHECK_OK)
356 0 : aclcheck_error(aclresult, get_relkind_objtype(get_rel_relkind(constrrelid)),
357 0 : get_rel_name(constrrelid));
358 : }
359 : }
360 :
361 : /*
362 : * When called on a partitioned table to create a FOR EACH ROW trigger
363 : * that's not internal, we create one trigger for each partition, too.
364 : *
365 : * For that, we'd better hold lock on all of them ahead of time.
366 : */
367 13487 : partition_recurse = !isInternal && stmt->row &&
368 1975 : rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE;
369 11512 : if (partition_recurse)
370 279 : list_free(find_all_inheritors(RelationGetRelid(rel),
371 : ShareRowExclusiveLock, NULL));
372 :
373 : /* Compute tgtype */
374 11512 : TRIGGER_CLEAR_TYPE(tgtype);
375 11512 : if (stmt->row)
376 10823 : TRIGGER_SETT_ROW(tgtype);
377 11512 : tgtype |= stmt->timing;
378 11512 : tgtype |= stmt->events;
379 :
380 : /* Disallow ROW-level TRUNCATE triggers */
381 11512 : if (TRIGGER_FOR_ROW(tgtype) && TRIGGER_FOR_TRUNCATE(tgtype))
382 0 : ereport(ERROR,
383 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
384 : errmsg("TRUNCATE FOR EACH ROW triggers are not supported")));
385 :
386 : /* INSTEAD triggers must be row-level, and can't have WHEN or columns */
387 11512 : if (TRIGGER_FOR_INSTEAD(tgtype))
388 : {
389 79 : if (!TRIGGER_FOR_ROW(tgtype))
390 4 : ereport(ERROR,
391 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
392 : errmsg("INSTEAD OF triggers must be FOR EACH ROW")));
393 75 : if (stmt->whenClause)
394 4 : ereport(ERROR,
395 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
396 : errmsg("INSTEAD OF triggers cannot have WHEN conditions")));
397 71 : if (stmt->columns != NIL)
398 4 : ereport(ERROR,
399 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
400 : errmsg("INSTEAD OF triggers cannot have column lists")));
401 : }
402 :
403 : /*
404 : * We don't yet support naming ROW transition variables, but the parser
405 : * recognizes the syntax so we can give a nicer message here.
406 : *
407 : * Per standard, REFERENCING TABLE names are only allowed on AFTER
408 : * triggers. Per standard, REFERENCING ROW names are not allowed with FOR
409 : * EACH STATEMENT. Per standard, each OLD/NEW, ROW/TABLE permutation is
410 : * only allowed once. Per standard, OLD may not be specified when
411 : * creating a trigger only for INSERT, and NEW may not be specified when
412 : * creating a trigger only for DELETE.
413 : *
414 : * Notice that the standard allows an AFTER ... FOR EACH ROW trigger to
415 : * reference both ROW and TABLE transition data.
416 : */
417 11500 : if (stmt->transitionRels != NIL)
418 : {
419 324 : List *varList = stmt->transitionRels;
420 : ListCell *lc;
421 :
422 708 : foreach(lc, varList)
423 : {
424 416 : TriggerTransition *tt = lfirst_node(TriggerTransition, lc);
425 :
426 416 : if (!(tt->isTable))
427 0 : ereport(ERROR,
428 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
429 : errmsg("ROW variable naming in the REFERENCING clause is not supported"),
430 : errhint("Use OLD TABLE or NEW TABLE for naming transition tables.")));
431 :
432 : /*
433 : * Because of the above test, we omit further ROW-related testing
434 : * below. If we later allow naming OLD and NEW ROW variables,
435 : * adjustments will be needed below.
436 : */
437 :
438 416 : if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
439 4 : ereport(ERROR,
440 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
441 : errmsg("\"%s\" is a foreign table",
442 : RelationGetRelationName(rel)),
443 : errdetail("Triggers on foreign tables cannot have transition tables.")));
444 :
445 412 : if (rel->rd_rel->relkind == RELKIND_VIEW)
446 4 : ereport(ERROR,
447 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
448 : errmsg("\"%s\" is a view",
449 : RelationGetRelationName(rel)),
450 : errdetail("Triggers on views cannot have transition tables.")));
451 :
452 : /*
453 : * We currently don't allow row-level triggers with transition
454 : * tables on partition or inheritance children. Such triggers
455 : * would somehow need to see tuples converted to the format of the
456 : * table they're attached to, and it's not clear which subset of
457 : * tuples each child should see. See also the prohibitions in
458 : * ATExecAttachPartition() and ATExecAddInherit().
459 : */
460 408 : if (TRIGGER_FOR_ROW(tgtype) && has_superclass(rel->rd_id))
461 : {
462 : /* Use appropriate error message. */
463 8 : if (rel->rd_rel->relispartition)
464 4 : ereport(ERROR,
465 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
466 : errmsg("ROW triggers with transition tables are not supported on partitions")));
467 : else
468 4 : ereport(ERROR,
469 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
470 : errmsg("ROW triggers with transition tables are not supported on inheritance children")));
471 : }
472 :
473 400 : if (stmt->timing != TRIGGER_TYPE_AFTER)
474 0 : ereport(ERROR,
475 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
476 : errmsg("transition table name can only be specified for an AFTER trigger")));
477 :
478 400 : if (TRIGGER_FOR_TRUNCATE(tgtype))
479 4 : ereport(ERROR,
480 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
481 : errmsg("TRUNCATE triggers with transition tables are not supported")));
482 :
483 : /*
484 : * We currently don't allow multi-event triggers ("INSERT OR
485 : * UPDATE") with transition tables, because it's not clear how to
486 : * handle INSERT ... ON CONFLICT statements which can fire both
487 : * INSERT and UPDATE triggers. We show the inserted tuples to
488 : * INSERT triggers and the updated tuples to UPDATE triggers, but
489 : * it's not yet clear what INSERT OR UPDATE trigger should see.
490 : * This restriction could be lifted if we can decide on the right
491 : * semantics in a later release.
492 : */
493 396 : if (((TRIGGER_FOR_INSERT(tgtype) ? 1 : 0) +
494 396 : (TRIGGER_FOR_UPDATE(tgtype) ? 1 : 0) +
495 396 : (TRIGGER_FOR_DELETE(tgtype) ? 1 : 0)) != 1)
496 4 : ereport(ERROR,
497 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
498 : errmsg("transition tables cannot be specified for triggers with more than one event")));
499 :
500 : /*
501 : * We currently don't allow column-specific triggers with
502 : * transition tables. Per spec, that seems to require
503 : * accumulating separate transition tables for each combination of
504 : * columns, which is a lot of work for a rather marginal feature.
505 : */
506 392 : if (stmt->columns != NIL)
507 4 : ereport(ERROR,
508 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
509 : errmsg("transition tables cannot be specified for triggers with column lists")));
510 :
511 : /*
512 : * We disallow constraint triggers with transition tables, to
513 : * protect the assumption that such triggers can't be deferred.
514 : * See notes with AfterTriggers data structures, below.
515 : *
516 : * Currently this is enforced by the grammar, so just Assert here.
517 : */
518 : Assert(!stmt->isconstraint);
519 :
520 388 : if (tt->isNew)
521 : {
522 204 : if (!(TRIGGER_FOR_INSERT(tgtype) ||
523 109 : TRIGGER_FOR_UPDATE(tgtype)))
524 0 : ereport(ERROR,
525 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
526 : errmsg("NEW TABLE can only be specified for an INSERT or UPDATE trigger")));
527 :
528 204 : if (newtablename != NULL)
529 0 : ereport(ERROR,
530 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
531 : errmsg("NEW TABLE cannot be specified multiple times")));
532 :
533 204 : newtablename = tt->name;
534 : }
535 : else
536 : {
537 184 : if (!(TRIGGER_FOR_DELETE(tgtype) ||
538 105 : TRIGGER_FOR_UPDATE(tgtype)))
539 4 : ereport(ERROR,
540 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
541 : errmsg("OLD TABLE can only be specified for a DELETE or UPDATE trigger")));
542 :
543 180 : if (oldtablename != NULL)
544 0 : ereport(ERROR,
545 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
546 : errmsg("OLD TABLE cannot be specified multiple times")));
547 :
548 180 : oldtablename = tt->name;
549 : }
550 : }
551 :
552 292 : if (newtablename != NULL && oldtablename != NULL &&
553 92 : strcmp(newtablename, oldtablename) == 0)
554 0 : ereport(ERROR,
555 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
556 : errmsg("OLD TABLE name and NEW TABLE name cannot be the same")));
557 : }
558 :
559 : /*
560 : * Parse the WHEN clause, if any and we weren't passed an already
561 : * transformed one.
562 : *
563 : * Note that as a side effect, we fill whenRtable when parsing. If we got
564 : * an already parsed clause, this does not occur, which is what we want --
565 : * no point in adding redundant dependencies below.
566 : */
567 11468 : if (!whenClause && stmt->whenClause)
568 88 : {
569 : ParseState *pstate;
570 : ParseNamespaceItem *nsitem;
571 : List *varList;
572 : ListCell *lc;
573 :
574 : /* Set up a pstate to parse with */
575 120 : pstate = make_parsestate(NULL);
576 120 : pstate->p_sourcetext = queryString;
577 :
578 : /*
579 : * Set up nsitems for OLD and NEW references.
580 : *
581 : * 'OLD' must always have varno equal to 1 and 'NEW' equal to 2.
582 : */
583 120 : nsitem = addRangeTableEntryForRelation(pstate, rel,
584 : AccessShareLock,
585 : makeAlias("old", NIL),
586 : false, false);
587 120 : addNSItemToQuery(pstate, nsitem, false, true, true);
588 120 : nsitem = addRangeTableEntryForRelation(pstate, rel,
589 : AccessShareLock,
590 : makeAlias("new", NIL),
591 : false, false);
592 120 : addNSItemToQuery(pstate, nsitem, false, true, true);
593 :
594 : /* Transform expression. Copy to be sure we don't modify original */
595 120 : whenClause = transformWhereClause(pstate,
596 120 : copyObject(stmt->whenClause),
597 : EXPR_KIND_TRIGGER_WHEN,
598 : "WHEN");
599 : /* we have to fix its collations too */
600 120 : assign_expr_collations(pstate, whenClause);
601 :
602 : /*
603 : * Check for disallowed references to OLD/NEW.
604 : *
605 : * NB: pull_var_clause is okay here only because we don't allow
606 : * subselects in WHEN clauses; it would fail to examine the contents
607 : * of subselects.
608 : */
609 120 : varList = pull_var_clause(whenClause, 0);
610 237 : foreach(lc, varList)
611 : {
612 149 : Var *var = (Var *) lfirst(lc);
613 :
614 149 : switch (var->varno)
615 : {
616 56 : case PRS2_OLD_VARNO:
617 56 : if (!TRIGGER_FOR_ROW(tgtype))
618 4 : ereport(ERROR,
619 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
620 : errmsg("statement trigger's WHEN condition cannot reference column values"),
621 : parser_errposition(pstate, var->location)));
622 52 : if (TRIGGER_FOR_INSERT(tgtype))
623 4 : ereport(ERROR,
624 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
625 : errmsg("INSERT trigger's WHEN condition cannot reference OLD values"),
626 : parser_errposition(pstate, var->location)));
627 : /* system columns are okay here */
628 48 : break;
629 93 : case PRS2_NEW_VARNO:
630 93 : if (!TRIGGER_FOR_ROW(tgtype))
631 0 : ereport(ERROR,
632 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
633 : errmsg("statement trigger's WHEN condition cannot reference column values"),
634 : parser_errposition(pstate, var->location)));
635 93 : if (TRIGGER_FOR_DELETE(tgtype))
636 4 : ereport(ERROR,
637 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
638 : errmsg("DELETE trigger's WHEN condition cannot reference NEW values"),
639 : parser_errposition(pstate, var->location)));
640 89 : if (var->varattno < 0 && TRIGGER_FOR_BEFORE(tgtype))
641 4 : ereport(ERROR,
642 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
643 : errmsg("BEFORE trigger's WHEN condition cannot reference NEW system columns"),
644 : parser_errposition(pstate, var->location)));
645 85 : if (TRIGGER_FOR_BEFORE(tgtype) &&
646 34 : var->varattno == 0 &&
647 12 : RelationGetDescr(rel)->constr &&
648 8 : (RelationGetDescr(rel)->constr->has_generated_stored ||
649 4 : RelationGetDescr(rel)->constr->has_generated_virtual))
650 8 : ereport(ERROR,
651 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
652 : errmsg("BEFORE trigger's WHEN condition cannot reference NEW generated columns"),
653 : errdetail("A whole-row reference is used and the table contains generated columns."),
654 : parser_errposition(pstate, var->location)));
655 77 : if (TRIGGER_FOR_BEFORE(tgtype) &&
656 26 : var->varattno > 0 &&
657 22 : TupleDescAttr(RelationGetDescr(rel), var->varattno - 1)->attgenerated)
658 8 : ereport(ERROR,
659 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
660 : errmsg("BEFORE trigger's WHEN condition cannot reference NEW generated columns"),
661 : errdetail("Column \"%s\" is a generated column.",
662 : NameStr(TupleDescAttr(RelationGetDescr(rel), var->varattno - 1)->attname)),
663 : parser_errposition(pstate, var->location)));
664 69 : break;
665 0 : default:
666 : /* can't happen without add_missing_from, so just elog */
667 0 : elog(ERROR, "trigger WHEN condition cannot contain references to other relations");
668 : break;
669 : }
670 : }
671 :
672 : /* we'll need the rtable for recordDependencyOnExpr */
673 88 : whenRtable = pstate->p_rtable;
674 :
675 88 : qual = nodeToString(whenClause);
676 :
677 88 : free_parsestate(pstate);
678 : }
679 11348 : else if (!whenClause)
680 : {
681 11320 : whenClause = NULL;
682 11320 : whenRtable = NIL;
683 11320 : qual = NULL;
684 : }
685 : else
686 : {
687 28 : qual = nodeToString(whenClause);
688 28 : whenRtable = NIL;
689 : }
690 :
691 : /*
692 : * Find and validate the trigger function.
693 : */
694 11436 : if (!OidIsValid(funcoid))
695 10833 : funcoid = LookupFuncName(stmt->funcname, 0, NULL, false);
696 11436 : if (!isInternal)
697 : {
698 2588 : aclresult = object_aclcheck(ProcedureRelationId, funcoid, GetUserId(), ACL_EXECUTE);
699 2588 : if (aclresult != ACLCHECK_OK)
700 0 : aclcheck_error(aclresult, OBJECT_FUNCTION,
701 0 : NameListToString(stmt->funcname));
702 : }
703 11436 : funcrettype = get_func_rettype(funcoid);
704 11436 : if (funcrettype != TRIGGEROID)
705 0 : ereport(ERROR,
706 : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
707 : errmsg("function %s must return type %s",
708 : NameListToString(stmt->funcname), "trigger")));
709 :
710 : /*
711 : * Scan pg_trigger to see if there is already a trigger of the same name.
712 : * Skip this for internally generated triggers, since we'll modify the
713 : * name to be unique below.
714 : *
715 : * NOTE that this is cool only because we have ShareRowExclusiveLock on
716 : * the relation, so the trigger set won't be changing underneath us.
717 : */
718 11436 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
719 11436 : if (!isInternal)
720 : {
721 : ScanKeyData skeys[2];
722 : SysScanDesc tgscan;
723 :
724 2588 : ScanKeyInit(&skeys[0],
725 : Anum_pg_trigger_tgrelid,
726 : BTEqualStrategyNumber, F_OIDEQ,
727 : ObjectIdGetDatum(RelationGetRelid(rel)));
728 :
729 2588 : ScanKeyInit(&skeys[1],
730 : Anum_pg_trigger_tgname,
731 : BTEqualStrategyNumber, F_NAMEEQ,
732 2588 : CStringGetDatum(stmt->trigname));
733 :
734 2588 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
735 : NULL, 2, skeys);
736 :
737 : /* There should be at most one matching tuple */
738 2588 : if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
739 : {
740 68 : Form_pg_trigger oldtrigger = (Form_pg_trigger) GETSTRUCT(tuple);
741 :
742 68 : trigoid = oldtrigger->oid;
743 68 : existing_constraint_oid = oldtrigger->tgconstraint;
744 68 : existing_isInternal = oldtrigger->tgisinternal;
745 68 : existing_isClone = OidIsValid(oldtrigger->tgparentid);
746 68 : trigger_exists = true;
747 : /* copy the tuple to use in CatalogTupleUpdate() */
748 68 : tuple = heap_copytuple(tuple);
749 : }
750 2588 : systable_endscan(tgscan);
751 : }
752 :
753 11436 : if (!trigger_exists)
754 : {
755 : /* Generate the OID for the new trigger. */
756 11368 : trigoid = GetNewOidWithIndex(tgrel, TriggerOidIndexId,
757 : Anum_pg_trigger_oid);
758 : }
759 : else
760 : {
761 : /*
762 : * If OR REPLACE was specified, we'll replace the old trigger;
763 : * otherwise complain about the duplicate name.
764 : */
765 68 : if (!stmt->replace)
766 12 : ereport(ERROR,
767 : (errcode(ERRCODE_DUPLICATE_OBJECT),
768 : errmsg("trigger \"%s\" for relation \"%s\" already exists",
769 : stmt->trigname, RelationGetRelationName(rel))));
770 :
771 : /*
772 : * An internal trigger or a child trigger (isClone) cannot be replaced
773 : * by a user-defined trigger. However, skip this test when
774 : * in_partition, because then we're recursing from a partitioned table
775 : * and the check was made at the parent level.
776 : */
777 56 : if ((existing_isInternal || existing_isClone) &&
778 40 : !isInternal && !in_partition)
779 4 : ereport(ERROR,
780 : (errcode(ERRCODE_DUPLICATE_OBJECT),
781 : errmsg("trigger \"%s\" for relation \"%s\" is an internal or a child trigger",
782 : stmt->trigname, RelationGetRelationName(rel))));
783 :
784 : /*
785 : * It is not allowed to replace with a constraint trigger; gram.y
786 : * should have enforced this already.
787 : */
788 : Assert(!stmt->isconstraint);
789 :
790 : /*
791 : * It is not allowed to replace an existing constraint trigger,
792 : * either. (The reason for these restrictions is partly that it seems
793 : * difficult to deal with pending trigger events in such cases, and
794 : * partly that the command might imply changing the constraint's
795 : * properties as well, which doesn't seem nice.)
796 : */
797 52 : if (OidIsValid(existing_constraint_oid))
798 0 : ereport(ERROR,
799 : (errcode(ERRCODE_DUPLICATE_OBJECT),
800 : errmsg("trigger \"%s\" for relation \"%s\" is a constraint trigger",
801 : stmt->trigname, RelationGetRelationName(rel))));
802 : }
803 :
804 : /*
805 : * If it's a user-entered CREATE CONSTRAINT TRIGGER command, make a
806 : * corresponding pg_constraint entry.
807 : */
808 11420 : if (stmt->isconstraint && !OidIsValid(constraintOid))
809 : {
810 : /* Internal callers should have made their own constraints */
811 : Assert(!isInternal);
812 117 : constraintOid = CreateConstraintEntry(stmt->trigname,
813 117 : RelationGetNamespace(rel),
814 : CONSTRAINT_TRIGGER,
815 117 : stmt->deferrable,
816 117 : stmt->initdeferred,
817 : true, /* Is Enforced */
818 : true,
819 : InvalidOid, /* no parent */
820 : RelationGetRelid(rel),
821 : NULL, /* no conkey */
822 : 0,
823 : 0,
824 : InvalidOid, /* no domain */
825 : InvalidOid, /* no index */
826 : InvalidOid, /* no foreign key */
827 : NULL,
828 : NULL,
829 : NULL,
830 : NULL,
831 : 0,
832 : ' ',
833 : ' ',
834 : NULL,
835 : 0,
836 : ' ',
837 : NULL, /* no exclusion */
838 : NULL, /* no check constraint */
839 : NULL,
840 : true, /* islocal */
841 : 0, /* inhcount */
842 : true, /* noinherit */
843 : false, /* conperiod */
844 : isInternal); /* is_internal */
845 : }
846 :
847 : /*
848 : * If trigger is internally generated, modify the provided trigger name to
849 : * ensure uniqueness by appending the trigger OID. (Callers will usually
850 : * supply a simple constant trigger name in these cases.)
851 : */
852 11420 : if (isInternal)
853 : {
854 8848 : snprintf(internaltrigname, sizeof(internaltrigname),
855 8848 : "%s_%u", stmt->trigname, trigoid);
856 8848 : trigname = internaltrigname;
857 : }
858 : else
859 : {
860 : /* user-defined trigger; use the specified trigger name as-is */
861 2572 : trigname = stmt->trigname;
862 : }
863 :
864 : /*
865 : * Build the new pg_trigger tuple.
866 : */
867 11420 : memset(nulls, false, sizeof(nulls));
868 :
869 11420 : values[Anum_pg_trigger_oid - 1] = ObjectIdGetDatum(trigoid);
870 11420 : values[Anum_pg_trigger_tgrelid - 1] = ObjectIdGetDatum(RelationGetRelid(rel));
871 11420 : values[Anum_pg_trigger_tgparentid - 1] = ObjectIdGetDatum(parentTriggerOid);
872 11420 : values[Anum_pg_trigger_tgname - 1] = DirectFunctionCall1(namein,
873 : CStringGetDatum(trigname));
874 11420 : values[Anum_pg_trigger_tgfoid - 1] = ObjectIdGetDatum(funcoid);
875 11420 : values[Anum_pg_trigger_tgtype - 1] = Int16GetDatum(tgtype);
876 11420 : values[Anum_pg_trigger_tgenabled - 1] = CharGetDatum(trigger_fires_when);
877 11420 : values[Anum_pg_trigger_tgisinternal - 1] = BoolGetDatum(isInternal);
878 11420 : values[Anum_pg_trigger_tgconstrrelid - 1] = ObjectIdGetDatum(constrrelid);
879 11420 : values[Anum_pg_trigger_tgconstrindid - 1] = ObjectIdGetDatum(indexOid);
880 11420 : values[Anum_pg_trigger_tgconstraint - 1] = ObjectIdGetDatum(constraintOid);
881 11420 : values[Anum_pg_trigger_tgdeferrable - 1] = BoolGetDatum(stmt->deferrable);
882 11420 : values[Anum_pg_trigger_tginitdeferred - 1] = BoolGetDatum(stmt->initdeferred);
883 :
884 11420 : if (stmt->args)
885 : {
886 : ListCell *le;
887 : char *args;
888 457 : int16 nargs = list_length(stmt->args);
889 457 : int len = 0;
890 :
891 1088 : foreach(le, stmt->args)
892 : {
893 631 : char *ar = strVal(lfirst(le));
894 :
895 631 : len += strlen(ar) + 4;
896 4992 : for (; *ar; ar++)
897 : {
898 4361 : if (*ar == '\\')
899 0 : len++;
900 : }
901 : }
902 457 : args = (char *) palloc(len + 1);
903 457 : args[0] = '\0';
904 1088 : foreach(le, stmt->args)
905 : {
906 631 : char *s = strVal(lfirst(le));
907 631 : char *d = args + strlen(args);
908 :
909 4992 : while (*s)
910 : {
911 4361 : if (*s == '\\')
912 0 : *d++ = '\\';
913 4361 : *d++ = *s++;
914 : }
915 631 : strcpy(d, "\\000");
916 : }
917 457 : values[Anum_pg_trigger_tgnargs - 1] = Int16GetDatum(nargs);
918 457 : values[Anum_pg_trigger_tgargs - 1] = DirectFunctionCall1(byteain,
919 : CStringGetDatum(args));
920 : }
921 : else
922 : {
923 10963 : values[Anum_pg_trigger_tgnargs - 1] = Int16GetDatum(0);
924 10963 : values[Anum_pg_trigger_tgargs - 1] = DirectFunctionCall1(byteain,
925 : CStringGetDatum(""));
926 : }
927 :
928 : /* build column number array if it's a column-specific trigger */
929 11420 : ncolumns = list_length(stmt->columns);
930 11420 : if (ncolumns == 0)
931 11351 : columns = NULL;
932 : else
933 : {
934 : ListCell *cell;
935 69 : int i = 0;
936 :
937 69 : columns = (int16 *) palloc(ncolumns * sizeof(int16));
938 143 : foreach(cell, stmt->columns)
939 : {
940 78 : char *name = strVal(lfirst(cell));
941 : int16 attnum;
942 : int j;
943 :
944 : /* Lookup column name. System columns are not allowed */
945 78 : attnum = attnameAttNum(rel, name, false);
946 78 : if (attnum == InvalidAttrNumber)
947 0 : ereport(ERROR,
948 : (errcode(ERRCODE_UNDEFINED_COLUMN),
949 : errmsg("column \"%s\" of relation \"%s\" does not exist",
950 : name, RelationGetRelationName(rel))));
951 :
952 : /* Check for duplicates */
953 83 : for (j = i - 1; j >= 0; j--)
954 : {
955 9 : if (columns[j] == attnum)
956 4 : ereport(ERROR,
957 : (errcode(ERRCODE_DUPLICATE_COLUMN),
958 : errmsg("column \"%s\" specified more than once",
959 : name)));
960 : }
961 :
962 74 : columns[i++] = attnum;
963 : }
964 : }
965 11416 : tgattr = buildint2vector(columns, ncolumns);
966 11416 : values[Anum_pg_trigger_tgattr - 1] = PointerGetDatum(tgattr);
967 :
968 : /* set tgqual if trigger has WHEN clause */
969 11416 : if (qual)
970 116 : values[Anum_pg_trigger_tgqual - 1] = CStringGetTextDatum(qual);
971 : else
972 11300 : nulls[Anum_pg_trigger_tgqual - 1] = true;
973 :
974 11416 : if (oldtablename)
975 180 : values[Anum_pg_trigger_tgoldtable - 1] = DirectFunctionCall1(namein,
976 : CStringGetDatum(oldtablename));
977 : else
978 11236 : nulls[Anum_pg_trigger_tgoldtable - 1] = true;
979 11416 : if (newtablename)
980 204 : values[Anum_pg_trigger_tgnewtable - 1] = DirectFunctionCall1(namein,
981 : CStringGetDatum(newtablename));
982 : else
983 11212 : nulls[Anum_pg_trigger_tgnewtable - 1] = true;
984 :
985 : /*
986 : * Insert or replace tuple in pg_trigger.
987 : */
988 11416 : if (!trigger_exists)
989 : {
990 11364 : tuple = heap_form_tuple(tgrel->rd_att, values, nulls);
991 11364 : CatalogTupleInsert(tgrel, tuple);
992 : }
993 : else
994 : {
995 : HeapTuple newtup;
996 :
997 52 : newtup = heap_form_tuple(tgrel->rd_att, values, nulls);
998 52 : CatalogTupleUpdate(tgrel, &tuple->t_self, newtup);
999 52 : heap_freetuple(newtup);
1000 : }
1001 :
1002 11416 : heap_freetuple(tuple); /* free either original or new tuple */
1003 11416 : table_close(tgrel, RowExclusiveLock);
1004 :
1005 11416 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgname - 1]));
1006 11416 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgargs - 1]));
1007 11416 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgattr - 1]));
1008 11416 : if (oldtablename)
1009 180 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgoldtable - 1]));
1010 11416 : if (newtablename)
1011 204 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgnewtable - 1]));
1012 :
1013 : /*
1014 : * Update relation's pg_class entry; if necessary; and if not, send an SI
1015 : * message to make other backends (and this one) rebuild relcache entries.
1016 : */
1017 11416 : pgrel = table_open(RelationRelationId, RowExclusiveLock);
1018 11416 : tuple = SearchSysCacheCopy1(RELOID,
1019 : ObjectIdGetDatum(RelationGetRelid(rel)));
1020 11416 : if (!HeapTupleIsValid(tuple))
1021 0 : elog(ERROR, "cache lookup failed for relation %u",
1022 : RelationGetRelid(rel));
1023 11416 : if (!((Form_pg_class) GETSTRUCT(tuple))->relhastriggers)
1024 : {
1025 4241 : ((Form_pg_class) GETSTRUCT(tuple))->relhastriggers = true;
1026 :
1027 4241 : CatalogTupleUpdate(pgrel, &tuple->t_self, tuple);
1028 :
1029 4241 : CommandCounterIncrement();
1030 : }
1031 : else
1032 7175 : CacheInvalidateRelcacheByTuple(tuple);
1033 :
1034 11416 : heap_freetuple(tuple);
1035 11416 : table_close(pgrel, RowExclusiveLock);
1036 :
1037 : /*
1038 : * If we're replacing a trigger, flush all the old dependencies before
1039 : * recording new ones.
1040 : */
1041 11416 : if (trigger_exists)
1042 52 : deleteDependencyRecordsFor(TriggerRelationId, trigoid, true);
1043 :
1044 : /*
1045 : * Record dependencies for trigger. Always place a normal dependency on
1046 : * the function.
1047 : */
1048 11416 : myself.classId = TriggerRelationId;
1049 11416 : myself.objectId = trigoid;
1050 11416 : myself.objectSubId = 0;
1051 :
1052 11416 : referenced.classId = ProcedureRelationId;
1053 11416 : referenced.objectId = funcoid;
1054 11416 : referenced.objectSubId = 0;
1055 11416 : recordDependencyOn(&myself, &referenced, DEPENDENCY_NORMAL);
1056 :
1057 11416 : if (isInternal && OidIsValid(constraintOid))
1058 : {
1059 : /*
1060 : * Internally-generated trigger for a constraint, so make it an
1061 : * internal dependency of the constraint. We can skip depending on
1062 : * the relation(s), as there'll be an indirect dependency via the
1063 : * constraint.
1064 : */
1065 8848 : referenced.classId = ConstraintRelationId;
1066 8848 : referenced.objectId = constraintOid;
1067 8848 : referenced.objectSubId = 0;
1068 8848 : recordDependencyOn(&myself, &referenced, DEPENDENCY_INTERNAL);
1069 : }
1070 : else
1071 : {
1072 : /*
1073 : * User CREATE TRIGGER, so place dependencies. We make trigger be
1074 : * auto-dropped if its relation is dropped or if the FK relation is
1075 : * dropped. (Auto drop is compatible with our pre-7.3 behavior.)
1076 : */
1077 2568 : referenced.classId = RelationRelationId;
1078 2568 : referenced.objectId = RelationGetRelid(rel);
1079 2568 : referenced.objectSubId = 0;
1080 2568 : recordDependencyOn(&myself, &referenced, DEPENDENCY_AUTO);
1081 :
1082 2568 : if (OidIsValid(constrrelid))
1083 : {
1084 28 : referenced.classId = RelationRelationId;
1085 28 : referenced.objectId = constrrelid;
1086 28 : referenced.objectSubId = 0;
1087 28 : recordDependencyOn(&myself, &referenced, DEPENDENCY_AUTO);
1088 : }
1089 : /* Not possible to have an index dependency in this case */
1090 : Assert(!OidIsValid(indexOid));
1091 :
1092 : /*
1093 : * If it's a user-specified constraint trigger, make the constraint
1094 : * internally dependent on the trigger instead of vice versa.
1095 : */
1096 2568 : if (OidIsValid(constraintOid))
1097 : {
1098 117 : referenced.classId = ConstraintRelationId;
1099 117 : referenced.objectId = constraintOid;
1100 117 : referenced.objectSubId = 0;
1101 117 : recordDependencyOn(&referenced, &myself, DEPENDENCY_INTERNAL);
1102 : }
1103 :
1104 : /*
1105 : * If it's a partition trigger, create the partition dependencies.
1106 : */
1107 2568 : if (OidIsValid(parentTriggerOid))
1108 : {
1109 595 : ObjectAddressSet(referenced, TriggerRelationId, parentTriggerOid);
1110 595 : recordDependencyOn(&myself, &referenced, DEPENDENCY_PARTITION_PRI);
1111 595 : ObjectAddressSet(referenced, RelationRelationId, RelationGetRelid(rel));
1112 595 : recordDependencyOn(&myself, &referenced, DEPENDENCY_PARTITION_SEC);
1113 : }
1114 : }
1115 :
1116 : /* If column-specific trigger, add normal dependencies on columns */
1117 11416 : if (columns != NULL)
1118 : {
1119 : int i;
1120 :
1121 65 : referenced.classId = RelationRelationId;
1122 65 : referenced.objectId = RelationGetRelid(rel);
1123 135 : for (i = 0; i < ncolumns; i++)
1124 : {
1125 70 : referenced.objectSubId = columns[i];
1126 70 : recordDependencyOn(&myself, &referenced, DEPENDENCY_NORMAL);
1127 : }
1128 : }
1129 :
1130 : /*
1131 : * If it has a WHEN clause, add dependencies on objects mentioned in the
1132 : * expression (eg, functions, as well as any columns used).
1133 : */
1134 11416 : if (whenRtable != NIL)
1135 88 : recordDependencyOnExpr(&myself, whenClause, whenRtable,
1136 : DEPENDENCY_NORMAL);
1137 :
1138 : /* Post creation hook for new trigger */
1139 11416 : InvokeObjectPostCreateHookArg(TriggerRelationId, trigoid, 0,
1140 : isInternal);
1141 :
1142 : /*
1143 : * Lastly, create the trigger on child relations, if needed.
1144 : */
1145 11416 : if (partition_recurse)
1146 : {
1147 271 : PartitionDesc partdesc = RelationGetPartitionDesc(rel, true);
1148 : int i;
1149 : MemoryContext oldcxt,
1150 : perChildCxt;
1151 :
1152 271 : perChildCxt = AllocSetContextCreate(CurrentMemoryContext,
1153 : "part trig clone",
1154 : ALLOCSET_SMALL_SIZES);
1155 :
1156 : /*
1157 : * We don't currently expect to be called with a valid indexOid. If
1158 : * that ever changes then we'll need to write code here to find the
1159 : * corresponding child index.
1160 : */
1161 : Assert(!OidIsValid(indexOid));
1162 :
1163 271 : oldcxt = MemoryContextSwitchTo(perChildCxt);
1164 :
1165 : /* Iterate to create the trigger on each existing partition */
1166 739 : for (i = 0; i < partdesc->nparts; i++)
1167 : {
1168 : CreateTrigStmt *childStmt;
1169 : Relation childTbl;
1170 : Node *qual;
1171 :
1172 472 : childTbl = table_open(partdesc->oids[i], ShareRowExclusiveLock);
1173 :
1174 : /*
1175 : * Initialize our fabricated parse node by copying the original
1176 : * one, then resetting fields that we pass separately.
1177 : */
1178 472 : childStmt = copyObject(stmt);
1179 472 : childStmt->funcname = NIL;
1180 472 : childStmt->whenClause = NULL;
1181 :
1182 : /* If there is a WHEN clause, create a modified copy of it */
1183 472 : qual = copyObject(whenClause);
1184 : qual = (Node *)
1185 472 : map_partition_varattnos((List *) qual, PRS2_OLD_VARNO,
1186 : childTbl, rel);
1187 : qual = (Node *)
1188 472 : map_partition_varattnos((List *) qual, PRS2_NEW_VARNO,
1189 : childTbl, rel);
1190 :
1191 472 : CreateTriggerFiringOn(childStmt, queryString,
1192 472 : partdesc->oids[i], refRelOid,
1193 : InvalidOid, InvalidOid,
1194 : funcoid, trigoid, qual,
1195 : isInternal, true, trigger_fires_when);
1196 :
1197 468 : table_close(childTbl, NoLock);
1198 :
1199 468 : MemoryContextReset(perChildCxt);
1200 : }
1201 :
1202 267 : MemoryContextSwitchTo(oldcxt);
1203 267 : MemoryContextDelete(perChildCxt);
1204 : }
1205 :
1206 : /* Keep lock on target rel until end of xact */
1207 11412 : table_close(rel, NoLock);
1208 :
1209 11412 : return myself;
1210 : }
1211 :
1212 : /*
1213 : * TriggerSetParentTrigger
1214 : * Set a partition's trigger as child of its parent trigger,
1215 : * or remove the linkage if parentTrigId is InvalidOid.
1216 : *
1217 : * This updates the constraint's pg_trigger row to show it as inherited, and
1218 : * adds PARTITION dependencies to prevent the trigger from being deleted
1219 : * on its own. Alternatively, reverse that.
1220 : */
1221 : void
1222 344 : TriggerSetParentTrigger(Relation trigRel,
1223 : Oid childTrigId,
1224 : Oid parentTrigId,
1225 : Oid childTableId)
1226 : {
1227 : SysScanDesc tgscan;
1228 : ScanKeyData skey[1];
1229 : Form_pg_trigger trigForm;
1230 : HeapTuple tuple,
1231 : newtup;
1232 : ObjectAddress depender;
1233 : ObjectAddress referenced;
1234 :
1235 : /*
1236 : * Find the trigger to delete.
1237 : */
1238 344 : ScanKeyInit(&skey[0],
1239 : Anum_pg_trigger_oid,
1240 : BTEqualStrategyNumber, F_OIDEQ,
1241 : ObjectIdGetDatum(childTrigId));
1242 :
1243 344 : tgscan = systable_beginscan(trigRel, TriggerOidIndexId, true,
1244 : NULL, 1, skey);
1245 :
1246 344 : tuple = systable_getnext(tgscan);
1247 344 : if (!HeapTupleIsValid(tuple))
1248 0 : elog(ERROR, "could not find tuple for trigger %u", childTrigId);
1249 344 : newtup = heap_copytuple(tuple);
1250 344 : trigForm = (Form_pg_trigger) GETSTRUCT(newtup);
1251 344 : if (OidIsValid(parentTrigId))
1252 : {
1253 : /* don't allow setting parent for a constraint that already has one */
1254 200 : if (OidIsValid(trigForm->tgparentid))
1255 0 : elog(ERROR, "trigger %u already has a parent trigger",
1256 : childTrigId);
1257 :
1258 200 : trigForm->tgparentid = parentTrigId;
1259 :
1260 200 : CatalogTupleUpdate(trigRel, &tuple->t_self, newtup);
1261 :
1262 200 : ObjectAddressSet(depender, TriggerRelationId, childTrigId);
1263 :
1264 200 : ObjectAddressSet(referenced, TriggerRelationId, parentTrigId);
1265 200 : recordDependencyOn(&depender, &referenced, DEPENDENCY_PARTITION_PRI);
1266 :
1267 200 : ObjectAddressSet(referenced, RelationRelationId, childTableId);
1268 200 : recordDependencyOn(&depender, &referenced, DEPENDENCY_PARTITION_SEC);
1269 : }
1270 : else
1271 : {
1272 144 : trigForm->tgparentid = InvalidOid;
1273 :
1274 144 : CatalogTupleUpdate(trigRel, &tuple->t_self, newtup);
1275 :
1276 144 : deleteDependencyRecordsForClass(TriggerRelationId, childTrigId,
1277 : TriggerRelationId,
1278 : DEPENDENCY_PARTITION_PRI);
1279 144 : deleteDependencyRecordsForClass(TriggerRelationId, childTrigId,
1280 : RelationRelationId,
1281 : DEPENDENCY_PARTITION_SEC);
1282 : }
1283 :
1284 344 : heap_freetuple(newtup);
1285 344 : systable_endscan(tgscan);
1286 344 : }
1287 :
1288 :
1289 : /*
1290 : * Guts of trigger deletion.
1291 : */
1292 : void
1293 9721 : RemoveTriggerById(Oid trigOid)
1294 : {
1295 : Relation tgrel;
1296 : SysScanDesc tgscan;
1297 : ScanKeyData skey[1];
1298 : HeapTuple tup;
1299 : Oid relid;
1300 : Relation rel;
1301 :
1302 9721 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1303 :
1304 : /*
1305 : * Find the trigger to delete.
1306 : */
1307 9721 : ScanKeyInit(&skey[0],
1308 : Anum_pg_trigger_oid,
1309 : BTEqualStrategyNumber, F_OIDEQ,
1310 : ObjectIdGetDatum(trigOid));
1311 :
1312 9721 : tgscan = systable_beginscan(tgrel, TriggerOidIndexId, true,
1313 : NULL, 1, skey);
1314 :
1315 9721 : tup = systable_getnext(tgscan);
1316 9721 : if (!HeapTupleIsValid(tup))
1317 0 : elog(ERROR, "could not find tuple for trigger %u", trigOid);
1318 :
1319 : /*
1320 : * Open and exclusive-lock the relation the trigger belongs to.
1321 : */
1322 9721 : relid = ((Form_pg_trigger) GETSTRUCT(tup))->tgrelid;
1323 :
1324 9721 : rel = table_open(relid, AccessExclusiveLock);
1325 :
1326 9721 : if (rel->rd_rel->relkind != RELKIND_RELATION &&
1327 1763 : rel->rd_rel->relkind != RELKIND_VIEW &&
1328 1673 : rel->rd_rel->relkind != RELKIND_FOREIGN_TABLE &&
1329 1623 : rel->rd_rel->relkind != RELKIND_PARTITIONED_TABLE)
1330 0 : ereport(ERROR,
1331 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
1332 : errmsg("relation \"%s\" cannot have triggers",
1333 : RelationGetRelationName(rel)),
1334 : errdetail_relkind_not_supported(rel->rd_rel->relkind)));
1335 :
1336 9721 : if (!allowSystemTableMods && IsSystemRelation(rel))
1337 0 : ereport(ERROR,
1338 : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1339 : errmsg("permission denied: \"%s\" is a system catalog",
1340 : RelationGetRelationName(rel))));
1341 :
1342 : /*
1343 : * Delete the pg_trigger tuple.
1344 : */
1345 9721 : CatalogTupleDelete(tgrel, &tup->t_self);
1346 :
1347 9721 : systable_endscan(tgscan);
1348 9721 : table_close(tgrel, RowExclusiveLock);
1349 :
1350 : /*
1351 : * We do not bother to try to determine whether any other triggers remain,
1352 : * which would be needed in order to decide whether it's safe to clear the
1353 : * relation's relhastriggers. (In any case, there might be a concurrent
1354 : * process adding new triggers.) Instead, just force a relcache inval to
1355 : * make other backends (and this one too!) rebuild their relcache entries.
1356 : * There's no great harm in leaving relhastriggers true even if there are
1357 : * no triggers left.
1358 : */
1359 9721 : CacheInvalidateRelcache(rel);
1360 :
1361 : /* Keep lock on trigger's rel until end of xact */
1362 9721 : table_close(rel, NoLock);
1363 9721 : }
1364 :
1365 : /*
1366 : * get_trigger_oid - Look up a trigger by name to find its OID.
1367 : *
1368 : * If missing_ok is false, throw an error if trigger not found. If
1369 : * true, just return InvalidOid.
1370 : */
1371 : Oid
1372 512 : get_trigger_oid(Oid relid, const char *trigname, bool missing_ok)
1373 : {
1374 : Relation tgrel;
1375 : ScanKeyData skey[2];
1376 : SysScanDesc tgscan;
1377 : HeapTuple tup;
1378 : Oid oid;
1379 :
1380 : /*
1381 : * Find the trigger, verify permissions, set up object address
1382 : */
1383 512 : tgrel = table_open(TriggerRelationId, AccessShareLock);
1384 :
1385 512 : ScanKeyInit(&skey[0],
1386 : Anum_pg_trigger_tgrelid,
1387 : BTEqualStrategyNumber, F_OIDEQ,
1388 : ObjectIdGetDatum(relid));
1389 512 : ScanKeyInit(&skey[1],
1390 : Anum_pg_trigger_tgname,
1391 : BTEqualStrategyNumber, F_NAMEEQ,
1392 : CStringGetDatum(trigname));
1393 :
1394 512 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1395 : NULL, 2, skey);
1396 :
1397 512 : tup = systable_getnext(tgscan);
1398 :
1399 512 : if (!HeapTupleIsValid(tup))
1400 : {
1401 20 : if (!missing_ok)
1402 16 : ereport(ERROR,
1403 : (errcode(ERRCODE_UNDEFINED_OBJECT),
1404 : errmsg("trigger \"%s\" for table \"%s\" does not exist",
1405 : trigname, get_rel_name(relid))));
1406 4 : oid = InvalidOid;
1407 : }
1408 : else
1409 : {
1410 492 : oid = ((Form_pg_trigger) GETSTRUCT(tup))->oid;
1411 : }
1412 :
1413 496 : systable_endscan(tgscan);
1414 496 : table_close(tgrel, AccessShareLock);
1415 496 : return oid;
1416 : }
1417 :
1418 : /*
1419 : * Perform permissions and integrity checks before acquiring a relation lock.
1420 : */
1421 : static void
1422 27 : RangeVarCallbackForRenameTrigger(const RangeVar *rv, Oid relid, Oid oldrelid,
1423 : void *arg)
1424 : {
1425 : HeapTuple tuple;
1426 : Form_pg_class form;
1427 :
1428 27 : tuple = SearchSysCache1(RELOID, ObjectIdGetDatum(relid));
1429 27 : if (!HeapTupleIsValid(tuple))
1430 0 : return; /* concurrently dropped */
1431 27 : form = (Form_pg_class) GETSTRUCT(tuple);
1432 :
1433 : /* only tables and views can have triggers */
1434 27 : if (form->relkind != RELKIND_RELATION && form->relkind != RELKIND_VIEW &&
1435 16 : form->relkind != RELKIND_FOREIGN_TABLE &&
1436 16 : form->relkind != RELKIND_PARTITIONED_TABLE)
1437 0 : ereport(ERROR,
1438 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
1439 : errmsg("relation \"%s\" cannot have triggers",
1440 : rv->relname),
1441 : errdetail_relkind_not_supported(form->relkind)));
1442 :
1443 : /* you must own the table to rename one of its triggers */
1444 27 : if (!object_ownercheck(RelationRelationId, relid, GetUserId()))
1445 0 : aclcheck_error(ACLCHECK_NOT_OWNER, get_relkind_objtype(get_rel_relkind(relid)), rv->relname);
1446 27 : if (!allowSystemTableMods && IsSystemClass(relid, form))
1447 1 : ereport(ERROR,
1448 : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1449 : errmsg("permission denied: \"%s\" is a system catalog",
1450 : rv->relname)));
1451 :
1452 26 : ReleaseSysCache(tuple);
1453 : }
1454 :
1455 : /*
1456 : * renametrig - changes the name of a trigger on a relation
1457 : *
1458 : * trigger name is changed in trigger catalog.
1459 : * No record of the previous name is kept.
1460 : *
1461 : * get proper relrelation from relation catalog (if not arg)
1462 : * scan trigger catalog
1463 : * for name conflict (within rel)
1464 : * for original trigger (if not arg)
1465 : * modify tgname in trigger tuple
1466 : * update row in catalog
1467 : */
1468 : ObjectAddress
1469 26 : renametrig(RenameStmt *stmt)
1470 : {
1471 : Oid tgoid;
1472 : Relation targetrel;
1473 : Relation tgrel;
1474 : HeapTuple tuple;
1475 : SysScanDesc tgscan;
1476 : ScanKeyData key[2];
1477 : Oid relid;
1478 : ObjectAddress address;
1479 :
1480 : /*
1481 : * Look up name, check permissions, and acquire lock (which we will NOT
1482 : * release until end of transaction).
1483 : */
1484 26 : relid = RangeVarGetRelidExtended(stmt->relation, AccessExclusiveLock,
1485 : 0,
1486 : RangeVarCallbackForRenameTrigger,
1487 : NULL);
1488 :
1489 : /* Have lock already, so just need to build relcache entry. */
1490 25 : targetrel = relation_open(relid, NoLock);
1491 :
1492 : /*
1493 : * On partitioned tables, this operation recurses to partitions. Lock all
1494 : * tables upfront.
1495 : */
1496 25 : if (targetrel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1497 16 : (void) find_all_inheritors(relid, AccessExclusiveLock, NULL);
1498 :
1499 25 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1500 :
1501 : /*
1502 : * Search for the trigger to modify.
1503 : */
1504 25 : ScanKeyInit(&key[0],
1505 : Anum_pg_trigger_tgrelid,
1506 : BTEqualStrategyNumber, F_OIDEQ,
1507 : ObjectIdGetDatum(relid));
1508 25 : ScanKeyInit(&key[1],
1509 : Anum_pg_trigger_tgname,
1510 : BTEqualStrategyNumber, F_NAMEEQ,
1511 25 : PointerGetDatum(stmt->subname));
1512 25 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1513 : NULL, 2, key);
1514 25 : if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1515 : {
1516 : Form_pg_trigger trigform;
1517 :
1518 25 : trigform = (Form_pg_trigger) GETSTRUCT(tuple);
1519 25 : tgoid = trigform->oid;
1520 :
1521 : /*
1522 : * If the trigger descends from a trigger on a parent partitioned
1523 : * table, reject the rename. We don't allow a trigger in a partition
1524 : * to differ in name from that of its parent: that would lead to an
1525 : * inconsistency that pg_dump would not reproduce.
1526 : */
1527 25 : if (OidIsValid(trigform->tgparentid))
1528 4 : ereport(ERROR,
1529 : errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
1530 : errmsg("cannot rename trigger \"%s\" on table \"%s\"",
1531 : stmt->subname, RelationGetRelationName(targetrel)),
1532 : errhint("Rename the trigger on the partitioned table \"%s\" instead.",
1533 : get_rel_name(get_partition_parent(relid, false))));
1534 :
1535 :
1536 : /* Rename the trigger on this relation ... */
1537 21 : renametrig_internal(tgrel, targetrel, tuple, stmt->newname,
1538 21 : stmt->subname);
1539 :
1540 : /* ... and if it is partitioned, recurse to its partitions */
1541 21 : if (targetrel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1542 : {
1543 12 : PartitionDesc partdesc = RelationGetPartitionDesc(targetrel, true);
1544 :
1545 20 : for (int i = 0; i < partdesc->nparts; i++)
1546 : {
1547 12 : Oid partitionId = partdesc->oids[i];
1548 :
1549 12 : renametrig_partition(tgrel, partitionId, trigform->oid,
1550 12 : stmt->newname, stmt->subname);
1551 : }
1552 : }
1553 : }
1554 : else
1555 : {
1556 0 : ereport(ERROR,
1557 : (errcode(ERRCODE_UNDEFINED_OBJECT),
1558 : errmsg("trigger \"%s\" for table \"%s\" does not exist",
1559 : stmt->subname, RelationGetRelationName(targetrel))));
1560 : }
1561 :
1562 17 : ObjectAddressSet(address, TriggerRelationId, tgoid);
1563 :
1564 17 : systable_endscan(tgscan);
1565 :
1566 17 : table_close(tgrel, RowExclusiveLock);
1567 :
1568 : /*
1569 : * Close rel, but keep exclusive lock!
1570 : */
1571 17 : relation_close(targetrel, NoLock);
1572 :
1573 17 : return address;
1574 : }
1575 :
1576 : /*
1577 : * Subroutine for renametrig -- perform the actual work of renaming one
1578 : * trigger on one table.
1579 : *
1580 : * If the trigger has a name different from the expected one, raise a
1581 : * NOTICE about it.
1582 : */
1583 : static void
1584 37 : renametrig_internal(Relation tgrel, Relation targetrel, HeapTuple trigtup,
1585 : const char *newname, const char *expected_name)
1586 : {
1587 : HeapTuple tuple;
1588 : Form_pg_trigger tgform;
1589 : ScanKeyData key[2];
1590 : SysScanDesc tgscan;
1591 :
1592 : /* If the trigger already has the new name, nothing to do. */
1593 37 : tgform = (Form_pg_trigger) GETSTRUCT(trigtup);
1594 37 : if (strcmp(NameStr(tgform->tgname), newname) == 0)
1595 0 : return;
1596 :
1597 : /*
1598 : * Before actually trying the rename, search for triggers with the same
1599 : * name. The update would fail with an ugly message in that case, and it
1600 : * is better to throw a nicer error.
1601 : */
1602 37 : ScanKeyInit(&key[0],
1603 : Anum_pg_trigger_tgrelid,
1604 : BTEqualStrategyNumber, F_OIDEQ,
1605 : ObjectIdGetDatum(RelationGetRelid(targetrel)));
1606 37 : ScanKeyInit(&key[1],
1607 : Anum_pg_trigger_tgname,
1608 : BTEqualStrategyNumber, F_NAMEEQ,
1609 : PointerGetDatum(newname));
1610 37 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1611 : NULL, 2, key);
1612 37 : if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1613 4 : ereport(ERROR,
1614 : (errcode(ERRCODE_DUPLICATE_OBJECT),
1615 : errmsg("trigger \"%s\" for relation \"%s\" already exists",
1616 : newname, RelationGetRelationName(targetrel))));
1617 33 : systable_endscan(tgscan);
1618 :
1619 : /*
1620 : * The target name is free; update the existing pg_trigger tuple with it.
1621 : */
1622 33 : tuple = heap_copytuple(trigtup); /* need a modifiable copy */
1623 33 : tgform = (Form_pg_trigger) GETSTRUCT(tuple);
1624 :
1625 : /*
1626 : * If the trigger has a name different from what we expected, let the user
1627 : * know. (We can proceed anyway, since we must have reached here following
1628 : * a tgparentid link.)
1629 : */
1630 33 : if (strcmp(NameStr(tgform->tgname), expected_name) != 0)
1631 0 : ereport(NOTICE,
1632 : errmsg("renamed trigger \"%s\" on relation \"%s\"",
1633 : NameStr(tgform->tgname),
1634 : RelationGetRelationName(targetrel)));
1635 :
1636 33 : namestrcpy(&tgform->tgname, newname);
1637 :
1638 33 : CatalogTupleUpdate(tgrel, &tuple->t_self, tuple);
1639 :
1640 33 : InvokeObjectPostAlterHook(TriggerRelationId, tgform->oid, 0);
1641 :
1642 : /*
1643 : * Invalidate relation's relcache entry so that other backends (and this
1644 : * one too!) are sent SI message to make them rebuild relcache entries.
1645 : * (Ideally this should happen automatically...)
1646 : */
1647 33 : CacheInvalidateRelcache(targetrel);
1648 : }
1649 :
1650 : /*
1651 : * Subroutine for renametrig -- Helper for recursing to partitions when
1652 : * renaming triggers on a partitioned table.
1653 : */
1654 : static void
1655 20 : renametrig_partition(Relation tgrel, Oid partitionId, Oid parentTriggerOid,
1656 : const char *newname, const char *expected_name)
1657 : {
1658 : SysScanDesc tgscan;
1659 : ScanKeyData key;
1660 : HeapTuple tuple;
1661 :
1662 : /*
1663 : * Given a relation and the OID of a trigger on parent relation, find the
1664 : * corresponding trigger in the child and rename that trigger to the given
1665 : * name.
1666 : */
1667 20 : ScanKeyInit(&key,
1668 : Anum_pg_trigger_tgrelid,
1669 : BTEqualStrategyNumber, F_OIDEQ,
1670 : ObjectIdGetDatum(partitionId));
1671 20 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1672 : NULL, 1, &key);
1673 32 : while (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1674 : {
1675 28 : Form_pg_trigger tgform = (Form_pg_trigger) GETSTRUCT(tuple);
1676 : Relation partitionRel;
1677 :
1678 28 : if (tgform->tgparentid != parentTriggerOid)
1679 12 : continue; /* not our trigger */
1680 :
1681 16 : partitionRel = table_open(partitionId, NoLock);
1682 :
1683 : /* Rename the trigger on this partition */
1684 16 : renametrig_internal(tgrel, partitionRel, tuple, newname, expected_name);
1685 :
1686 : /* And if this relation is partitioned, recurse to its partitions */
1687 12 : if (partitionRel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1688 : {
1689 4 : PartitionDesc partdesc = RelationGetPartitionDesc(partitionRel,
1690 : true);
1691 :
1692 12 : for (int i = 0; i < partdesc->nparts; i++)
1693 : {
1694 8 : Oid partoid = partdesc->oids[i];
1695 :
1696 8 : renametrig_partition(tgrel, partoid, tgform->oid, newname,
1697 8 : NameStr(tgform->tgname));
1698 : }
1699 : }
1700 12 : table_close(partitionRel, NoLock);
1701 :
1702 : /* There should be at most one matching tuple */
1703 12 : break;
1704 : }
1705 16 : systable_endscan(tgscan);
1706 16 : }
1707 :
1708 : /*
1709 : * EnableDisableTrigger()
1710 : *
1711 : * Called by ALTER TABLE ENABLE/DISABLE [ REPLICA | ALWAYS ] TRIGGER
1712 : * to change 'tgenabled' field for the specified trigger(s)
1713 : *
1714 : * rel: relation to process (caller must hold suitable lock on it)
1715 : * tgname: name of trigger to process, or NULL to scan all triggers
1716 : * tgparent: if not zero, process only triggers with this tgparentid
1717 : * fires_when: new value for tgenabled field. In addition to generic
1718 : * enablement/disablement, this also defines when the trigger
1719 : * should be fired in session replication roles.
1720 : * skip_system: if true, skip "system" triggers (constraint triggers)
1721 : * recurse: if true, recurse to partitions
1722 : *
1723 : * Caller should have checked permissions for the table; here we also
1724 : * enforce that superuser privilege is required to alter the state of
1725 : * system triggers
1726 : */
1727 : void
1728 264 : EnableDisableTrigger(Relation rel, const char *tgname, Oid tgparent,
1729 : char fires_when, bool skip_system, bool recurse,
1730 : LOCKMODE lockmode)
1731 : {
1732 : Relation tgrel;
1733 : int nkeys;
1734 : ScanKeyData keys[2];
1735 : SysScanDesc tgscan;
1736 : HeapTuple tuple;
1737 : bool found;
1738 : bool changed;
1739 :
1740 : /* Scan the relevant entries in pg_triggers */
1741 264 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1742 :
1743 264 : ScanKeyInit(&keys[0],
1744 : Anum_pg_trigger_tgrelid,
1745 : BTEqualStrategyNumber, F_OIDEQ,
1746 : ObjectIdGetDatum(RelationGetRelid(rel)));
1747 264 : if (tgname)
1748 : {
1749 175 : ScanKeyInit(&keys[1],
1750 : Anum_pg_trigger_tgname,
1751 : BTEqualStrategyNumber, F_NAMEEQ,
1752 : CStringGetDatum(tgname));
1753 175 : nkeys = 2;
1754 : }
1755 : else
1756 89 : nkeys = 1;
1757 :
1758 264 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1759 : NULL, nkeys, keys);
1760 :
1761 264 : found = changed = false;
1762 :
1763 712 : while (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1764 : {
1765 448 : Form_pg_trigger oldtrig = (Form_pg_trigger) GETSTRUCT(tuple);
1766 :
1767 448 : if (OidIsValid(tgparent) && tgparent != oldtrig->tgparentid)
1768 128 : continue;
1769 :
1770 320 : if (oldtrig->tgisinternal)
1771 : {
1772 : /* system trigger ... ok to process? */
1773 48 : if (skip_system)
1774 8 : continue;
1775 40 : if (!superuser())
1776 0 : ereport(ERROR,
1777 : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1778 : errmsg("permission denied: \"%s\" is a system trigger",
1779 : NameStr(oldtrig->tgname))));
1780 : }
1781 :
1782 312 : found = true;
1783 :
1784 312 : if (oldtrig->tgenabled != fires_when)
1785 : {
1786 : /* need to change this one ... make a copy to scribble on */
1787 292 : HeapTuple newtup = heap_copytuple(tuple);
1788 292 : Form_pg_trigger newtrig = (Form_pg_trigger) GETSTRUCT(newtup);
1789 :
1790 292 : newtrig->tgenabled = fires_when;
1791 :
1792 292 : CatalogTupleUpdate(tgrel, &newtup->t_self, newtup);
1793 :
1794 292 : heap_freetuple(newtup);
1795 :
1796 292 : changed = true;
1797 : }
1798 :
1799 : /*
1800 : * When altering FOR EACH ROW triggers on a partitioned table, do the
1801 : * same on the partitions as well, unless ONLY is specified.
1802 : *
1803 : * Note that we recurse even if we didn't change the trigger above,
1804 : * because the partitions' copy of the trigger may have a different
1805 : * value of tgenabled than the parent's trigger and thus might need to
1806 : * be changed.
1807 : */
1808 312 : if (recurse &&
1809 295 : rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE &&
1810 57 : (TRIGGER_FOR_ROW(oldtrig->tgtype)))
1811 : {
1812 49 : PartitionDesc partdesc = RelationGetPartitionDesc(rel, true);
1813 : int i;
1814 :
1815 122 : for (i = 0; i < partdesc->nparts; i++)
1816 : {
1817 : Relation part;
1818 :
1819 73 : part = relation_open(partdesc->oids[i], lockmode);
1820 : /* Match on child triggers' tgparentid, not their name */
1821 73 : EnableDisableTrigger(part, NULL, oldtrig->oid,
1822 : fires_when, skip_system, recurse,
1823 : lockmode);
1824 73 : table_close(part, NoLock); /* keep lock till commit */
1825 : }
1826 : }
1827 :
1828 312 : InvokeObjectPostAlterHook(TriggerRelationId,
1829 : oldtrig->oid, 0);
1830 : }
1831 :
1832 264 : systable_endscan(tgscan);
1833 :
1834 264 : table_close(tgrel, RowExclusiveLock);
1835 :
1836 264 : if (tgname && !found)
1837 0 : ereport(ERROR,
1838 : (errcode(ERRCODE_UNDEFINED_OBJECT),
1839 : errmsg("trigger \"%s\" for table \"%s\" does not exist",
1840 : tgname, RelationGetRelationName(rel))));
1841 :
1842 : /*
1843 : * If we changed anything, broadcast a SI inval message to force each
1844 : * backend (including our own!) to rebuild relation's relcache entry.
1845 : * Otherwise they will fail to apply the change promptly.
1846 : */
1847 264 : if (changed)
1848 260 : CacheInvalidateRelcache(rel);
1849 264 : }
1850 :
1851 :
1852 : /*
1853 : * Build trigger data to attach to the given relcache entry.
1854 : *
1855 : * Note that trigger data attached to a relcache entry must be stored in
1856 : * CacheMemoryContext to ensure it survives as long as the relcache entry.
1857 : * But we should be running in a less long-lived working context. To avoid
1858 : * leaking cache memory if this routine fails partway through, we build a
1859 : * temporary TriggerDesc in working memory and then copy the completed
1860 : * structure into cache memory.
1861 : */
1862 : void
1863 42220 : RelationBuildTriggers(Relation relation)
1864 : {
1865 : TriggerDesc *trigdesc;
1866 : int numtrigs;
1867 : int maxtrigs;
1868 : Trigger *triggers;
1869 : Relation tgrel;
1870 : ScanKeyData skey;
1871 : SysScanDesc tgscan;
1872 : HeapTuple htup;
1873 : MemoryContext oldContext;
1874 : int i;
1875 :
1876 : /*
1877 : * Allocate a working array to hold the triggers (the array is extended if
1878 : * necessary)
1879 : */
1880 42220 : maxtrigs = 16;
1881 42220 : triggers = (Trigger *) palloc(maxtrigs * sizeof(Trigger));
1882 42220 : numtrigs = 0;
1883 :
1884 : /*
1885 : * Note: since we scan the triggers using TriggerRelidNameIndexId, we will
1886 : * be reading the triggers in name order, except possibly during
1887 : * emergency-recovery operations (ie, IgnoreSystemIndexes). This in turn
1888 : * ensures that triggers will be fired in name order.
1889 : */
1890 42220 : ScanKeyInit(&skey,
1891 : Anum_pg_trigger_tgrelid,
1892 : BTEqualStrategyNumber, F_OIDEQ,
1893 : ObjectIdGetDatum(RelationGetRelid(relation)));
1894 :
1895 42220 : tgrel = table_open(TriggerRelationId, AccessShareLock);
1896 42220 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1897 : NULL, 1, &skey);
1898 :
1899 118568 : while (HeapTupleIsValid(htup = systable_getnext(tgscan)))
1900 : {
1901 76348 : Form_pg_trigger pg_trigger = (Form_pg_trigger) GETSTRUCT(htup);
1902 : Trigger *build;
1903 : Datum datum;
1904 : bool isnull;
1905 :
1906 76348 : if (numtrigs >= maxtrigs)
1907 : {
1908 32 : maxtrigs *= 2;
1909 32 : triggers = (Trigger *) repalloc(triggers, maxtrigs * sizeof(Trigger));
1910 : }
1911 76348 : build = &(triggers[numtrigs]);
1912 :
1913 76348 : build->tgoid = pg_trigger->oid;
1914 76348 : build->tgname = DatumGetCString(DirectFunctionCall1(nameout,
1915 : NameGetDatum(&pg_trigger->tgname)));
1916 76348 : build->tgfoid = pg_trigger->tgfoid;
1917 76348 : build->tgtype = pg_trigger->tgtype;
1918 76348 : build->tgenabled = pg_trigger->tgenabled;
1919 76348 : build->tgisinternal = pg_trigger->tgisinternal;
1920 76348 : build->tgisclone = OidIsValid(pg_trigger->tgparentid);
1921 76348 : build->tgconstrrelid = pg_trigger->tgconstrrelid;
1922 76348 : build->tgconstrindid = pg_trigger->tgconstrindid;
1923 76348 : build->tgconstraint = pg_trigger->tgconstraint;
1924 76348 : build->tgdeferrable = pg_trigger->tgdeferrable;
1925 76348 : build->tginitdeferred = pg_trigger->tginitdeferred;
1926 76348 : build->tgnargs = pg_trigger->tgnargs;
1927 : /* tgattr is first var-width field, so OK to access directly */
1928 76348 : build->tgnattr = pg_trigger->tgattr.dim1;
1929 76348 : if (build->tgnattr > 0)
1930 : {
1931 342 : build->tgattr = (int16 *) palloc(build->tgnattr * sizeof(int16));
1932 342 : memcpy(build->tgattr, &(pg_trigger->tgattr.values),
1933 342 : build->tgnattr * sizeof(int16));
1934 : }
1935 : else
1936 76006 : build->tgattr = NULL;
1937 76348 : if (build->tgnargs > 0)
1938 : {
1939 : bytea *val;
1940 : char *p;
1941 :
1942 2788 : val = DatumGetByteaPP(fastgetattr(htup,
1943 : Anum_pg_trigger_tgargs,
1944 : tgrel->rd_att, &isnull));
1945 2788 : if (isnull)
1946 0 : elog(ERROR, "tgargs is null in trigger for relation \"%s\"",
1947 : RelationGetRelationName(relation));
1948 2788 : p = (char *) VARDATA_ANY(val);
1949 2788 : build->tgargs = (char **) palloc(build->tgnargs * sizeof(char *));
1950 6387 : for (i = 0; i < build->tgnargs; i++)
1951 : {
1952 3599 : build->tgargs[i] = pstrdup(p);
1953 3599 : p += strlen(p) + 1;
1954 : }
1955 : }
1956 : else
1957 73560 : build->tgargs = NULL;
1958 :
1959 76348 : datum = fastgetattr(htup, Anum_pg_trigger_tgoldtable,
1960 : tgrel->rd_att, &isnull);
1961 76348 : if (!isnull)
1962 755 : build->tgoldtable =
1963 755 : DatumGetCString(DirectFunctionCall1(nameout, datum));
1964 : else
1965 75593 : build->tgoldtable = NULL;
1966 :
1967 76348 : datum = fastgetattr(htup, Anum_pg_trigger_tgnewtable,
1968 : tgrel->rd_att, &isnull);
1969 76348 : if (!isnull)
1970 992 : build->tgnewtable =
1971 992 : DatumGetCString(DirectFunctionCall1(nameout, datum));
1972 : else
1973 75356 : build->tgnewtable = NULL;
1974 :
1975 76348 : datum = fastgetattr(htup, Anum_pg_trigger_tgqual,
1976 : tgrel->rd_att, &isnull);
1977 76348 : if (!isnull)
1978 577 : build->tgqual = TextDatumGetCString(datum);
1979 : else
1980 75771 : build->tgqual = NULL;
1981 :
1982 76348 : numtrigs++;
1983 : }
1984 :
1985 42220 : systable_endscan(tgscan);
1986 42220 : table_close(tgrel, AccessShareLock);
1987 :
1988 : /* There might not be any triggers */
1989 42220 : if (numtrigs == 0)
1990 : {
1991 9806 : pfree(triggers);
1992 9806 : return;
1993 : }
1994 :
1995 : /* Build trigdesc */
1996 32414 : trigdesc = palloc0_object(TriggerDesc);
1997 32414 : trigdesc->triggers = triggers;
1998 32414 : trigdesc->numtriggers = numtrigs;
1999 108762 : for (i = 0; i < numtrigs; i++)
2000 76348 : SetTriggerFlags(trigdesc, &(triggers[i]));
2001 :
2002 : /* Copy completed trigdesc into cache storage */
2003 32414 : oldContext = MemoryContextSwitchTo(CacheMemoryContext);
2004 32414 : relation->trigdesc = CopyTriggerDesc(trigdesc);
2005 32414 : MemoryContextSwitchTo(oldContext);
2006 :
2007 : /* Release working memory */
2008 32414 : FreeTriggerDesc(trigdesc);
2009 : }
2010 :
2011 : /*
2012 : * Update the TriggerDesc's hint flags to include the specified trigger
2013 : */
2014 : static void
2015 76348 : SetTriggerFlags(TriggerDesc *trigdesc, Trigger *trigger)
2016 : {
2017 76348 : int16 tgtype = trigger->tgtype;
2018 :
2019 76348 : trigdesc->trig_insert_before_row |=
2020 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2021 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_INSERT);
2022 76348 : trigdesc->trig_insert_after_row |=
2023 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2024 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_INSERT);
2025 76348 : trigdesc->trig_insert_instead_row |=
2026 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2027 : TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_INSERT);
2028 76348 : trigdesc->trig_insert_before_statement |=
2029 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2030 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_INSERT);
2031 76348 : trigdesc->trig_insert_after_statement |=
2032 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2033 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_INSERT);
2034 76348 : trigdesc->trig_update_before_row |=
2035 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2036 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_UPDATE);
2037 76348 : trigdesc->trig_update_after_row |=
2038 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2039 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_UPDATE);
2040 76348 : trigdesc->trig_update_instead_row |=
2041 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2042 : TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_UPDATE);
2043 76348 : trigdesc->trig_update_before_statement |=
2044 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2045 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_UPDATE);
2046 76348 : trigdesc->trig_update_after_statement |=
2047 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2048 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_UPDATE);
2049 76348 : trigdesc->trig_delete_before_row |=
2050 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2051 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_DELETE);
2052 76348 : trigdesc->trig_delete_after_row |=
2053 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2054 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_DELETE);
2055 76348 : trigdesc->trig_delete_instead_row |=
2056 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2057 : TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_DELETE);
2058 76348 : trigdesc->trig_delete_before_statement |=
2059 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2060 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_DELETE);
2061 76348 : trigdesc->trig_delete_after_statement |=
2062 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2063 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_DELETE);
2064 : /* there are no row-level truncate triggers */
2065 76348 : trigdesc->trig_truncate_before_statement |=
2066 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2067 : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_TRUNCATE);
2068 76348 : trigdesc->trig_truncate_after_statement |=
2069 76348 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2070 : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_TRUNCATE);
2071 :
2072 152696 : trigdesc->trig_insert_new_table |=
2073 101819 : (TRIGGER_FOR_INSERT(tgtype) &&
2074 25471 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgnewtable));
2075 152696 : trigdesc->trig_update_old_table |=
2076 110756 : (TRIGGER_FOR_UPDATE(tgtype) &&
2077 34408 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgoldtable));
2078 152696 : trigdesc->trig_update_new_table |=
2079 110756 : (TRIGGER_FOR_UPDATE(tgtype) &&
2080 34408 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgnewtable));
2081 152696 : trigdesc->trig_delete_old_table |=
2082 97416 : (TRIGGER_FOR_DELETE(tgtype) &&
2083 21068 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgoldtable));
2084 76348 : }
2085 :
2086 : /*
2087 : * Copy a TriggerDesc data structure.
2088 : *
2089 : * The copy is allocated in the current memory context.
2090 : */
2091 : TriggerDesc *
2092 287826 : CopyTriggerDesc(TriggerDesc *trigdesc)
2093 : {
2094 : TriggerDesc *newdesc;
2095 : Trigger *trigger;
2096 : int i;
2097 :
2098 287826 : if (trigdesc == NULL || trigdesc->numtriggers <= 0)
2099 243019 : return NULL;
2100 :
2101 44807 : newdesc = palloc_object(TriggerDesc);
2102 44807 : memcpy(newdesc, trigdesc, sizeof(TriggerDesc));
2103 :
2104 44807 : trigger = (Trigger *) palloc(trigdesc->numtriggers * sizeof(Trigger));
2105 44807 : memcpy(trigger, trigdesc->triggers,
2106 44807 : trigdesc->numtriggers * sizeof(Trigger));
2107 44807 : newdesc->triggers = trigger;
2108 :
2109 155177 : for (i = 0; i < trigdesc->numtriggers; i++)
2110 : {
2111 110370 : trigger->tgname = pstrdup(trigger->tgname);
2112 110370 : if (trigger->tgnattr > 0)
2113 : {
2114 : int16 *newattr;
2115 :
2116 681 : newattr = (int16 *) palloc(trigger->tgnattr * sizeof(int16));
2117 681 : memcpy(newattr, trigger->tgattr,
2118 681 : trigger->tgnattr * sizeof(int16));
2119 681 : trigger->tgattr = newattr;
2120 : }
2121 110370 : if (trigger->tgnargs > 0)
2122 : {
2123 : char **newargs;
2124 : int16 j;
2125 :
2126 7166 : newargs = (char **) palloc(trigger->tgnargs * sizeof(char *));
2127 15739 : for (j = 0; j < trigger->tgnargs; j++)
2128 8573 : newargs[j] = pstrdup(trigger->tgargs[j]);
2129 7166 : trigger->tgargs = newargs;
2130 : }
2131 110370 : if (trigger->tgqual)
2132 981 : trigger->tgqual = pstrdup(trigger->tgqual);
2133 110370 : if (trigger->tgoldtable)
2134 1610 : trigger->tgoldtable = pstrdup(trigger->tgoldtable);
2135 110370 : if (trigger->tgnewtable)
2136 1883 : trigger->tgnewtable = pstrdup(trigger->tgnewtable);
2137 110370 : trigger++;
2138 : }
2139 :
2140 44807 : return newdesc;
2141 : }
2142 :
2143 : /*
2144 : * Free a TriggerDesc data structure.
2145 : */
2146 : void
2147 891674 : FreeTriggerDesc(TriggerDesc *trigdesc)
2148 : {
2149 : Trigger *trigger;
2150 : int i;
2151 :
2152 891674 : if (trigdesc == NULL)
2153 829076 : return;
2154 :
2155 62598 : trigger = trigdesc->triggers;
2156 208382 : for (i = 0; i < trigdesc->numtriggers; i++)
2157 : {
2158 145784 : pfree(trigger->tgname);
2159 145784 : if (trigger->tgnattr > 0)
2160 647 : pfree(trigger->tgattr);
2161 145784 : if (trigger->tgnargs > 0)
2162 : {
2163 12350 : while (--(trigger->tgnargs) >= 0)
2164 6971 : pfree(trigger->tgargs[trigger->tgnargs]);
2165 5379 : pfree(trigger->tgargs);
2166 : }
2167 145784 : if (trigger->tgqual)
2168 1090 : pfree(trigger->tgqual);
2169 145784 : if (trigger->tgoldtable)
2170 1465 : pfree(trigger->tgoldtable);
2171 145784 : if (trigger->tgnewtable)
2172 1930 : pfree(trigger->tgnewtable);
2173 145784 : trigger++;
2174 : }
2175 62598 : pfree(trigdesc->triggers);
2176 62598 : pfree(trigdesc);
2177 : }
2178 :
2179 : /*
2180 : * Compare two TriggerDesc structures for logical equality.
2181 : */
2182 : #ifdef NOT_USED
2183 : bool
2184 : equalTriggerDescs(TriggerDesc *trigdesc1, TriggerDesc *trigdesc2)
2185 : {
2186 : int i,
2187 : j;
2188 :
2189 : /*
2190 : * We need not examine the hint flags, just the trigger array itself; if
2191 : * we have the same triggers with the same types, the flags should match.
2192 : *
2193 : * As of 7.3 we assume trigger set ordering is significant in the
2194 : * comparison; so we just compare corresponding slots of the two sets.
2195 : *
2196 : * Note: comparing the stringToNode forms of the WHEN clauses means that
2197 : * parse column locations will affect the result. This is okay as long as
2198 : * this function is only used for detecting exact equality, as for example
2199 : * in checking for staleness of a cache entry.
2200 : */
2201 : if (trigdesc1 != NULL)
2202 : {
2203 : if (trigdesc2 == NULL)
2204 : return false;
2205 : if (trigdesc1->numtriggers != trigdesc2->numtriggers)
2206 : return false;
2207 : for (i = 0; i < trigdesc1->numtriggers; i++)
2208 : {
2209 : Trigger *trig1 = trigdesc1->triggers + i;
2210 : Trigger *trig2 = trigdesc2->triggers + i;
2211 :
2212 : if (trig1->tgoid != trig2->tgoid)
2213 : return false;
2214 : if (strcmp(trig1->tgname, trig2->tgname) != 0)
2215 : return false;
2216 : if (trig1->tgfoid != trig2->tgfoid)
2217 : return false;
2218 : if (trig1->tgtype != trig2->tgtype)
2219 : return false;
2220 : if (trig1->tgenabled != trig2->tgenabled)
2221 : return false;
2222 : if (trig1->tgisinternal != trig2->tgisinternal)
2223 : return false;
2224 : if (trig1->tgisclone != trig2->tgisclone)
2225 : return false;
2226 : if (trig1->tgconstrrelid != trig2->tgconstrrelid)
2227 : return false;
2228 : if (trig1->tgconstrindid != trig2->tgconstrindid)
2229 : return false;
2230 : if (trig1->tgconstraint != trig2->tgconstraint)
2231 : return false;
2232 : if (trig1->tgdeferrable != trig2->tgdeferrable)
2233 : return false;
2234 : if (trig1->tginitdeferred != trig2->tginitdeferred)
2235 : return false;
2236 : if (trig1->tgnargs != trig2->tgnargs)
2237 : return false;
2238 : if (trig1->tgnattr != trig2->tgnattr)
2239 : return false;
2240 : if (trig1->tgnattr > 0 &&
2241 : memcmp(trig1->tgattr, trig2->tgattr,
2242 : trig1->tgnattr * sizeof(int16)) != 0)
2243 : return false;
2244 : for (j = 0; j < trig1->tgnargs; j++)
2245 : if (strcmp(trig1->tgargs[j], trig2->tgargs[j]) != 0)
2246 : return false;
2247 : if (trig1->tgqual == NULL && trig2->tgqual == NULL)
2248 : /* ok */ ;
2249 : else if (trig1->tgqual == NULL || trig2->tgqual == NULL)
2250 : return false;
2251 : else if (strcmp(trig1->tgqual, trig2->tgqual) != 0)
2252 : return false;
2253 : if (trig1->tgoldtable == NULL && trig2->tgoldtable == NULL)
2254 : /* ok */ ;
2255 : else if (trig1->tgoldtable == NULL || trig2->tgoldtable == NULL)
2256 : return false;
2257 : else if (strcmp(trig1->tgoldtable, trig2->tgoldtable) != 0)
2258 : return false;
2259 : if (trig1->tgnewtable == NULL && trig2->tgnewtable == NULL)
2260 : /* ok */ ;
2261 : else if (trig1->tgnewtable == NULL || trig2->tgnewtable == NULL)
2262 : return false;
2263 : else if (strcmp(trig1->tgnewtable, trig2->tgnewtable) != 0)
2264 : return false;
2265 : }
2266 : }
2267 : else if (trigdesc2 != NULL)
2268 : return false;
2269 : return true;
2270 : }
2271 : #endif /* NOT_USED */
2272 :
2273 : /*
2274 : * Check if there is a row-level trigger with transition tables that prevents
2275 : * a table from becoming an inheritance child or partition. Return the name
2276 : * of the first such incompatible trigger, or NULL if there is none.
2277 : */
2278 : const char *
2279 1797 : FindTriggerIncompatibleWithInheritance(TriggerDesc *trigdesc)
2280 : {
2281 1797 : if (trigdesc != NULL)
2282 : {
2283 : int i;
2284 :
2285 408 : for (i = 0; i < trigdesc->numtriggers; ++i)
2286 : {
2287 292 : Trigger *trigger = &trigdesc->triggers[i];
2288 :
2289 292 : if (!TRIGGER_FOR_ROW(trigger->tgtype))
2290 24 : continue;
2291 268 : if (trigger->tgoldtable != NULL || trigger->tgnewtable != NULL)
2292 8 : return trigger->tgname;
2293 : }
2294 : }
2295 :
2296 1789 : return NULL;
2297 : }
2298 :
2299 : /*
2300 : * Call a trigger function.
2301 : *
2302 : * trigdata: trigger descriptor.
2303 : * tgindx: trigger's index in finfo and instr arrays.
2304 : * finfo: array of cached trigger function call information.
2305 : * instr: optional array of EXPLAIN ANALYZE instrumentation state.
2306 : * per_tuple_context: memory context to execute the function in.
2307 : *
2308 : * Returns the tuple (or NULL) as returned by the function.
2309 : */
2310 : static HeapTuple
2311 616405 : ExecCallTriggerFunc(TriggerData *trigdata,
2312 : int tgindx,
2313 : FmgrInfo *finfo,
2314 : TriggerInstrumentation *instr,
2315 : MemoryContext per_tuple_context)
2316 : {
2317 616405 : LOCAL_FCINFO(fcinfo, 0);
2318 : PgStat_FunctionCallUsage fcusage;
2319 : Datum result;
2320 : MemoryContext oldContext;
2321 :
2322 : /*
2323 : * Protect against code paths that may fail to initialize transition table
2324 : * info.
2325 : */
2326 : Assert(((TRIGGER_FIRED_BY_INSERT(trigdata->tg_event) ||
2327 : TRIGGER_FIRED_BY_UPDATE(trigdata->tg_event) ||
2328 : TRIGGER_FIRED_BY_DELETE(trigdata->tg_event)) &&
2329 : TRIGGER_FIRED_AFTER(trigdata->tg_event) &&
2330 : !(trigdata->tg_event & AFTER_TRIGGER_DEFERRABLE) &&
2331 : !(trigdata->tg_event & AFTER_TRIGGER_INITDEFERRED)) ||
2332 : (trigdata->tg_oldtable == NULL && trigdata->tg_newtable == NULL));
2333 :
2334 616405 : finfo += tgindx;
2335 :
2336 : /*
2337 : * We cache fmgr lookup info, to avoid making the lookup again on each
2338 : * call.
2339 : */
2340 616405 : if (finfo->fn_oid == InvalidOid)
2341 12575 : fmgr_info(trigdata->tg_trigger->tgfoid, finfo);
2342 :
2343 : Assert(finfo->fn_oid == trigdata->tg_trigger->tgfoid);
2344 :
2345 : /*
2346 : * If doing EXPLAIN ANALYZE, start charging time to this trigger.
2347 : */
2348 616405 : if (instr)
2349 0 : InstrStartTrigger(instr + tgindx);
2350 :
2351 : /*
2352 : * Do the function evaluation in the per-tuple memory context, so that
2353 : * leaked memory will be reclaimed once per tuple. Note in particular that
2354 : * any new tuple created by the trigger function will live till the end of
2355 : * the tuple cycle.
2356 : */
2357 616405 : oldContext = MemoryContextSwitchTo(per_tuple_context);
2358 :
2359 : /*
2360 : * Call the function, passing no arguments but setting a context.
2361 : */
2362 616405 : InitFunctionCallInfoData(*fcinfo, finfo, 0,
2363 : InvalidOid, (Node *) trigdata, NULL);
2364 :
2365 616405 : pgstat_init_function_usage(fcinfo, &fcusage);
2366 :
2367 616405 : MyTriggerDepth++;
2368 616405 : PG_TRY();
2369 : {
2370 616405 : result = FunctionCallInvoke(fcinfo);
2371 : }
2372 690 : PG_FINALLY();
2373 : {
2374 616405 : MyTriggerDepth--;
2375 : }
2376 616405 : PG_END_TRY();
2377 :
2378 615715 : pgstat_end_function_usage(&fcusage, true);
2379 :
2380 615715 : MemoryContextSwitchTo(oldContext);
2381 :
2382 : /*
2383 : * Trigger protocol allows function to return a null pointer, but NOT to
2384 : * set the isnull result flag.
2385 : */
2386 615715 : if (fcinfo->isnull)
2387 0 : ereport(ERROR,
2388 : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2389 : errmsg("trigger function %u returned null value",
2390 : fcinfo->flinfo->fn_oid)));
2391 :
2392 : /*
2393 : * If doing EXPLAIN ANALYZE, stop charging time to this trigger, and count
2394 : * the firing of the trigger.
2395 : */
2396 615715 : if (instr)
2397 0 : InstrStopTrigger(instr + tgindx, 1);
2398 :
2399 615715 : return (HeapTuple) DatumGetPointer(result);
2400 : }
2401 :
2402 : void
2403 56086 : ExecBSInsertTriggers(EState *estate, ResultRelInfo *relinfo)
2404 : {
2405 : TriggerDesc *trigdesc;
2406 : int i;
2407 56086 : TriggerData LocTriggerData = {0};
2408 :
2409 56086 : trigdesc = relinfo->ri_TrigDesc;
2410 :
2411 56086 : if (trigdesc == NULL)
2412 55915 : return;
2413 5032 : if (!trigdesc->trig_insert_before_statement)
2414 4861 : return;
2415 :
2416 : /* no-op if we already fired BS triggers in this context */
2417 171 : if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2418 : CMD_INSERT))
2419 0 : return;
2420 :
2421 171 : LocTriggerData.type = T_TriggerData;
2422 171 : LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2423 : TRIGGER_EVENT_BEFORE;
2424 171 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2425 1501 : for (i = 0; i < trigdesc->numtriggers; i++)
2426 : {
2427 1338 : Trigger *trigger = &trigdesc->triggers[i];
2428 : HeapTuple newtuple;
2429 :
2430 1338 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2431 : TRIGGER_TYPE_STATEMENT,
2432 : TRIGGER_TYPE_BEFORE,
2433 : TRIGGER_TYPE_INSERT))
2434 1159 : continue;
2435 179 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2436 : NULL, NULL, NULL))
2437 20 : continue;
2438 :
2439 159 : LocTriggerData.tg_trigger = trigger;
2440 159 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2441 : i,
2442 : relinfo->ri_TrigFunctions,
2443 : relinfo->ri_TrigInstrument,
2444 159 : GetPerTupleMemoryContext(estate));
2445 :
2446 151 : if (newtuple)
2447 0 : ereport(ERROR,
2448 : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2449 : errmsg("BEFORE STATEMENT trigger cannot return a value")));
2450 : }
2451 : }
2452 :
2453 : void
2454 54296 : ExecASInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2455 : TransitionCaptureState *transition_capture)
2456 : {
2457 54296 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2458 :
2459 54296 : if (trigdesc && trigdesc->trig_insert_after_statement)
2460 349 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2461 : TRIGGER_EVENT_INSERT,
2462 : false, NULL, NULL, NIL, NULL, transition_capture,
2463 : false);
2464 54296 : }
2465 :
2466 : bool
2467 1610 : ExecBRInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2468 : TupleTableSlot *slot)
2469 : {
2470 1610 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2471 1610 : HeapTuple newtuple = NULL;
2472 : bool should_free;
2473 1610 : TriggerData LocTriggerData = {0};
2474 : int i;
2475 :
2476 1610 : LocTriggerData.type = T_TriggerData;
2477 1610 : LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2478 : TRIGGER_EVENT_ROW |
2479 : TRIGGER_EVENT_BEFORE;
2480 1610 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2481 7566 : for (i = 0; i < trigdesc->numtriggers; i++)
2482 : {
2483 6160 : Trigger *trigger = &trigdesc->triggers[i];
2484 : HeapTuple oldtuple;
2485 :
2486 6160 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2487 : TRIGGER_TYPE_ROW,
2488 : TRIGGER_TYPE_BEFORE,
2489 : TRIGGER_TYPE_INSERT))
2490 3024 : continue;
2491 3136 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2492 : NULL, NULL, slot))
2493 41 : continue;
2494 :
2495 3095 : if (!newtuple)
2496 1587 : newtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2497 :
2498 3095 : LocTriggerData.tg_trigslot = slot;
2499 3095 : LocTriggerData.tg_trigtuple = oldtuple = newtuple;
2500 3095 : LocTriggerData.tg_trigger = trigger;
2501 3095 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2502 : i,
2503 : relinfo->ri_TrigFunctions,
2504 : relinfo->ri_TrigInstrument,
2505 3095 : GetPerTupleMemoryContext(estate));
2506 3049 : if (newtuple == NULL)
2507 : {
2508 142 : if (should_free)
2509 13 : heap_freetuple(oldtuple);
2510 142 : return false; /* "do nothing" */
2511 : }
2512 2907 : else if (newtuple != oldtuple)
2513 : {
2514 532 : newtuple = check_modified_virtual_generated(RelationGetDescr(relinfo->ri_RelationDesc), newtuple);
2515 :
2516 532 : ExecForceStoreHeapTuple(newtuple, slot, false);
2517 :
2518 : /*
2519 : * After a tuple in a partition goes through a trigger, the user
2520 : * could have changed the partition key enough that the tuple no
2521 : * longer fits the partition. Verify that.
2522 : */
2523 532 : if (trigger->tgisclone &&
2524 44 : !ExecPartitionCheck(relinfo, slot, estate, false))
2525 16 : ereport(ERROR,
2526 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2527 : errmsg("moving row to another partition during a BEFORE FOR EACH ROW trigger is not supported"),
2528 : errdetail("Before executing trigger \"%s\", the row was to be in partition \"%s.%s\".",
2529 : trigger->tgname,
2530 : get_namespace_name(RelationGetNamespace(relinfo->ri_RelationDesc)),
2531 : RelationGetRelationName(relinfo->ri_RelationDesc))));
2532 :
2533 516 : if (should_free)
2534 26 : heap_freetuple(oldtuple);
2535 :
2536 : /* signal tuple should be re-fetched if used */
2537 516 : newtuple = NULL;
2538 : }
2539 : }
2540 :
2541 1406 : return true;
2542 : }
2543 :
2544 : void
2545 7867391 : ExecARInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2546 : TupleTableSlot *slot, List *recheckIndexes,
2547 : TransitionCaptureState *transition_capture)
2548 : {
2549 7867391 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2550 :
2551 7867391 : if (relinfo->ri_FdwRoutine && transition_capture &&
2552 4 : transition_capture->tcs_insert_new_table)
2553 : {
2554 : Assert(relinfo->ri_RootResultRelInfo);
2555 4 : ereport(ERROR,
2556 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2557 : errmsg("cannot collect transition tuples from child foreign tables")));
2558 : }
2559 :
2560 7867387 : if ((trigdesc && trigdesc->trig_insert_after_row) ||
2561 40224 : (transition_capture && transition_capture->tcs_insert_new_table))
2562 445528 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2563 : TRIGGER_EVENT_INSERT,
2564 : true, NULL, slot,
2565 : recheckIndexes, NULL,
2566 : transition_capture,
2567 : false);
2568 7867387 : }
2569 :
2570 : bool
2571 119 : ExecIRInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2572 : TupleTableSlot *slot)
2573 : {
2574 119 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2575 119 : HeapTuple newtuple = NULL;
2576 : bool should_free;
2577 119 : TriggerData LocTriggerData = {0};
2578 : int i;
2579 :
2580 119 : LocTriggerData.type = T_TriggerData;
2581 119 : LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2582 : TRIGGER_EVENT_ROW |
2583 : TRIGGER_EVENT_INSTEAD;
2584 119 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2585 362 : for (i = 0; i < trigdesc->numtriggers; i++)
2586 : {
2587 255 : Trigger *trigger = &trigdesc->triggers[i];
2588 : HeapTuple oldtuple;
2589 :
2590 255 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2591 : TRIGGER_TYPE_ROW,
2592 : TRIGGER_TYPE_INSTEAD,
2593 : TRIGGER_TYPE_INSERT))
2594 136 : continue;
2595 119 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2596 : NULL, NULL, slot))
2597 0 : continue;
2598 :
2599 119 : if (!newtuple)
2600 119 : newtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2601 :
2602 119 : LocTriggerData.tg_trigslot = slot;
2603 119 : LocTriggerData.tg_trigtuple = oldtuple = newtuple;
2604 119 : LocTriggerData.tg_trigger = trigger;
2605 119 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2606 : i,
2607 : relinfo->ri_TrigFunctions,
2608 : relinfo->ri_TrigInstrument,
2609 119 : GetPerTupleMemoryContext(estate));
2610 119 : if (newtuple == NULL)
2611 : {
2612 12 : if (should_free)
2613 12 : heap_freetuple(oldtuple);
2614 12 : return false; /* "do nothing" */
2615 : }
2616 107 : else if (newtuple != oldtuple)
2617 : {
2618 36 : ExecForceStoreHeapTuple(newtuple, slot, false);
2619 :
2620 36 : if (should_free)
2621 36 : heap_freetuple(oldtuple);
2622 :
2623 : /* signal tuple should be re-fetched if used */
2624 36 : newtuple = NULL;
2625 : }
2626 : }
2627 :
2628 107 : return true;
2629 : }
2630 :
2631 : void
2632 8557 : ExecBSDeleteTriggers(EState *estate, ResultRelInfo *relinfo)
2633 : {
2634 : TriggerDesc *trigdesc;
2635 : int i;
2636 8557 : TriggerData LocTriggerData = {0};
2637 :
2638 8557 : trigdesc = relinfo->ri_TrigDesc;
2639 :
2640 8557 : if (trigdesc == NULL)
2641 8498 : return;
2642 1066 : if (!trigdesc->trig_delete_before_statement)
2643 979 : return;
2644 :
2645 : /* no-op if we already fired BS triggers in this context */
2646 87 : if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2647 : CMD_DELETE))
2648 28 : return;
2649 :
2650 59 : LocTriggerData.type = T_TriggerData;
2651 59 : LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2652 : TRIGGER_EVENT_BEFORE;
2653 59 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2654 540 : for (i = 0; i < trigdesc->numtriggers; i++)
2655 : {
2656 481 : Trigger *trigger = &trigdesc->triggers[i];
2657 : HeapTuple newtuple;
2658 :
2659 481 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2660 : TRIGGER_TYPE_STATEMENT,
2661 : TRIGGER_TYPE_BEFORE,
2662 : TRIGGER_TYPE_DELETE))
2663 422 : continue;
2664 59 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2665 : NULL, NULL, NULL))
2666 8 : continue;
2667 :
2668 51 : LocTriggerData.tg_trigger = trigger;
2669 51 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2670 : i,
2671 : relinfo->ri_TrigFunctions,
2672 : relinfo->ri_TrigInstrument,
2673 51 : GetPerTupleMemoryContext(estate));
2674 :
2675 51 : if (newtuple)
2676 0 : ereport(ERROR,
2677 : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2678 : errmsg("BEFORE STATEMENT trigger cannot return a value")));
2679 : }
2680 : }
2681 :
2682 : void
2683 8434 : ExecASDeleteTriggers(EState *estate, ResultRelInfo *relinfo,
2684 : TransitionCaptureState *transition_capture)
2685 : {
2686 8434 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2687 :
2688 8434 : if (trigdesc && trigdesc->trig_delete_after_statement)
2689 165 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2690 : TRIGGER_EVENT_DELETE,
2691 : false, NULL, NULL, NIL, NULL, transition_capture,
2692 : false);
2693 8434 : }
2694 :
2695 : /*
2696 : * Execute BEFORE ROW DELETE triggers.
2697 : *
2698 : * True indicates caller can proceed with the delete. False indicates caller
2699 : * need to suppress the delete and additionally if requested, we need to pass
2700 : * back the concurrently updated tuple if any.
2701 : */
2702 : bool
2703 217 : ExecBRDeleteTriggers(EState *estate, EPQState *epqstate,
2704 : ResultRelInfo *relinfo,
2705 : ItemPointer tupleid,
2706 : HeapTuple fdw_trigtuple,
2707 : TupleTableSlot **epqslot,
2708 : TM_Result *tmresult,
2709 : TM_FailureData *tmfd,
2710 : bool is_merge_delete)
2711 : {
2712 217 : TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2713 217 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2714 217 : bool result = true;
2715 217 : TriggerData LocTriggerData = {0};
2716 : HeapTuple trigtuple;
2717 217 : bool should_free = false;
2718 : int i;
2719 :
2720 : Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2721 217 : if (fdw_trigtuple == NULL)
2722 : {
2723 209 : TupleTableSlot *epqslot_candidate = NULL;
2724 :
2725 : /*
2726 : * Get a copy of the on-disk tuple we are planning to delete. In
2727 : * general, if the tuple has been concurrently updated, we should
2728 : * recheck it using EPQ. However, if this is a MERGE DELETE action,
2729 : * we skip this EPQ recheck and leave it to the caller (it must do
2730 : * additional rechecking, and might end up executing a different
2731 : * action entirely).
2732 : */
2733 205 : if (!GetTupleForTrigger(estate, epqstate, relinfo, tupleid,
2734 : LockTupleExclusive, slot, !is_merge_delete,
2735 209 : &epqslot_candidate, tmresult, tmfd))
2736 6 : return false;
2737 :
2738 : /*
2739 : * If the tuple was concurrently updated and the caller of this
2740 : * function requested for the updated tuple, skip the trigger
2741 : * execution.
2742 : */
2743 200 : if (epqslot_candidate != NULL && epqslot != NULL)
2744 : {
2745 1 : *epqslot = epqslot_candidate;
2746 1 : return false;
2747 : }
2748 :
2749 199 : trigtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2750 : }
2751 : else
2752 : {
2753 8 : trigtuple = fdw_trigtuple;
2754 8 : ExecForceStoreHeapTuple(trigtuple, slot, false);
2755 : }
2756 :
2757 207 : LocTriggerData.type = T_TriggerData;
2758 207 : LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2759 : TRIGGER_EVENT_ROW |
2760 : TRIGGER_EVENT_BEFORE;
2761 207 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2762 832 : for (i = 0; i < trigdesc->numtriggers; i++)
2763 : {
2764 : HeapTuple newtuple;
2765 665 : Trigger *trigger = &trigdesc->triggers[i];
2766 :
2767 665 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2768 : TRIGGER_TYPE_ROW,
2769 : TRIGGER_TYPE_BEFORE,
2770 : TRIGGER_TYPE_DELETE))
2771 454 : continue;
2772 211 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2773 : NULL, slot, NULL))
2774 9 : continue;
2775 :
2776 202 : LocTriggerData.tg_trigslot = slot;
2777 202 : LocTriggerData.tg_trigtuple = trigtuple;
2778 202 : LocTriggerData.tg_trigger = trigger;
2779 202 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2780 : i,
2781 : relinfo->ri_TrigFunctions,
2782 : relinfo->ri_TrigInstrument,
2783 202 : GetPerTupleMemoryContext(estate));
2784 196 : if (newtuple == NULL)
2785 : {
2786 34 : result = false; /* tell caller to suppress delete */
2787 34 : break;
2788 : }
2789 162 : if (newtuple != trigtuple)
2790 33 : heap_freetuple(newtuple);
2791 : }
2792 201 : if (should_free)
2793 0 : heap_freetuple(trigtuple);
2794 :
2795 201 : return result;
2796 : }
2797 :
2798 : /*
2799 : * Note: is_crosspart_update must be true if the DELETE is being performed
2800 : * as part of a cross-partition update.
2801 : */
2802 : void
2803 1061757 : ExecARDeleteTriggers(EState *estate,
2804 : ResultRelInfo *relinfo,
2805 : ItemPointer tupleid,
2806 : HeapTuple fdw_trigtuple,
2807 : TransitionCaptureState *transition_capture,
2808 : bool is_crosspart_update)
2809 : {
2810 1061757 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2811 :
2812 1061757 : if (relinfo->ri_FdwRoutine && transition_capture &&
2813 2 : transition_capture->tcs_delete_old_table)
2814 : {
2815 : Assert(relinfo->ri_RootResultRelInfo);
2816 2 : ereport(ERROR,
2817 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2818 : errmsg("cannot collect transition tuples from child foreign tables")));
2819 : }
2820 :
2821 1061755 : if ((trigdesc && trigdesc->trig_delete_after_row) ||
2822 3344 : (transition_capture && transition_capture->tcs_delete_old_table))
2823 : {
2824 4150 : TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2825 :
2826 : Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2827 4150 : if (fdw_trigtuple == NULL)
2828 4142 : GetTupleForTrigger(estate,
2829 : NULL,
2830 : relinfo,
2831 : tupleid,
2832 : LockTupleExclusive,
2833 : slot,
2834 : false,
2835 : NULL,
2836 : NULL,
2837 : NULL);
2838 : else
2839 8 : ExecForceStoreHeapTuple(fdw_trigtuple, slot, false);
2840 :
2841 4150 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2842 : TRIGGER_EVENT_DELETE,
2843 : true, slot, NULL, NIL, NULL,
2844 : transition_capture,
2845 : is_crosspart_update);
2846 : }
2847 1061755 : }
2848 :
2849 : bool
2850 39 : ExecIRDeleteTriggers(EState *estate, ResultRelInfo *relinfo,
2851 : HeapTuple trigtuple)
2852 : {
2853 39 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2854 39 : TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2855 39 : TriggerData LocTriggerData = {0};
2856 : int i;
2857 :
2858 39 : LocTriggerData.type = T_TriggerData;
2859 39 : LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2860 : TRIGGER_EVENT_ROW |
2861 : TRIGGER_EVENT_INSTEAD;
2862 39 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2863 :
2864 39 : ExecForceStoreHeapTuple(trigtuple, slot, false);
2865 :
2866 234 : for (i = 0; i < trigdesc->numtriggers; i++)
2867 : {
2868 : HeapTuple rettuple;
2869 199 : Trigger *trigger = &trigdesc->triggers[i];
2870 :
2871 199 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2872 : TRIGGER_TYPE_ROW,
2873 : TRIGGER_TYPE_INSTEAD,
2874 : TRIGGER_TYPE_DELETE))
2875 160 : continue;
2876 39 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2877 : NULL, slot, NULL))
2878 0 : continue;
2879 :
2880 39 : LocTriggerData.tg_trigslot = slot;
2881 39 : LocTriggerData.tg_trigtuple = trigtuple;
2882 39 : LocTriggerData.tg_trigger = trigger;
2883 39 : rettuple = ExecCallTriggerFunc(&LocTriggerData,
2884 : i,
2885 : relinfo->ri_TrigFunctions,
2886 : relinfo->ri_TrigInstrument,
2887 39 : GetPerTupleMemoryContext(estate));
2888 39 : if (rettuple == NULL)
2889 4 : return false; /* Delete was suppressed */
2890 35 : if (rettuple != trigtuple)
2891 0 : heap_freetuple(rettuple);
2892 : }
2893 35 : return true;
2894 : }
2895 :
2896 : void
2897 10147 : ExecBSUpdateTriggers(EState *estate, ResultRelInfo *relinfo)
2898 : {
2899 : TriggerDesc *trigdesc;
2900 : int i;
2901 10147 : TriggerData LocTriggerData = {0};
2902 : Bitmapset *updatedCols;
2903 :
2904 10147 : trigdesc = relinfo->ri_TrigDesc;
2905 :
2906 10147 : if (trigdesc == NULL)
2907 10018 : return;
2908 2743 : if (!trigdesc->trig_update_before_statement)
2909 2614 : return;
2910 :
2911 : /* no-op if we already fired BS triggers in this context */
2912 129 : if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2913 : CMD_UPDATE))
2914 0 : return;
2915 :
2916 : /* statement-level triggers operate on the parent table */
2917 : Assert(relinfo->ri_RootResultRelInfo == NULL);
2918 :
2919 129 : updatedCols = ExecGetAllUpdatedCols(relinfo, estate);
2920 :
2921 129 : LocTriggerData.type = T_TriggerData;
2922 129 : LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
2923 : TRIGGER_EVENT_BEFORE;
2924 129 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2925 129 : LocTriggerData.tg_updatedcols = updatedCols;
2926 1168 : for (i = 0; i < trigdesc->numtriggers; i++)
2927 : {
2928 1039 : Trigger *trigger = &trigdesc->triggers[i];
2929 : HeapTuple newtuple;
2930 :
2931 1039 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2932 : TRIGGER_TYPE_STATEMENT,
2933 : TRIGGER_TYPE_BEFORE,
2934 : TRIGGER_TYPE_UPDATE))
2935 910 : continue;
2936 129 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2937 : updatedCols, NULL, NULL))
2938 4 : continue;
2939 :
2940 125 : LocTriggerData.tg_trigger = trigger;
2941 125 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2942 : i,
2943 : relinfo->ri_TrigFunctions,
2944 : relinfo->ri_TrigInstrument,
2945 125 : GetPerTupleMemoryContext(estate));
2946 :
2947 125 : if (newtuple)
2948 0 : ereport(ERROR,
2949 : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2950 : errmsg("BEFORE STATEMENT trigger cannot return a value")));
2951 : }
2952 : }
2953 :
2954 : void
2955 9505 : ExecASUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
2956 : TransitionCaptureState *transition_capture)
2957 : {
2958 9505 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2959 :
2960 : /* statement-level triggers operate on the parent table */
2961 : Assert(relinfo->ri_RootResultRelInfo == NULL);
2962 :
2963 9505 : if (trigdesc && trigdesc->trig_update_after_statement)
2964 277 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2965 : TRIGGER_EVENT_UPDATE,
2966 : false, NULL, NULL, NIL,
2967 : ExecGetAllUpdatedCols(relinfo, estate),
2968 : transition_capture,
2969 : false);
2970 9505 : }
2971 :
2972 : bool
2973 1572 : ExecBRUpdateTriggers(EState *estate, EPQState *epqstate,
2974 : ResultRelInfo *relinfo,
2975 : ItemPointer tupleid,
2976 : HeapTuple fdw_trigtuple,
2977 : TupleTableSlot *newslot,
2978 : TM_Result *tmresult,
2979 : TM_FailureData *tmfd,
2980 : bool is_merge_update)
2981 : {
2982 1572 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2983 1572 : TupleTableSlot *oldslot = ExecGetTriggerOldSlot(estate, relinfo);
2984 1572 : HeapTuple newtuple = NULL;
2985 : HeapTuple trigtuple;
2986 1572 : bool should_free_trig = false;
2987 1572 : bool should_free_new = false;
2988 1572 : TriggerData LocTriggerData = {0};
2989 : int i;
2990 : Bitmapset *updatedCols;
2991 : LockTupleMode lockmode;
2992 :
2993 : /* Determine lock mode to use */
2994 1572 : lockmode = ExecUpdateLockMode(estate, relinfo);
2995 :
2996 : Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2997 1572 : if (fdw_trigtuple == NULL)
2998 : {
2999 1553 : TupleTableSlot *epqslot_candidate = NULL;
3000 :
3001 : /*
3002 : * Get a copy of the on-disk tuple we are planning to update. In
3003 : * general, if the tuple has been concurrently updated, we should
3004 : * recheck it using EPQ. However, if this is a MERGE UPDATE action,
3005 : * we skip this EPQ recheck and leave it to the caller (it must do
3006 : * additional rechecking, and might end up executing a different
3007 : * action entirely).
3008 : */
3009 1549 : if (!GetTupleForTrigger(estate, epqstate, relinfo, tupleid,
3010 : lockmode, oldslot, !is_merge_update,
3011 1553 : &epqslot_candidate, tmresult, tmfd))
3012 12 : return false; /* cancel the update action */
3013 :
3014 : /*
3015 : * In READ COMMITTED isolation level it's possible that target tuple
3016 : * was changed due to concurrent update. In that case we have a raw
3017 : * subplan output tuple in epqslot_candidate, and need to form a new
3018 : * insertable tuple using ExecGetUpdateNewTuple to replace the one we
3019 : * received in newslot. Neither we nor our callers have any further
3020 : * interest in the passed-in tuple, so it's okay to overwrite newslot
3021 : * with the newer data.
3022 : */
3023 1537 : if (epqslot_candidate != NULL)
3024 : {
3025 : TupleTableSlot *epqslot_clean;
3026 :
3027 3 : epqslot_clean = ExecGetUpdateNewTuple(relinfo, epqslot_candidate,
3028 : oldslot);
3029 :
3030 : /*
3031 : * Typically, the caller's newslot was also generated by
3032 : * ExecGetUpdateNewTuple, so that epqslot_clean will be the same
3033 : * slot and copying is not needed. But do the right thing if it
3034 : * isn't.
3035 : */
3036 3 : if (unlikely(newslot != epqslot_clean))
3037 0 : ExecCopySlot(newslot, epqslot_clean);
3038 :
3039 : /*
3040 : * At this point newslot contains a virtual tuple that may
3041 : * reference some fields of oldslot's tuple in some disk buffer.
3042 : * If that tuple is in a different page than the original target
3043 : * tuple, then our only pin on that buffer is oldslot's, and we're
3044 : * about to release it. Hence we'd better materialize newslot to
3045 : * ensure it doesn't contain references into an unpinned buffer.
3046 : * (We'd materialize it below anyway, but too late for safety.)
3047 : */
3048 3 : ExecMaterializeSlot(newslot);
3049 : }
3050 :
3051 : /*
3052 : * Here we convert oldslot to a materialized slot holding trigtuple.
3053 : * Neither slot passed to the triggers will hold any buffer pin.
3054 : */
3055 1537 : trigtuple = ExecFetchSlotHeapTuple(oldslot, true, &should_free_trig);
3056 : }
3057 : else
3058 : {
3059 : /* Put the FDW-supplied tuple into oldslot to unify the cases */
3060 19 : ExecForceStoreHeapTuple(fdw_trigtuple, oldslot, false);
3061 19 : trigtuple = fdw_trigtuple;
3062 : }
3063 :
3064 1556 : LocTriggerData.type = T_TriggerData;
3065 1556 : LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
3066 : TRIGGER_EVENT_ROW |
3067 : TRIGGER_EVENT_BEFORE;
3068 1556 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3069 1556 : updatedCols = ExecGetAllUpdatedCols(relinfo, estate);
3070 1556 : LocTriggerData.tg_updatedcols = updatedCols;
3071 7408 : for (i = 0; i < trigdesc->numtriggers; i++)
3072 : {
3073 5946 : Trigger *trigger = &trigdesc->triggers[i];
3074 : HeapTuple oldtuple;
3075 :
3076 5946 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3077 : TRIGGER_TYPE_ROW,
3078 : TRIGGER_TYPE_BEFORE,
3079 : TRIGGER_TYPE_UPDATE))
3080 3018 : continue;
3081 2928 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3082 : updatedCols, oldslot, newslot))
3083 65 : continue;
3084 :
3085 2863 : if (!newtuple)
3086 1547 : newtuple = ExecFetchSlotHeapTuple(newslot, true, &should_free_new);
3087 :
3088 2863 : LocTriggerData.tg_trigslot = oldslot;
3089 2863 : LocTriggerData.tg_trigtuple = trigtuple;
3090 2863 : LocTriggerData.tg_newtuple = oldtuple = newtuple;
3091 2863 : LocTriggerData.tg_newslot = newslot;
3092 2863 : LocTriggerData.tg_trigger = trigger;
3093 2863 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
3094 : i,
3095 : relinfo->ri_TrigFunctions,
3096 : relinfo->ri_TrigInstrument,
3097 2863 : GetPerTupleMemoryContext(estate));
3098 :
3099 2855 : if (newtuple == NULL)
3100 : {
3101 86 : if (should_free_trig)
3102 0 : heap_freetuple(trigtuple);
3103 86 : if (should_free_new)
3104 2 : heap_freetuple(oldtuple);
3105 86 : return false; /* "do nothing" */
3106 : }
3107 2769 : else if (newtuple != oldtuple)
3108 : {
3109 736 : newtuple = check_modified_virtual_generated(RelationGetDescr(relinfo->ri_RelationDesc), newtuple);
3110 :
3111 736 : ExecForceStoreHeapTuple(newtuple, newslot, false);
3112 :
3113 : /*
3114 : * If the tuple returned by the trigger / being stored, is the old
3115 : * row version, and the heap tuple passed to the trigger was
3116 : * allocated locally, materialize the slot. Otherwise we might
3117 : * free it while still referenced by the slot.
3118 : */
3119 736 : if (should_free_trig && newtuple == trigtuple)
3120 0 : ExecMaterializeSlot(newslot);
3121 :
3122 736 : if (should_free_new)
3123 1 : heap_freetuple(oldtuple);
3124 :
3125 : /* signal tuple should be re-fetched if used */
3126 736 : newtuple = NULL;
3127 : }
3128 : }
3129 1462 : if (should_free_trig)
3130 0 : heap_freetuple(trigtuple);
3131 :
3132 1462 : return true;
3133 : }
3134 :
3135 : /*
3136 : * Note: 'src_partinfo' and 'dst_partinfo', when non-NULL, refer to the source
3137 : * and destination partitions, respectively, of a cross-partition update of
3138 : * the root partitioned table mentioned in the query, given by 'relinfo'.
3139 : * 'tupleid' in that case refers to the ctid of the "old" tuple in the source
3140 : * partition, and 'newslot' contains the "new" tuple in the destination
3141 : * partition. This interface allows to support the requirements of
3142 : * ExecCrossPartitionUpdateForeignKey(); is_crosspart_update must be true in
3143 : * that case.
3144 : */
3145 : void
3146 2245485 : ExecARUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
3147 : ResultRelInfo *src_partinfo,
3148 : ResultRelInfo *dst_partinfo,
3149 : ItemPointer tupleid,
3150 : HeapTuple fdw_trigtuple,
3151 : TupleTableSlot *newslot,
3152 : List *recheckIndexes,
3153 : TransitionCaptureState *transition_capture,
3154 : bool is_crosspart_update)
3155 : {
3156 2245485 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3157 :
3158 2245485 : if (relinfo->ri_FdwRoutine && transition_capture &&
3159 2 : (transition_capture->tcs_update_old_table ||
3160 0 : transition_capture->tcs_update_new_table))
3161 : {
3162 : Assert(relinfo->ri_RootResultRelInfo);
3163 2 : ereport(ERROR,
3164 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
3165 : errmsg("cannot collect transition tuples from child foreign tables")));
3166 : }
3167 :
3168 2245483 : if ((trigdesc && trigdesc->trig_update_after_row) ||
3169 247 : (transition_capture &&
3170 247 : (transition_capture->tcs_update_old_table ||
3171 12 : transition_capture->tcs_update_new_table)))
3172 : {
3173 : /*
3174 : * Note: if the UPDATE is converted into a DELETE+INSERT as part of
3175 : * update-partition-key operation, then this function is also called
3176 : * separately for DELETE and INSERT to capture transition table rows.
3177 : * In such case, either old tuple or new tuple can be NULL.
3178 : */
3179 : TupleTableSlot *oldslot;
3180 : ResultRelInfo *tupsrc;
3181 :
3182 : Assert((src_partinfo != NULL && dst_partinfo != NULL) ||
3183 : !is_crosspart_update);
3184 :
3185 2444 : tupsrc = src_partinfo ? src_partinfo : relinfo;
3186 2444 : oldslot = ExecGetTriggerOldSlot(estate, tupsrc);
3187 :
3188 2444 : if (fdw_trigtuple == NULL && ItemPointerIsValid(tupleid))
3189 2402 : GetTupleForTrigger(estate,
3190 : NULL,
3191 : tupsrc,
3192 : tupleid,
3193 : LockTupleExclusive,
3194 : oldslot,
3195 : false,
3196 : NULL,
3197 : NULL,
3198 : NULL);
3199 42 : else if (fdw_trigtuple != NULL)
3200 10 : ExecForceStoreHeapTuple(fdw_trigtuple, oldslot, false);
3201 : else
3202 32 : ExecClearTuple(oldslot);
3203 :
3204 2444 : AfterTriggerSaveEvent(estate, relinfo,
3205 : src_partinfo, dst_partinfo,
3206 : TRIGGER_EVENT_UPDATE,
3207 : true,
3208 : oldslot, newslot, recheckIndexes,
3209 : ExecGetAllUpdatedCols(relinfo, estate),
3210 : transition_capture,
3211 : is_crosspart_update);
3212 : }
3213 2245483 : }
3214 :
3215 : bool
3216 135 : ExecIRUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
3217 : HeapTuple trigtuple, TupleTableSlot *newslot)
3218 : {
3219 135 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3220 135 : TupleTableSlot *oldslot = ExecGetTriggerOldSlot(estate, relinfo);
3221 135 : HeapTuple newtuple = NULL;
3222 : bool should_free;
3223 135 : TriggerData LocTriggerData = {0};
3224 : int i;
3225 :
3226 135 : LocTriggerData.type = T_TriggerData;
3227 135 : LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
3228 : TRIGGER_EVENT_ROW |
3229 : TRIGGER_EVENT_INSTEAD;
3230 135 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3231 :
3232 135 : ExecForceStoreHeapTuple(trigtuple, oldslot, false);
3233 :
3234 502 : for (i = 0; i < trigdesc->numtriggers; i++)
3235 : {
3236 387 : Trigger *trigger = &trigdesc->triggers[i];
3237 : HeapTuple oldtuple;
3238 :
3239 387 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3240 : TRIGGER_TYPE_ROW,
3241 : TRIGGER_TYPE_INSTEAD,
3242 : TRIGGER_TYPE_UPDATE))
3243 252 : continue;
3244 135 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3245 : NULL, oldslot, newslot))
3246 0 : continue;
3247 :
3248 135 : if (!newtuple)
3249 135 : newtuple = ExecFetchSlotHeapTuple(newslot, true, &should_free);
3250 :
3251 135 : LocTriggerData.tg_trigslot = oldslot;
3252 135 : LocTriggerData.tg_trigtuple = trigtuple;
3253 135 : LocTriggerData.tg_newslot = newslot;
3254 135 : LocTriggerData.tg_newtuple = oldtuple = newtuple;
3255 :
3256 135 : LocTriggerData.tg_trigger = trigger;
3257 135 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
3258 : i,
3259 : relinfo->ri_TrigFunctions,
3260 : relinfo->ri_TrigInstrument,
3261 135 : GetPerTupleMemoryContext(estate));
3262 127 : if (newtuple == NULL)
3263 : {
3264 12 : return false; /* "do nothing" */
3265 : }
3266 115 : else if (newtuple != oldtuple)
3267 : {
3268 92 : ExecForceStoreHeapTuple(newtuple, newslot, false);
3269 :
3270 92 : if (should_free)
3271 92 : heap_freetuple(oldtuple);
3272 :
3273 : /* signal tuple should be re-fetched if used */
3274 92 : newtuple = NULL;
3275 : }
3276 : }
3277 :
3278 115 : return true;
3279 : }
3280 :
3281 : void
3282 2513 : ExecBSTruncateTriggers(EState *estate, ResultRelInfo *relinfo)
3283 : {
3284 : TriggerDesc *trigdesc;
3285 : int i;
3286 2513 : TriggerData LocTriggerData = {0};
3287 :
3288 2513 : trigdesc = relinfo->ri_TrigDesc;
3289 :
3290 2513 : if (trigdesc == NULL)
3291 2506 : return;
3292 491 : if (!trigdesc->trig_truncate_before_statement)
3293 484 : return;
3294 :
3295 7 : LocTriggerData.type = T_TriggerData;
3296 7 : LocTriggerData.tg_event = TRIGGER_EVENT_TRUNCATE |
3297 : TRIGGER_EVENT_BEFORE;
3298 7 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3299 :
3300 20 : for (i = 0; i < trigdesc->numtriggers; i++)
3301 : {
3302 13 : Trigger *trigger = &trigdesc->triggers[i];
3303 : HeapTuple newtuple;
3304 :
3305 13 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3306 : TRIGGER_TYPE_STATEMENT,
3307 : TRIGGER_TYPE_BEFORE,
3308 : TRIGGER_TYPE_TRUNCATE))
3309 6 : continue;
3310 7 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3311 : NULL, NULL, NULL))
3312 0 : continue;
3313 :
3314 7 : LocTriggerData.tg_trigger = trigger;
3315 7 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
3316 : i,
3317 : relinfo->ri_TrigFunctions,
3318 : relinfo->ri_TrigInstrument,
3319 7 : GetPerTupleMemoryContext(estate));
3320 :
3321 7 : if (newtuple)
3322 0 : ereport(ERROR,
3323 : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
3324 : errmsg("BEFORE STATEMENT trigger cannot return a value")));
3325 : }
3326 : }
3327 :
3328 : void
3329 2509 : ExecASTruncateTriggers(EState *estate, ResultRelInfo *relinfo)
3330 : {
3331 2509 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3332 :
3333 2509 : if (trigdesc && trigdesc->trig_truncate_after_statement)
3334 5 : AfterTriggerSaveEvent(estate, relinfo,
3335 : NULL, NULL,
3336 : TRIGGER_EVENT_TRUNCATE,
3337 : false, NULL, NULL, NIL, NULL, NULL,
3338 : false);
3339 2509 : }
3340 :
3341 :
3342 : /*
3343 : * Fetch tuple into "oldslot", dealing with locking and EPQ if necessary
3344 : */
3345 : static bool
3346 8306 : GetTupleForTrigger(EState *estate,
3347 : EPQState *epqstate,
3348 : ResultRelInfo *relinfo,
3349 : ItemPointer tid,
3350 : LockTupleMode lockmode,
3351 : TupleTableSlot *oldslot,
3352 : bool do_epq_recheck,
3353 : TupleTableSlot **epqslot,
3354 : TM_Result *tmresultp,
3355 : TM_FailureData *tmfdp)
3356 : {
3357 8306 : Relation relation = relinfo->ri_RelationDesc;
3358 :
3359 8306 : if (epqslot != NULL)
3360 : {
3361 : TM_Result test;
3362 : TM_FailureData tmfd;
3363 1762 : int lockflags = 0;
3364 :
3365 1762 : *epqslot = NULL;
3366 :
3367 : /* caller must pass an epqstate if EvalPlanQual is possible */
3368 : Assert(epqstate != NULL);
3369 :
3370 : /*
3371 : * lock tuple for update
3372 : */
3373 1762 : if (!IsolationUsesXactSnapshot())
3374 1329 : lockflags |= TUPLE_LOCK_FLAG_FIND_LAST_VERSION;
3375 1762 : test = table_tuple_lock(relation, tid, estate->es_snapshot, oldslot,
3376 : estate->es_output_cid,
3377 : lockmode, LockWaitBlock,
3378 : lockflags,
3379 : &tmfd);
3380 :
3381 : /* Let the caller know about the status of this operation */
3382 1760 : if (tmresultp)
3383 148 : *tmresultp = test;
3384 1760 : if (tmfdp)
3385 1757 : *tmfdp = tmfd;
3386 :
3387 1760 : switch (test)
3388 : {
3389 4 : case TM_SelfModified:
3390 :
3391 : /*
3392 : * The target tuple was already updated or deleted by the
3393 : * current command, or by a later command in the current
3394 : * transaction. We ignore the tuple in the former case, and
3395 : * throw error in the latter case, for the same reasons
3396 : * enumerated in ExecUpdate and ExecDelete in
3397 : * nodeModifyTable.c.
3398 : */
3399 4 : if (tmfd.cmax != estate->es_output_cid)
3400 4 : ereport(ERROR,
3401 : (errcode(ERRCODE_TRIGGERED_DATA_CHANGE_VIOLATION),
3402 : errmsg("tuple to be updated was already modified by an operation triggered by the current command"),
3403 : errhint("Consider using an AFTER trigger instead of a BEFORE trigger to propagate changes to other rows.")));
3404 :
3405 : /* treat it as deleted; do not process */
3406 17 : return false;
3407 :
3408 1747 : case TM_Ok:
3409 1747 : if (tmfd.traversed)
3410 : {
3411 : /*
3412 : * Recheck the tuple using EPQ, if requested. Otherwise,
3413 : * just return that it was concurrently updated.
3414 : */
3415 14 : if (do_epq_recheck)
3416 : {
3417 6 : *epqslot = EvalPlanQual(epqstate,
3418 : relation,
3419 : relinfo->ri_RangeTableIndex,
3420 : oldslot);
3421 :
3422 : /*
3423 : * If PlanQual failed for updated tuple - we must not
3424 : * process this tuple!
3425 : */
3426 6 : if (TupIsNull(*epqslot))
3427 : {
3428 2 : *epqslot = NULL;
3429 2 : return false;
3430 : }
3431 : }
3432 : else
3433 : {
3434 8 : if (tmresultp)
3435 8 : *tmresultp = TM_Updated;
3436 8 : return false;
3437 : }
3438 : }
3439 1737 : break;
3440 :
3441 1 : case TM_Updated:
3442 1 : if (IsolationUsesXactSnapshot())
3443 1 : ereport(ERROR,
3444 : (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE),
3445 : errmsg("could not serialize access due to concurrent update")));
3446 0 : elog(ERROR, "unexpected table_tuple_lock status: %u", test);
3447 : break;
3448 :
3449 8 : case TM_Deleted:
3450 8 : if (IsolationUsesXactSnapshot())
3451 1 : ereport(ERROR,
3452 : (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE),
3453 : errmsg("could not serialize access due to concurrent delete")));
3454 : /* tuple was deleted */
3455 7 : return false;
3456 :
3457 0 : case TM_Invisible:
3458 0 : elog(ERROR, "attempted to lock invisible tuple");
3459 : break;
3460 :
3461 0 : default:
3462 0 : elog(ERROR, "unrecognized table_tuple_lock status: %u", test);
3463 : return false; /* keep compiler quiet */
3464 : }
3465 : }
3466 : else
3467 : {
3468 : /*
3469 : * We expect the tuple to be present, thus very simple error handling
3470 : * suffices.
3471 : */
3472 6544 : if (!table_tuple_fetch_row_version(relation, tid, SnapshotAny,
3473 : oldslot))
3474 0 : elog(ERROR, "failed to fetch tuple for trigger");
3475 : }
3476 :
3477 8281 : return true;
3478 : }
3479 :
3480 : /*
3481 : * Is trigger enabled to fire?
3482 : */
3483 : static bool
3484 618025 : TriggerEnabled(EState *estate, ResultRelInfo *relinfo,
3485 : Trigger *trigger, TriggerEvent event,
3486 : Bitmapset *modifiedCols,
3487 : TupleTableSlot *oldslot, TupleTableSlot *newslot)
3488 : {
3489 : /* Check replication-role-dependent enable state */
3490 618025 : if (SessionReplicationRole == SESSION_REPLICATION_ROLE_REPLICA)
3491 : {
3492 72 : if (trigger->tgenabled == TRIGGER_FIRES_ON_ORIGIN ||
3493 47 : trigger->tgenabled == TRIGGER_DISABLED)
3494 49 : return false;
3495 : }
3496 : else /* ORIGIN or LOCAL role */
3497 : {
3498 617953 : if (trigger->tgenabled == TRIGGER_FIRES_ON_REPLICA ||
3499 617952 : trigger->tgenabled == TRIGGER_DISABLED)
3500 105 : return false;
3501 : }
3502 :
3503 : /*
3504 : * Check for column-specific trigger (only possible for UPDATE, and in
3505 : * fact we *must* ignore tgattr for other event types)
3506 : */
3507 617871 : if (trigger->tgnattr > 0 && TRIGGER_FIRED_BY_UPDATE(event))
3508 : {
3509 : int i;
3510 : bool modified;
3511 :
3512 286 : modified = false;
3513 374 : for (i = 0; i < trigger->tgnattr; i++)
3514 : {
3515 318 : if (bms_is_member(trigger->tgattr[i] - FirstLowInvalidHeapAttributeNumber,
3516 : modifiedCols))
3517 : {
3518 230 : modified = true;
3519 230 : break;
3520 : }
3521 : }
3522 286 : if (!modified)
3523 56 : return false;
3524 : }
3525 :
3526 : /* Check for WHEN clause */
3527 617815 : if (trigger->tgqual)
3528 : {
3529 : ExprState **predicate;
3530 : ExprContext *econtext;
3531 : MemoryContext oldContext;
3532 : int i;
3533 :
3534 : Assert(estate != NULL);
3535 :
3536 : /*
3537 : * trigger is an element of relinfo->ri_TrigDesc->triggers[]; find the
3538 : * matching element of relinfo->ri_TrigWhenExprs[]
3539 : */
3540 376 : i = trigger - relinfo->ri_TrigDesc->triggers;
3541 376 : predicate = &relinfo->ri_TrigWhenExprs[i];
3542 :
3543 : /*
3544 : * If first time through for this WHEN expression, build expression
3545 : * nodetrees for it. Keep them in the per-query memory context so
3546 : * they'll survive throughout the query.
3547 : */
3548 376 : if (*predicate == NULL)
3549 : {
3550 : Node *tgqual;
3551 :
3552 198 : oldContext = MemoryContextSwitchTo(estate->es_query_cxt);
3553 198 : tgqual = stringToNode(trigger->tgqual);
3554 198 : tgqual = expand_generated_columns_in_expr(tgqual, relinfo->ri_RelationDesc, PRS2_OLD_VARNO);
3555 198 : tgqual = expand_generated_columns_in_expr(tgqual, relinfo->ri_RelationDesc, PRS2_NEW_VARNO);
3556 : /* Change references to OLD and NEW to INNER_VAR and OUTER_VAR */
3557 198 : ChangeVarNodes(tgqual, PRS2_OLD_VARNO, INNER_VAR, 0);
3558 198 : ChangeVarNodes(tgqual, PRS2_NEW_VARNO, OUTER_VAR, 0);
3559 : /* ExecPrepareQual wants implicit-AND form */
3560 198 : tgqual = (Node *) make_ands_implicit((Expr *) tgqual);
3561 198 : *predicate = ExecPrepareQual((List *) tgqual, estate);
3562 198 : MemoryContextSwitchTo(oldContext);
3563 : }
3564 :
3565 : /*
3566 : * We will use the EState's per-tuple context for evaluating WHEN
3567 : * expressions (creating it if it's not already there).
3568 : */
3569 376 : econtext = GetPerTupleExprContext(estate);
3570 :
3571 : /*
3572 : * Finally evaluate the expression, making the old and/or new tuples
3573 : * available as INNER_VAR/OUTER_VAR respectively.
3574 : */
3575 376 : econtext->ecxt_innertuple = oldslot;
3576 376 : econtext->ecxt_outertuple = newslot;
3577 376 : if (!ExecQual(*predicate, econtext))
3578 210 : return false;
3579 : }
3580 :
3581 617605 : return true;
3582 : }
3583 :
3584 :
3585 : /* ----------
3586 : * After-trigger stuff
3587 : *
3588 : * The AfterTriggersData struct holds data about pending AFTER trigger events
3589 : * during the current transaction tree. (BEFORE triggers are fired
3590 : * immediately so we don't need any persistent state about them.) The struct
3591 : * and most of its subsidiary data are kept in TopTransactionContext; however
3592 : * some data that can be discarded sooner appears in the CurTransactionContext
3593 : * of the relevant subtransaction. Also, the individual event records are
3594 : * kept in a separate sub-context of TopTransactionContext. This is done
3595 : * mainly so that it's easy to tell from a memory context dump how much space
3596 : * is being eaten by trigger events.
3597 : *
3598 : * Because the list of pending events can grow large, we go to some
3599 : * considerable effort to minimize per-event memory consumption. The event
3600 : * records are grouped into chunks and common data for similar events in the
3601 : * same chunk is only stored once.
3602 : *
3603 : * XXX We need to be able to save the per-event data in a file if it grows too
3604 : * large.
3605 : * ----------
3606 : */
3607 :
3608 : /* Per-trigger SET CONSTRAINT status */
3609 : typedef struct SetConstraintTriggerData
3610 : {
3611 : Oid sct_tgoid;
3612 : bool sct_tgisdeferred;
3613 : } SetConstraintTriggerData;
3614 :
3615 : typedef struct SetConstraintTriggerData *SetConstraintTrigger;
3616 :
3617 : /*
3618 : * SET CONSTRAINT intra-transaction status.
3619 : *
3620 : * We make this a single palloc'd object so it can be copied and freed easily.
3621 : *
3622 : * all_isset and all_isdeferred are used to keep track
3623 : * of SET CONSTRAINTS ALL {DEFERRED, IMMEDIATE}.
3624 : *
3625 : * trigstates[] stores per-trigger tgisdeferred settings.
3626 : */
3627 : typedef struct SetConstraintStateData
3628 : {
3629 : bool all_isset;
3630 : bool all_isdeferred;
3631 : int numstates; /* number of trigstates[] entries in use */
3632 : int numalloc; /* allocated size of trigstates[] */
3633 : SetConstraintTriggerData trigstates[FLEXIBLE_ARRAY_MEMBER];
3634 : } SetConstraintStateData;
3635 :
3636 : typedef SetConstraintStateData *SetConstraintState;
3637 :
3638 :
3639 : /*
3640 : * Per-trigger-event data
3641 : *
3642 : * The actual per-event data, AfterTriggerEventData, includes DONE/IN_PROGRESS
3643 : * status bits, up to two tuple CTIDs, and optionally two OIDs of partitions.
3644 : * Each event record also has an associated AfterTriggerSharedData that is
3645 : * shared across all instances of similar events within a "chunk".
3646 : *
3647 : * For row-level triggers, we arrange not to waste storage on unneeded ctid
3648 : * fields. Updates of regular tables use two; inserts and deletes of regular
3649 : * tables use one; foreign tables always use zero and save the tuple(s) to a
3650 : * tuplestore. AFTER_TRIGGER_FDW_FETCH directs AfterTriggerExecute() to
3651 : * retrieve a fresh tuple or pair of tuples from that tuplestore, while
3652 : * AFTER_TRIGGER_FDW_REUSE directs it to use the most-recently-retrieved
3653 : * tuple(s). This permits storing tuples once regardless of the number of
3654 : * row-level triggers on a foreign table.
3655 : *
3656 : * When updates on partitioned tables cause rows to move between partitions,
3657 : * the OIDs of both partitions are stored too, so that the tuples can be
3658 : * fetched; such entries are marked AFTER_TRIGGER_CP_UPDATE (for "cross-
3659 : * partition update").
3660 : *
3661 : * Note that we need triggers on foreign tables to be fired in exactly the
3662 : * order they were queued, so that the tuples come out of the tuplestore in
3663 : * the right order. To ensure that, we forbid deferrable (constraint)
3664 : * triggers on foreign tables. This also ensures that such triggers do not
3665 : * get deferred into outer trigger query levels, meaning that it's okay to
3666 : * destroy the tuplestore at the end of the query level.
3667 : *
3668 : * Statement-level triggers always bear AFTER_TRIGGER_1CTID, though they
3669 : * require no ctid field. We lack the flag bit space to neatly represent that
3670 : * distinct case, and it seems unlikely to be worth much trouble.
3671 : *
3672 : * Note: ats_firing_id is initially zero and is set to something else when
3673 : * AFTER_TRIGGER_IN_PROGRESS is set. It indicates which trigger firing
3674 : * cycle the trigger will be fired in (or was fired in, if DONE is set).
3675 : * Although this is mutable state, we can keep it in AfterTriggerSharedData
3676 : * because all instances of the same type of event in a given event list will
3677 : * be fired at the same time, if they were queued between the same firing
3678 : * cycles. So we need only ensure that ats_firing_id is zero when attaching
3679 : * a new event to an existing AfterTriggerSharedData record.
3680 : */
3681 : typedef uint32 TriggerFlags;
3682 :
3683 : #define AFTER_TRIGGER_OFFSET 0x07FFFFFF /* must be low-order bits */
3684 : #define AFTER_TRIGGER_DONE 0x80000000
3685 : #define AFTER_TRIGGER_IN_PROGRESS 0x40000000
3686 : /* bits describing the size and tuple sources of this event */
3687 : #define AFTER_TRIGGER_FDW_REUSE 0x00000000
3688 : #define AFTER_TRIGGER_FDW_FETCH 0x20000000
3689 : #define AFTER_TRIGGER_1CTID 0x10000000
3690 : #define AFTER_TRIGGER_2CTID 0x30000000
3691 : #define AFTER_TRIGGER_CP_UPDATE 0x08000000
3692 : #define AFTER_TRIGGER_TUP_BITS 0x38000000
3693 : typedef struct AfterTriggerSharedData *AfterTriggerShared;
3694 :
3695 : typedef struct AfterTriggerSharedData
3696 : {
3697 : TriggerEvent ats_event; /* event type indicator, see trigger.h */
3698 : Oid ats_tgoid; /* the trigger's ID */
3699 : Oid ats_relid; /* the relation it's on */
3700 : Oid ats_rolid; /* role to execute the trigger */
3701 : CommandId ats_firing_id; /* ID for firing cycle */
3702 : struct AfterTriggersTableData *ats_table; /* transition table access */
3703 : Bitmapset *ats_modifiedcols; /* modified columns */
3704 : } AfterTriggerSharedData;
3705 :
3706 : typedef struct AfterTriggerEventData *AfterTriggerEvent;
3707 :
3708 : typedef struct AfterTriggerEventData
3709 : {
3710 : TriggerFlags ate_flags; /* status bits and offset to shared data */
3711 : ItemPointerData ate_ctid1; /* inserted, deleted, or old updated tuple */
3712 : ItemPointerData ate_ctid2; /* new updated tuple */
3713 :
3714 : /*
3715 : * During a cross-partition update of a partitioned table, we also store
3716 : * the OIDs of source and destination partitions that are needed to fetch
3717 : * the old (ctid1) and the new tuple (ctid2) from, respectively.
3718 : */
3719 : Oid ate_src_part;
3720 : Oid ate_dst_part;
3721 : } AfterTriggerEventData;
3722 :
3723 : /* AfterTriggerEventData, minus ate_src_part, ate_dst_part */
3724 : typedef struct AfterTriggerEventDataNoOids
3725 : {
3726 : TriggerFlags ate_flags;
3727 : ItemPointerData ate_ctid1;
3728 : ItemPointerData ate_ctid2;
3729 : } AfterTriggerEventDataNoOids;
3730 :
3731 : /* AfterTriggerEventData, minus ate_*_part and ate_ctid2 */
3732 : typedef struct AfterTriggerEventDataOneCtid
3733 : {
3734 : TriggerFlags ate_flags; /* status bits and offset to shared data */
3735 : ItemPointerData ate_ctid1; /* inserted, deleted, or old updated tuple */
3736 : } AfterTriggerEventDataOneCtid;
3737 :
3738 : /* AfterTriggerEventData, minus ate_*_part, ate_ctid1 and ate_ctid2 */
3739 : typedef struct AfterTriggerEventDataZeroCtids
3740 : {
3741 : TriggerFlags ate_flags; /* status bits and offset to shared data */
3742 : } AfterTriggerEventDataZeroCtids;
3743 :
3744 : #define SizeofTriggerEvent(evt) \
3745 : (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_CP_UPDATE ? \
3746 : sizeof(AfterTriggerEventData) : \
3747 : (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_2CTID ? \
3748 : sizeof(AfterTriggerEventDataNoOids) : \
3749 : (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_1CTID ? \
3750 : sizeof(AfterTriggerEventDataOneCtid) : \
3751 : sizeof(AfterTriggerEventDataZeroCtids))))
3752 :
3753 : #define GetTriggerSharedData(evt) \
3754 : ((AfterTriggerShared) ((char *) (evt) + ((evt)->ate_flags & AFTER_TRIGGER_OFFSET)))
3755 :
3756 : /*
3757 : * To avoid palloc overhead, we keep trigger events in arrays in successively-
3758 : * larger chunks (a slightly more sophisticated version of an expansible
3759 : * array). The space between CHUNK_DATA_START and freeptr is occupied by
3760 : * AfterTriggerEventData records; the space between endfree and endptr is
3761 : * occupied by AfterTriggerSharedData records.
3762 : */
3763 : typedef struct AfterTriggerEventChunk
3764 : {
3765 : struct AfterTriggerEventChunk *next; /* list link */
3766 : char *freeptr; /* start of free space in chunk */
3767 : char *endfree; /* end of free space in chunk */
3768 : char *endptr; /* end of chunk */
3769 : /* event data follows here */
3770 : } AfterTriggerEventChunk;
3771 :
3772 : #define CHUNK_DATA_START(cptr) ((char *) (cptr) + MAXALIGN(sizeof(AfterTriggerEventChunk)))
3773 :
3774 : /* A list of events */
3775 : typedef struct AfterTriggerEventList
3776 : {
3777 : AfterTriggerEventChunk *head;
3778 : AfterTriggerEventChunk *tail;
3779 : char *tailfree; /* freeptr of tail chunk */
3780 : } AfterTriggerEventList;
3781 :
3782 : /* Macros to help in iterating over a list of events */
3783 : #define for_each_chunk(cptr, evtlist) \
3784 : for (cptr = (evtlist).head; cptr != NULL; cptr = cptr->next)
3785 : #define for_each_event(eptr, cptr) \
3786 : for (eptr = (AfterTriggerEvent) CHUNK_DATA_START(cptr); \
3787 : (char *) eptr < (cptr)->freeptr; \
3788 : eptr = (AfterTriggerEvent) (((char *) eptr) + SizeofTriggerEvent(eptr)))
3789 : /* Use this if no special per-chunk processing is needed */
3790 : #define for_each_event_chunk(eptr, cptr, evtlist) \
3791 : for_each_chunk(cptr, evtlist) for_each_event(eptr, cptr)
3792 :
3793 : /* Macros for iterating from a start point that might not be list start */
3794 : #define for_each_chunk_from(cptr) \
3795 : for (; cptr != NULL; cptr = cptr->next)
3796 : #define for_each_event_from(eptr, cptr) \
3797 : for (; \
3798 : (char *) eptr < (cptr)->freeptr; \
3799 : eptr = (AfterTriggerEvent) (((char *) eptr) + SizeofTriggerEvent(eptr)))
3800 :
3801 :
3802 : /*
3803 : * All per-transaction data for the AFTER TRIGGERS module.
3804 : *
3805 : * AfterTriggersData has the following fields:
3806 : *
3807 : * firing_counter is incremented for each call of afterTriggerInvokeEvents.
3808 : * We mark firable events with the current firing cycle's ID so that we can
3809 : * tell which ones to work on. This ensures sane behavior if a trigger
3810 : * function chooses to do SET CONSTRAINTS: the inner SET CONSTRAINTS will
3811 : * only fire those events that weren't already scheduled for firing.
3812 : *
3813 : * state keeps track of the transaction-local effects of SET CONSTRAINTS.
3814 : * This is saved and restored across failed subtransactions.
3815 : *
3816 : * events is the current list of deferred events. This is global across
3817 : * all subtransactions of the current transaction. In a subtransaction
3818 : * abort, we know that the events added by the subtransaction are at the
3819 : * end of the list, so it is relatively easy to discard them. The event
3820 : * list chunks themselves are stored in event_cxt.
3821 : *
3822 : * query_depth is the current depth of nested AfterTriggerBeginQuery calls
3823 : * (-1 when the stack is empty).
3824 : *
3825 : * query_stack[query_depth] is the per-query-level data, including these fields:
3826 : *
3827 : * events is a list of AFTER trigger events queued by the current query.
3828 : * None of these are valid until the matching AfterTriggerEndQuery call
3829 : * occurs. At that point we fire immediate-mode triggers, and append any
3830 : * deferred events to the main events list.
3831 : *
3832 : * fdw_tuplestore is a tuplestore containing the foreign-table tuples
3833 : * needed by events queued by the current query. (Note: we use just one
3834 : * tuplestore even though more than one foreign table might be involved.
3835 : * This is okay because tuplestores don't really care what's in the tuples
3836 : * they store; but it's possible that someday it'd break.)
3837 : *
3838 : * tables is a List of AfterTriggersTableData structs for target tables
3839 : * of the current query (see below).
3840 : *
3841 : * maxquerydepth is just the allocated length of query_stack.
3842 : *
3843 : * trans_stack holds per-subtransaction data, including these fields:
3844 : *
3845 : * state is NULL or a pointer to a saved copy of the SET CONSTRAINTS
3846 : * state data. Each subtransaction level that modifies that state first
3847 : * saves a copy, which we use to restore the state if we abort.
3848 : *
3849 : * events is a copy of the events head/tail pointers,
3850 : * which we use to restore those values during subtransaction abort.
3851 : *
3852 : * query_depth is the subtransaction-start-time value of query_depth,
3853 : * which we similarly use to clean up at subtransaction abort.
3854 : *
3855 : * firing_counter is the subtransaction-start-time value of firing_counter.
3856 : * We use this to recognize which deferred triggers were fired (or marked
3857 : * for firing) within an aborted subtransaction.
3858 : *
3859 : * We use GetCurrentTransactionNestLevel() to determine the correct array
3860 : * index in trans_stack. maxtransdepth is the number of allocated entries in
3861 : * trans_stack. (By not keeping our own stack pointer, we can avoid trouble
3862 : * in cases where errors during subxact abort cause multiple invocations
3863 : * of AfterTriggerEndSubXact() at the same nesting depth.)
3864 : *
3865 : * We create an AfterTriggersTableData struct for each target table of the
3866 : * current query, and each operation mode (INSERT/UPDATE/DELETE), that has
3867 : * either transition tables or statement-level triggers. This is used to
3868 : * hold the relevant transition tables, as well as info tracking whether
3869 : * we already queued the statement triggers. (We use that info to prevent
3870 : * firing the same statement triggers more than once per statement, or really
3871 : * once per transition table set.) These structs, along with the transition
3872 : * table tuplestores, live in the (sub)transaction's CurTransactionContext.
3873 : * That's sufficient lifespan because we don't allow transition tables to be
3874 : * used by deferrable triggers, so they only need to survive until
3875 : * AfterTriggerEndQuery.
3876 : */
3877 : typedef struct AfterTriggersQueryData AfterTriggersQueryData;
3878 : typedef struct AfterTriggersTransData AfterTriggersTransData;
3879 : typedef struct AfterTriggersTableData AfterTriggersTableData;
3880 :
3881 : typedef struct AfterTriggersData
3882 : {
3883 : CommandId firing_counter; /* next firing ID to assign */
3884 : SetConstraintState state; /* the active S C state */
3885 : AfterTriggerEventList events; /* deferred-event list */
3886 : MemoryContext event_cxt; /* memory context for events, if any */
3887 :
3888 : /* per-query-level data: */
3889 : AfterTriggersQueryData *query_stack; /* array of structs shown below */
3890 : int query_depth; /* current index in above array */
3891 : int maxquerydepth; /* allocated len of above array */
3892 :
3893 : /* per-subtransaction-level data: */
3894 : AfterTriggersTransData *trans_stack; /* array of structs shown below */
3895 : int maxtransdepth; /* allocated len of above array */
3896 :
3897 : List *batch_callbacks; /* List of AfterTriggerCallbackItem */
3898 : } AfterTriggersData;
3899 :
3900 : struct AfterTriggersQueryData
3901 : {
3902 : AfterTriggerEventList events; /* events pending from this query */
3903 : Tuplestorestate *fdw_tuplestore; /* foreign tuples for said events */
3904 : List *tables; /* list of AfterTriggersTableData, see below */
3905 : };
3906 :
3907 : struct AfterTriggersTransData
3908 : {
3909 : /* these fields are just for resetting at subtrans abort: */
3910 : SetConstraintState state; /* saved S C state, or NULL if not yet saved */
3911 : AfterTriggerEventList events; /* saved list pointer */
3912 : int query_depth; /* saved query_depth */
3913 : CommandId firing_counter; /* saved firing_counter */
3914 : };
3915 :
3916 : struct AfterTriggersTableData
3917 : {
3918 : /* relid + cmdType form the lookup key for these structs: */
3919 : Oid relid; /* target table's OID */
3920 : CmdType cmdType; /* event type, CMD_INSERT/UPDATE/DELETE */
3921 : bool closed; /* true when no longer OK to add tuples */
3922 : bool before_trig_done; /* did we already queue BS triggers? */
3923 : bool after_trig_done; /* did we already queue AS triggers? */
3924 : AfterTriggerEventList after_trig_events; /* if so, saved list pointer */
3925 :
3926 : /* "old" transition table for UPDATE/DELETE, if any */
3927 : Tuplestorestate *old_tuplestore;
3928 : /* "new" transition table for INSERT/UPDATE, if any */
3929 : Tuplestorestate *new_tuplestore;
3930 :
3931 : TupleTableSlot *storeslot; /* for converting to tuplestore's format */
3932 : };
3933 :
3934 : /* Entry in afterTriggers.batch_callbacks */
3935 : typedef struct AfterTriggerCallbackItem
3936 : {
3937 : AfterTriggerBatchCallback callback;
3938 : void *arg;
3939 : } AfterTriggerCallbackItem;
3940 :
3941 : static AfterTriggersData afterTriggers;
3942 :
3943 : /*
3944 : * Incremented before invoking afterTriggerInvokeEvents(). Used by
3945 : * AfterTriggerIsActive() to determine whether batch callbacks will fire,
3946 : * so that RI trigger functions can take the batched fast path.
3947 : */
3948 : static int afterTriggerFiringDepth = 0;
3949 :
3950 : static void AfterTriggerExecute(EState *estate,
3951 : AfterTriggerEvent event,
3952 : ResultRelInfo *relInfo,
3953 : ResultRelInfo *src_relInfo,
3954 : ResultRelInfo *dst_relInfo,
3955 : TriggerDesc *trigdesc,
3956 : FmgrInfo *finfo,
3957 : TriggerInstrumentation *instr,
3958 : MemoryContext per_tuple_context,
3959 : TupleTableSlot *trig_tuple_slot1,
3960 : TupleTableSlot *trig_tuple_slot2);
3961 : static AfterTriggersTableData *GetAfterTriggersTableData(Oid relid,
3962 : CmdType cmdType);
3963 : static TupleTableSlot *GetAfterTriggersStoreSlot(AfterTriggersTableData *table,
3964 : TupleDesc tupdesc);
3965 : static Tuplestorestate *GetAfterTriggersTransitionTable(int event,
3966 : TupleTableSlot *oldslot,
3967 : TupleTableSlot *newslot,
3968 : TransitionCaptureState *transition_capture);
3969 : static void TransitionTableAddTuple(EState *estate,
3970 : int event,
3971 : TransitionCaptureState *transition_capture,
3972 : ResultRelInfo *relinfo,
3973 : TupleTableSlot *slot,
3974 : TupleTableSlot *original_insert_tuple,
3975 : Tuplestorestate *tuplestore);
3976 : static void AfterTriggerFreeQuery(AfterTriggersQueryData *qs);
3977 : static SetConstraintState SetConstraintStateCreate(int numalloc);
3978 : static SetConstraintState SetConstraintStateCopy(SetConstraintState origstate);
3979 : static SetConstraintState SetConstraintStateAddItem(SetConstraintState state,
3980 : Oid tgoid, bool tgisdeferred);
3981 : static void cancel_prior_stmt_triggers(Oid relid, CmdType cmdType, int tgevent);
3982 :
3983 : static void FireAfterTriggerBatchCallbacks(void);
3984 :
3985 : /*
3986 : * Get the FDW tuplestore for the current trigger query level, creating it
3987 : * if necessary.
3988 : */
3989 : static Tuplestorestate *
3990 50 : GetCurrentFDWTuplestore(void)
3991 : {
3992 : Tuplestorestate *ret;
3993 :
3994 50 : ret = afterTriggers.query_stack[afterTriggers.query_depth].fdw_tuplestore;
3995 50 : if (ret == NULL)
3996 : {
3997 : MemoryContext oldcxt;
3998 : ResourceOwner saveResourceOwner;
3999 :
4000 : /*
4001 : * Make the tuplestore valid until end of subtransaction. We really
4002 : * only need it until AfterTriggerEndQuery().
4003 : */
4004 18 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4005 18 : saveResourceOwner = CurrentResourceOwner;
4006 18 : CurrentResourceOwner = CurTransactionResourceOwner;
4007 :
4008 18 : ret = tuplestore_begin_heap(false, false, work_mem);
4009 :
4010 18 : CurrentResourceOwner = saveResourceOwner;
4011 18 : MemoryContextSwitchTo(oldcxt);
4012 :
4013 18 : afterTriggers.query_stack[afterTriggers.query_depth].fdw_tuplestore = ret;
4014 : }
4015 :
4016 50 : return ret;
4017 : }
4018 :
4019 : /* ----------
4020 : * afterTriggerCheckState()
4021 : *
4022 : * Returns true if the trigger event is actually in state DEFERRED.
4023 : * ----------
4024 : */
4025 : static bool
4026 609796 : afterTriggerCheckState(AfterTriggerShared evtshared)
4027 : {
4028 609796 : Oid tgoid = evtshared->ats_tgoid;
4029 609796 : SetConstraintState state = afterTriggers.state;
4030 : int i;
4031 :
4032 : /*
4033 : * For not-deferrable triggers (i.e. normal AFTER ROW triggers and
4034 : * constraints declared NOT DEFERRABLE), the state is always false.
4035 : */
4036 609796 : if ((evtshared->ats_event & AFTER_TRIGGER_DEFERRABLE) == 0)
4037 609255 : return false;
4038 :
4039 : /*
4040 : * If constraint state exists, SET CONSTRAINTS might have been executed
4041 : * either for this trigger or for all triggers.
4042 : */
4043 541 : if (state != NULL)
4044 : {
4045 : /* Check for SET CONSTRAINTS for this specific trigger. */
4046 224 : for (i = 0; i < state->numstates; i++)
4047 : {
4048 169 : if (state->trigstates[i].sct_tgoid == tgoid)
4049 40 : return state->trigstates[i].sct_tgisdeferred;
4050 : }
4051 :
4052 : /* Check for SET CONSTRAINTS ALL. */
4053 55 : if (state->all_isset)
4054 47 : return state->all_isdeferred;
4055 : }
4056 :
4057 : /*
4058 : * Otherwise return the default state for the trigger.
4059 : */
4060 454 : return ((evtshared->ats_event & AFTER_TRIGGER_INITDEFERRED) != 0);
4061 : }
4062 :
4063 : /* ----------
4064 : * afterTriggerCopyBitmap()
4065 : *
4066 : * Copy bitmap into AfterTriggerEvents memory context, which is where the after
4067 : * trigger events are kept.
4068 : * ----------
4069 : */
4070 : static Bitmapset *
4071 7609 : afterTriggerCopyBitmap(Bitmapset *src)
4072 : {
4073 : Bitmapset *dst;
4074 : MemoryContext oldcxt;
4075 :
4076 7609 : if (src == NULL)
4077 5386 : return NULL;
4078 :
4079 2223 : oldcxt = MemoryContextSwitchTo(afterTriggers.event_cxt);
4080 :
4081 2223 : dst = bms_copy(src);
4082 :
4083 2223 : MemoryContextSwitchTo(oldcxt);
4084 :
4085 2223 : return dst;
4086 : }
4087 :
4088 : /* ----------
4089 : * afterTriggerAddEvent()
4090 : *
4091 : * Add a new trigger event to the specified queue.
4092 : * The passed-in event data is copied.
4093 : * ----------
4094 : */
4095 : static void
4096 610281 : afterTriggerAddEvent(AfterTriggerEventList *events,
4097 : AfterTriggerEvent event, AfterTriggerShared evtshared)
4098 : {
4099 610281 : Size eventsize = SizeofTriggerEvent(event);
4100 610281 : Size needed = eventsize + sizeof(AfterTriggerSharedData);
4101 : AfterTriggerEventChunk *chunk;
4102 : AfterTriggerShared newshared;
4103 : AfterTriggerEvent newevent;
4104 :
4105 : /*
4106 : * If empty list or not enough room in the tail chunk, make a new chunk.
4107 : * We assume here that a new shared record will always be needed.
4108 : */
4109 610281 : chunk = events->tail;
4110 610281 : if (chunk == NULL ||
4111 604918 : chunk->endfree - chunk->freeptr < needed)
4112 : {
4113 : Size chunksize;
4114 :
4115 : /* Create event context if we didn't already */
4116 5417 : if (afterTriggers.event_cxt == NULL)
4117 4455 : afterTriggers.event_cxt =
4118 4455 : AllocSetContextCreate(TopTransactionContext,
4119 : "AfterTriggerEvents",
4120 : ALLOCSET_DEFAULT_SIZES);
4121 :
4122 : /*
4123 : * Chunk size starts at 1KB and is allowed to increase up to 1MB.
4124 : * These numbers are fairly arbitrary, though there is a hard limit at
4125 : * AFTER_TRIGGER_OFFSET; else we couldn't link event records to their
4126 : * shared records using the available space in ate_flags. Another
4127 : * constraint is that if the chunk size gets too huge, the search loop
4128 : * below would get slow given a (not too common) usage pattern with
4129 : * many distinct event types in a chunk. Therefore, we double the
4130 : * preceding chunk size only if there weren't too many shared records
4131 : * in the preceding chunk; otherwise we halve it. This gives us some
4132 : * ability to adapt to the actual usage pattern of the current query
4133 : * while still having large chunk sizes in typical usage. All chunk
4134 : * sizes used should be MAXALIGN multiples, to ensure that the shared
4135 : * records will be aligned safely.
4136 : */
4137 : #define MIN_CHUNK_SIZE 1024
4138 : #define MAX_CHUNK_SIZE (1024*1024)
4139 :
4140 : #if MAX_CHUNK_SIZE > (AFTER_TRIGGER_OFFSET+1)
4141 : #error MAX_CHUNK_SIZE must not exceed AFTER_TRIGGER_OFFSET
4142 : #endif
4143 :
4144 5417 : if (chunk == NULL)
4145 5363 : chunksize = MIN_CHUNK_SIZE;
4146 : else
4147 : {
4148 : /* preceding chunk size... */
4149 54 : chunksize = chunk->endptr - (char *) chunk;
4150 : /* check number of shared records in preceding chunk */
4151 54 : if ((chunk->endptr - chunk->endfree) <=
4152 : (100 * sizeof(AfterTriggerSharedData)))
4153 54 : chunksize *= 2; /* okay, double it */
4154 : else
4155 0 : chunksize /= 2; /* too many shared records */
4156 54 : chunksize = Min(chunksize, MAX_CHUNK_SIZE);
4157 : }
4158 5417 : chunk = MemoryContextAlloc(afterTriggers.event_cxt, chunksize);
4159 5417 : chunk->next = NULL;
4160 5417 : chunk->freeptr = CHUNK_DATA_START(chunk);
4161 5417 : chunk->endptr = chunk->endfree = (char *) chunk + chunksize;
4162 : Assert(chunk->endfree - chunk->freeptr >= needed);
4163 :
4164 5417 : if (events->tail == NULL)
4165 : {
4166 : Assert(events->head == NULL);
4167 5363 : events->head = chunk;
4168 : }
4169 : else
4170 54 : events->tail->next = chunk;
4171 5417 : events->tail = chunk;
4172 : /* events->tailfree is now out of sync, but we'll fix it below */
4173 : }
4174 :
4175 : /*
4176 : * Try to locate a matching shared-data record already in the chunk. If
4177 : * none, make a new one. The search begins with the most recently added
4178 : * record, since newer ones are most likely to match.
4179 : */
4180 610281 : for (newshared = (AfterTriggerShared) chunk->endfree;
4181 813848 : (char *) newshared < chunk->endptr;
4182 203567 : newshared++)
4183 : {
4184 : /* compare fields roughly by probability of them being different */
4185 806239 : if (newshared->ats_tgoid == evtshared->ats_tgoid &&
4186 602817 : newshared->ats_event == evtshared->ats_event &&
4187 602813 : newshared->ats_firing_id == 0 &&
4188 602697 : newshared->ats_table == evtshared->ats_table &&
4189 602697 : newshared->ats_relid == evtshared->ats_relid &&
4190 1205390 : newshared->ats_rolid == evtshared->ats_rolid &&
4191 602693 : bms_equal(newshared->ats_modifiedcols,
4192 602693 : evtshared->ats_modifiedcols))
4193 602672 : break;
4194 : }
4195 610281 : if ((char *) newshared >= chunk->endptr)
4196 : {
4197 7609 : newshared = ((AfterTriggerShared) chunk->endfree) - 1;
4198 7609 : *newshared = *evtshared;
4199 : /* now we must make a suitably-long-lived copy of the bitmap */
4200 7609 : newshared->ats_modifiedcols = afterTriggerCopyBitmap(evtshared->ats_modifiedcols);
4201 7609 : newshared->ats_firing_id = 0; /* just to be sure */
4202 7609 : chunk->endfree = (char *) newshared;
4203 : }
4204 :
4205 : /* Insert the data */
4206 610281 : newevent = (AfterTriggerEvent) chunk->freeptr;
4207 610281 : memcpy(newevent, event, eventsize);
4208 : /* ... and link the new event to its shared record */
4209 610281 : newevent->ate_flags &= ~AFTER_TRIGGER_OFFSET;
4210 610281 : newevent->ate_flags |= (char *) newshared - (char *) newevent;
4211 :
4212 610281 : chunk->freeptr += eventsize;
4213 610281 : events->tailfree = chunk->freeptr;
4214 610281 : }
4215 :
4216 : /* ----------
4217 : * afterTriggerFreeEventList()
4218 : *
4219 : * Free all the event storage in the given list.
4220 : * ----------
4221 : */
4222 : static void
4223 11193 : afterTriggerFreeEventList(AfterTriggerEventList *events)
4224 : {
4225 : AfterTriggerEventChunk *chunk;
4226 :
4227 15537 : while ((chunk = events->head) != NULL)
4228 : {
4229 4344 : events->head = chunk->next;
4230 4344 : pfree(chunk);
4231 : }
4232 11193 : events->tail = NULL;
4233 11193 : events->tailfree = NULL;
4234 11193 : }
4235 :
4236 : /* ----------
4237 : * afterTriggerRestoreEventList()
4238 : *
4239 : * Restore an event list to its prior length, removing all the events
4240 : * added since it had the value old_events.
4241 : * ----------
4242 : */
4243 : static void
4244 6007 : afterTriggerRestoreEventList(AfterTriggerEventList *events,
4245 : const AfterTriggerEventList *old_events)
4246 : {
4247 : AfterTriggerEventChunk *chunk;
4248 : AfterTriggerEventChunk *next_chunk;
4249 :
4250 6007 : if (old_events->tail == NULL)
4251 : {
4252 : /* restoring to a completely empty state, so free everything */
4253 5993 : afterTriggerFreeEventList(events);
4254 : }
4255 : else
4256 : {
4257 14 : *events = *old_events;
4258 : /* free any chunks after the last one we want to keep */
4259 14 : for (chunk = events->tail->next; chunk != NULL; chunk = next_chunk)
4260 : {
4261 0 : next_chunk = chunk->next;
4262 0 : pfree(chunk);
4263 : }
4264 : /* and clean up the tail chunk to be the right length */
4265 14 : events->tail->next = NULL;
4266 14 : events->tail->freeptr = events->tailfree;
4267 :
4268 : /*
4269 : * We don't make any effort to remove now-unused shared data records.
4270 : * They might still be useful, anyway.
4271 : */
4272 : }
4273 6007 : }
4274 :
4275 : /* ----------
4276 : * afterTriggerDeleteHeadEventChunk()
4277 : *
4278 : * Remove the first chunk of events from the query level's event list.
4279 : * Keep any event list pointers elsewhere in the query level's data
4280 : * structures in sync.
4281 : * ----------
4282 : */
4283 : static void
4284 0 : afterTriggerDeleteHeadEventChunk(AfterTriggersQueryData *qs)
4285 : {
4286 0 : AfterTriggerEventChunk *target = qs->events.head;
4287 : ListCell *lc;
4288 :
4289 : Assert(target && target->next);
4290 :
4291 : /*
4292 : * First, update any pointers in the per-table data, so that they won't be
4293 : * dangling. Resetting obsoleted pointers to NULL will make
4294 : * cancel_prior_stmt_triggers start from the list head, which is fine.
4295 : */
4296 0 : foreach(lc, qs->tables)
4297 : {
4298 0 : AfterTriggersTableData *table = (AfterTriggersTableData *) lfirst(lc);
4299 :
4300 0 : if (table->after_trig_done &&
4301 0 : table->after_trig_events.tail == target)
4302 : {
4303 0 : table->after_trig_events.head = NULL;
4304 0 : table->after_trig_events.tail = NULL;
4305 0 : table->after_trig_events.tailfree = NULL;
4306 : }
4307 : }
4308 :
4309 : /* Now we can flush the head chunk */
4310 0 : qs->events.head = target->next;
4311 0 : pfree(target);
4312 0 : }
4313 :
4314 :
4315 : /* ----------
4316 : * AfterTriggerExecute()
4317 : *
4318 : * Fetch the required tuples back from the heap and fire one
4319 : * single trigger function.
4320 : *
4321 : * Frequently, this will be fired many times in a row for triggers of
4322 : * a single relation. Therefore, we cache the open relation and provide
4323 : * fmgr lookup cache space at the caller level. (For triggers fired at
4324 : * the end of a query, we can even piggyback on the executor's state.)
4325 : *
4326 : * When fired for a cross-partition update of a partitioned table, the old
4327 : * tuple is fetched using 'src_relInfo' (the source leaf partition) and
4328 : * the new tuple using 'dst_relInfo' (the destination leaf partition), though
4329 : * both are converted into the root partitioned table's format before passing
4330 : * to the trigger function.
4331 : *
4332 : * event: event currently being fired.
4333 : * relInfo: result relation for event.
4334 : * src_relInfo: source partition of a cross-partition update
4335 : * dst_relInfo: its destination partition
4336 : * trigdesc: working copy of rel's trigger info.
4337 : * finfo: array of fmgr lookup cache entries (one per trigger in trigdesc).
4338 : * instr: array of EXPLAIN ANALYZE instrumentation nodes (one per trigger),
4339 : * or NULL if no instrumentation is wanted.
4340 : * per_tuple_context: memory context to call trigger function in.
4341 : * trig_tuple_slot1: scratch slot for tg_trigtuple (foreign tables only)
4342 : * trig_tuple_slot2: scratch slot for tg_newtuple (foreign tables only)
4343 : * ----------
4344 : */
4345 : static void
4346 609614 : AfterTriggerExecute(EState *estate,
4347 : AfterTriggerEvent event,
4348 : ResultRelInfo *relInfo,
4349 : ResultRelInfo *src_relInfo,
4350 : ResultRelInfo *dst_relInfo,
4351 : TriggerDesc *trigdesc,
4352 : FmgrInfo *finfo, TriggerInstrumentation *instr,
4353 : MemoryContext per_tuple_context,
4354 : TupleTableSlot *trig_tuple_slot1,
4355 : TupleTableSlot *trig_tuple_slot2)
4356 : {
4357 609614 : Relation rel = relInfo->ri_RelationDesc;
4358 609614 : Relation src_rel = src_relInfo->ri_RelationDesc;
4359 609614 : Relation dst_rel = dst_relInfo->ri_RelationDesc;
4360 609614 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
4361 609614 : Oid tgoid = evtshared->ats_tgoid;
4362 609614 : TriggerData LocTriggerData = {0};
4363 : Oid save_rolid;
4364 : int save_sec_context;
4365 : HeapTuple rettuple;
4366 : int tgindx;
4367 609614 : bool should_free_trig = false;
4368 609614 : bool should_free_new = false;
4369 :
4370 : /*
4371 : * Locate trigger in trigdesc. It might not be present, and in fact the
4372 : * trigdesc could be NULL, if the trigger was dropped since the event was
4373 : * queued. In that case, silently do nothing.
4374 : */
4375 609614 : if (trigdesc == NULL)
4376 4 : return;
4377 1019450 : for (tgindx = 0; tgindx < trigdesc->numtriggers; tgindx++)
4378 : {
4379 1019450 : if (trigdesc->triggers[tgindx].tgoid == tgoid)
4380 : {
4381 609610 : LocTriggerData.tg_trigger = &(trigdesc->triggers[tgindx]);
4382 609610 : break;
4383 : }
4384 : }
4385 609610 : if (LocTriggerData.tg_trigger == NULL)
4386 0 : return;
4387 :
4388 : /*
4389 : * If doing EXPLAIN ANALYZE, start charging time to this trigger. We want
4390 : * to include time spent re-fetching tuples in the trigger cost.
4391 : */
4392 609610 : if (instr)
4393 0 : InstrStartTrigger(instr + tgindx);
4394 :
4395 : /*
4396 : * Fetch the required tuple(s).
4397 : */
4398 609610 : switch (event->ate_flags & AFTER_TRIGGER_TUP_BITS)
4399 : {
4400 25 : case AFTER_TRIGGER_FDW_FETCH:
4401 : {
4402 25 : Tuplestorestate *fdw_tuplestore = GetCurrentFDWTuplestore();
4403 :
4404 25 : if (!tuplestore_gettupleslot(fdw_tuplestore, true, false,
4405 : trig_tuple_slot1))
4406 0 : elog(ERROR, "failed to fetch tuple1 for AFTER trigger");
4407 :
4408 25 : if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) ==
4409 9 : TRIGGER_EVENT_UPDATE &&
4410 9 : !tuplestore_gettupleslot(fdw_tuplestore, true, false,
4411 : trig_tuple_slot2))
4412 0 : elog(ERROR, "failed to fetch tuple2 for AFTER trigger");
4413 : }
4414 : pg_fallthrough;
4415 : case AFTER_TRIGGER_FDW_REUSE:
4416 :
4417 : /*
4418 : * Store tuple in the slot so that tg_trigtuple does not reference
4419 : * tuplestore memory. (It is formally possible for the trigger
4420 : * function to queue trigger events that add to the same
4421 : * tuplestore, which can push other tuples out of memory.) The
4422 : * distinction is academic, because we start with a minimal tuple
4423 : * that is stored as a heap tuple, constructed in different memory
4424 : * context, in the slot anyway.
4425 : */
4426 29 : LocTriggerData.tg_trigslot = trig_tuple_slot1;
4427 29 : LocTriggerData.tg_trigtuple =
4428 29 : ExecFetchSlotHeapTuple(trig_tuple_slot1, true, &should_free_trig);
4429 :
4430 29 : if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) ==
4431 : TRIGGER_EVENT_UPDATE)
4432 : {
4433 11 : LocTriggerData.tg_newslot = trig_tuple_slot2;
4434 11 : LocTriggerData.tg_newtuple =
4435 11 : ExecFetchSlotHeapTuple(trig_tuple_slot2, true, &should_free_new);
4436 : }
4437 : else
4438 : {
4439 18 : LocTriggerData.tg_newtuple = NULL;
4440 : }
4441 29 : break;
4442 :
4443 609581 : default:
4444 609581 : if (ItemPointerIsValid(&(event->ate_ctid1)))
4445 : {
4446 608825 : TupleTableSlot *src_slot = ExecGetTriggerOldSlot(estate,
4447 : src_relInfo);
4448 :
4449 608825 : if (!table_tuple_fetch_row_version(src_rel,
4450 : &(event->ate_ctid1),
4451 : SnapshotAny,
4452 : src_slot))
4453 0 : elog(ERROR, "failed to fetch tuple1 for AFTER trigger");
4454 :
4455 : /*
4456 : * Store the tuple fetched from the source partition into the
4457 : * target (root partitioned) table slot, converting if needed.
4458 : */
4459 608825 : if (src_relInfo != relInfo)
4460 : {
4461 100 : TupleConversionMap *map = ExecGetChildToRootMap(src_relInfo);
4462 :
4463 100 : LocTriggerData.tg_trigslot = ExecGetTriggerOldSlot(estate, relInfo);
4464 100 : if (map)
4465 : {
4466 24 : execute_attr_map_slot(map->attrMap,
4467 : src_slot,
4468 : LocTriggerData.tg_trigslot);
4469 : }
4470 : else
4471 76 : ExecCopySlot(LocTriggerData.tg_trigslot, src_slot);
4472 : }
4473 : else
4474 608725 : LocTriggerData.tg_trigslot = src_slot;
4475 608825 : LocTriggerData.tg_trigtuple =
4476 608825 : ExecFetchSlotHeapTuple(LocTriggerData.tg_trigslot, false, &should_free_trig);
4477 : }
4478 : else
4479 : {
4480 756 : LocTriggerData.tg_trigtuple = NULL;
4481 : }
4482 :
4483 : /* don't touch ctid2 if not there */
4484 609581 : if (((event->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_2CTID ||
4485 609681 : (event->ate_flags & AFTER_TRIGGER_CP_UPDATE)) &&
4486 2094 : ItemPointerIsValid(&(event->ate_ctid2)))
4487 2094 : {
4488 2094 : TupleTableSlot *dst_slot = ExecGetTriggerNewSlot(estate,
4489 : dst_relInfo);
4490 :
4491 2094 : if (!table_tuple_fetch_row_version(dst_rel,
4492 : &(event->ate_ctid2),
4493 : SnapshotAny,
4494 : dst_slot))
4495 0 : elog(ERROR, "failed to fetch tuple2 for AFTER trigger");
4496 :
4497 : /*
4498 : * Store the tuple fetched from the destination partition into
4499 : * the target (root partitioned) table slot, converting if
4500 : * needed.
4501 : */
4502 2094 : if (dst_relInfo != relInfo)
4503 : {
4504 100 : TupleConversionMap *map = ExecGetChildToRootMap(dst_relInfo);
4505 :
4506 100 : LocTriggerData.tg_newslot = ExecGetTriggerNewSlot(estate, relInfo);
4507 100 : if (map)
4508 : {
4509 28 : execute_attr_map_slot(map->attrMap,
4510 : dst_slot,
4511 : LocTriggerData.tg_newslot);
4512 : }
4513 : else
4514 72 : ExecCopySlot(LocTriggerData.tg_newslot, dst_slot);
4515 : }
4516 : else
4517 1994 : LocTriggerData.tg_newslot = dst_slot;
4518 2094 : LocTriggerData.tg_newtuple =
4519 2094 : ExecFetchSlotHeapTuple(LocTriggerData.tg_newslot, false, &should_free_new);
4520 : }
4521 : else
4522 : {
4523 607487 : LocTriggerData.tg_newtuple = NULL;
4524 : }
4525 : }
4526 :
4527 : /*
4528 : * Set up the tuplestore information to let the trigger have access to
4529 : * transition tables. When we first make a transition table available to
4530 : * a trigger, mark it "closed" so that it cannot change anymore. If any
4531 : * additional events of the same type get queued in the current trigger
4532 : * query level, they'll go into new transition tables.
4533 : */
4534 609610 : LocTriggerData.tg_oldtable = LocTriggerData.tg_newtable = NULL;
4535 609610 : if (evtshared->ats_table)
4536 : {
4537 435 : if (LocTriggerData.tg_trigger->tgoldtable)
4538 : {
4539 231 : LocTriggerData.tg_oldtable = evtshared->ats_table->old_tuplestore;
4540 231 : evtshared->ats_table->closed = true;
4541 : }
4542 :
4543 435 : if (LocTriggerData.tg_trigger->tgnewtable)
4544 : {
4545 319 : LocTriggerData.tg_newtable = evtshared->ats_table->new_tuplestore;
4546 319 : evtshared->ats_table->closed = true;
4547 : }
4548 : }
4549 :
4550 : /*
4551 : * Setup the remaining trigger information
4552 : */
4553 609610 : LocTriggerData.type = T_TriggerData;
4554 609610 : LocTriggerData.tg_event =
4555 609610 : evtshared->ats_event & (TRIGGER_EVENT_OPMASK | TRIGGER_EVENT_ROW);
4556 609610 : LocTriggerData.tg_relation = rel;
4557 609610 : if (TRIGGER_FOR_UPDATE(LocTriggerData.tg_trigger->tgtype))
4558 3640 : LocTriggerData.tg_updatedcols = evtshared->ats_modifiedcols;
4559 :
4560 609610 : MemoryContextReset(per_tuple_context);
4561 :
4562 : /*
4563 : * If necessary, become the role that was active when the trigger got
4564 : * queued. Note that the role might have been dropped since the trigger
4565 : * was queued, but if that is a problem, we will get an error later.
4566 : * Checking here would still leave a race condition.
4567 : */
4568 609610 : GetUserIdAndSecContext(&save_rolid, &save_sec_context);
4569 609610 : if (save_rolid != evtshared->ats_rolid)
4570 16 : SetUserIdAndSecContext(evtshared->ats_rolid,
4571 : save_sec_context | SECURITY_LOCAL_USERID_CHANGE);
4572 :
4573 : /*
4574 : * Call the trigger and throw away any possibly returned updated tuple.
4575 : * (Don't let ExecCallTriggerFunc measure EXPLAIN time.)
4576 : */
4577 609610 : rettuple = ExecCallTriggerFunc(&LocTriggerData,
4578 : tgindx,
4579 : finfo,
4580 : NULL,
4581 : per_tuple_context);
4582 608996 : if (rettuple != NULL &&
4583 2361 : rettuple != LocTriggerData.tg_trigtuple &&
4584 979 : rettuple != LocTriggerData.tg_newtuple)
4585 0 : heap_freetuple(rettuple);
4586 :
4587 : /* Restore the current role if necessary */
4588 608996 : if (save_rolid != evtshared->ats_rolid)
4589 12 : SetUserIdAndSecContext(save_rolid, save_sec_context);
4590 :
4591 : /*
4592 : * Release resources
4593 : */
4594 608996 : if (should_free_trig)
4595 109 : heap_freetuple(LocTriggerData.tg_trigtuple);
4596 608996 : if (should_free_new)
4597 91 : heap_freetuple(LocTriggerData.tg_newtuple);
4598 :
4599 : /* don't clear slots' contents if foreign table */
4600 608996 : if (trig_tuple_slot1 == NULL)
4601 : {
4602 608961 : if (LocTriggerData.tg_trigslot)
4603 608243 : ExecClearTuple(LocTriggerData.tg_trigslot);
4604 608961 : if (LocTriggerData.tg_newslot)
4605 1897 : ExecClearTuple(LocTriggerData.tg_newslot);
4606 : }
4607 :
4608 : /*
4609 : * If doing EXPLAIN ANALYZE, stop charging time to this trigger, and count
4610 : * the firing of the trigger.
4611 : */
4612 608996 : if (instr)
4613 0 : InstrStopTrigger(instr + tgindx, 1);
4614 : }
4615 :
4616 :
4617 : /*
4618 : * afterTriggerMarkEvents()
4619 : *
4620 : * Scan the given event list for not yet invoked events. Mark the ones
4621 : * that can be invoked now with the current firing ID.
4622 : *
4623 : * If move_list isn't NULL, events that are not to be invoked now are
4624 : * transferred to move_list.
4625 : *
4626 : * When immediate_only is true, do not invoke currently-deferred triggers.
4627 : * (This will be false only at main transaction exit.)
4628 : *
4629 : * Returns true if any invokable events were found.
4630 : */
4631 : static bool
4632 605833 : afterTriggerMarkEvents(AfterTriggerEventList *events,
4633 : AfterTriggerEventList *move_list,
4634 : bool immediate_only)
4635 : {
4636 605833 : bool found = false;
4637 605833 : bool deferred_found = false;
4638 : AfterTriggerEvent event;
4639 : AfterTriggerEventChunk *chunk;
4640 :
4641 1222159 : for_each_event_chunk(event, chunk, *events)
4642 : {
4643 610751 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
4644 610751 : bool defer_it = false;
4645 :
4646 610751 : if (!(event->ate_flags &
4647 : (AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS)))
4648 : {
4649 : /*
4650 : * This trigger hasn't been called or scheduled yet. Check if we
4651 : * should call it now.
4652 : */
4653 610175 : if (immediate_only && afterTriggerCheckState(evtshared))
4654 : {
4655 451 : defer_it = true;
4656 : }
4657 : else
4658 : {
4659 : /*
4660 : * Mark it as to be fired in this firing cycle.
4661 : */
4662 609724 : evtshared->ats_firing_id = afterTriggers.firing_counter;
4663 609724 : event->ate_flags |= AFTER_TRIGGER_IN_PROGRESS;
4664 609724 : found = true;
4665 : }
4666 : }
4667 :
4668 : /*
4669 : * If it's deferred, move it to move_list, if requested.
4670 : */
4671 610751 : if (defer_it && move_list != NULL)
4672 : {
4673 451 : deferred_found = true;
4674 : /* add it to move_list */
4675 451 : afterTriggerAddEvent(move_list, event, evtshared);
4676 : /* mark original copy "done" so we don't do it again */
4677 451 : event->ate_flags |= AFTER_TRIGGER_DONE;
4678 : }
4679 : }
4680 :
4681 : /*
4682 : * We could allow deferred triggers if, before the end of the
4683 : * security-restricted operation, we were to verify that a SET CONSTRAINTS
4684 : * ... IMMEDIATE has fired all such triggers. For now, don't bother.
4685 : */
4686 605833 : if (deferred_found && InSecurityRestrictedOperation())
4687 8 : ereport(ERROR,
4688 : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
4689 : errmsg("cannot fire deferred trigger within security-restricted operation")));
4690 :
4691 605825 : return found;
4692 : }
4693 :
4694 : /*
4695 : * afterTriggerInvokeEvents()
4696 : *
4697 : * Scan the given event list for events that are marked as to be fired
4698 : * in the current firing cycle, and fire them.
4699 : *
4700 : * If estate isn't NULL, we use its result relation info to avoid repeated
4701 : * openings and closing of trigger target relations. If it is NULL, we
4702 : * make one locally to cache the info in case there are multiple trigger
4703 : * events per rel.
4704 : *
4705 : * When delete_ok is true, it's safe to delete fully-processed events.
4706 : * (We are not very tense about that: we simply reset a chunk to be empty
4707 : * if all its events got fired. The objective here is just to avoid useless
4708 : * rescanning of events when a trigger queues new events during transaction
4709 : * end, so it's not necessary to worry much about the case where only
4710 : * some events are fired.)
4711 : *
4712 : * Returns true if no unfired events remain in the list (this allows us
4713 : * to avoid repeating afterTriggerMarkEvents).
4714 : */
4715 : static bool
4716 5174 : afterTriggerInvokeEvents(AfterTriggerEventList *events,
4717 : CommandId firing_id,
4718 : EState *estate,
4719 : bool delete_ok)
4720 : {
4721 5174 : bool all_fired = true;
4722 : AfterTriggerEventChunk *chunk;
4723 : MemoryContext per_tuple_context;
4724 5174 : bool local_estate = false;
4725 5174 : ResultRelInfo *rInfo = NULL;
4726 5174 : Relation rel = NULL;
4727 5174 : TriggerDesc *trigdesc = NULL;
4728 5174 : FmgrInfo *finfo = NULL;
4729 5174 : TriggerInstrumentation *instr = NULL;
4730 5174 : TupleTableSlot *slot1 = NULL,
4731 5174 : *slot2 = NULL;
4732 :
4733 : /* Make a local EState if need be */
4734 5174 : if (estate == NULL)
4735 : {
4736 252 : estate = CreateExecutorState();
4737 252 : local_estate = true;
4738 : }
4739 :
4740 : /* Make a per-tuple memory context for trigger function calls */
4741 : per_tuple_context =
4742 5174 : AllocSetContextCreate(CurrentMemoryContext,
4743 : "AfterTriggerTupleContext",
4744 : ALLOCSET_DEFAULT_SIZES);
4745 :
4746 9788 : for_each_chunk(chunk, *events)
4747 : {
4748 : AfterTriggerEvent event;
4749 5228 : bool all_fired_in_chunk = true;
4750 :
4751 615240 : for_each_event(event, chunk)
4752 : {
4753 610626 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
4754 :
4755 : /*
4756 : * Is it one for me to fire?
4757 : */
4758 610626 : if ((event->ate_flags & AFTER_TRIGGER_IN_PROGRESS) &&
4759 609614 : evtshared->ats_firing_id == firing_id)
4760 609000 : {
4761 : ResultRelInfo *src_rInfo,
4762 : *dst_rInfo;
4763 :
4764 : /*
4765 : * So let's fire it... but first, find the correct relation if
4766 : * this is not the same relation as before.
4767 : */
4768 609614 : if (rel == NULL || RelationGetRelid(rel) != evtshared->ats_relid)
4769 : {
4770 5399 : rInfo = ExecGetTriggerResultRel(estate, evtshared->ats_relid,
4771 : NULL);
4772 5399 : rel = rInfo->ri_RelationDesc;
4773 : /* Catch calls with insufficient relcache refcounting */
4774 : Assert(!RelationHasReferenceCountZero(rel));
4775 5399 : trigdesc = rInfo->ri_TrigDesc;
4776 : /* caution: trigdesc could be NULL here */
4777 5399 : finfo = rInfo->ri_TrigFunctions;
4778 5399 : instr = rInfo->ri_TrigInstrument;
4779 5399 : if (slot1 != NULL)
4780 : {
4781 0 : ExecDropSingleTupleTableSlot(slot1);
4782 0 : ExecDropSingleTupleTableSlot(slot2);
4783 0 : slot1 = slot2 = NULL;
4784 : }
4785 5399 : if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
4786 : {
4787 19 : slot1 = MakeSingleTupleTableSlot(rel->rd_att,
4788 : &TTSOpsMinimalTuple);
4789 19 : slot2 = MakeSingleTupleTableSlot(rel->rd_att,
4790 : &TTSOpsMinimalTuple);
4791 : }
4792 : }
4793 :
4794 : /*
4795 : * Look up source and destination partition result rels of a
4796 : * cross-partition update event.
4797 : */
4798 609614 : if ((event->ate_flags & AFTER_TRIGGER_TUP_BITS) ==
4799 : AFTER_TRIGGER_CP_UPDATE)
4800 : {
4801 : Assert(OidIsValid(event->ate_src_part) &&
4802 : OidIsValid(event->ate_dst_part));
4803 100 : src_rInfo = ExecGetTriggerResultRel(estate,
4804 : event->ate_src_part,
4805 : rInfo);
4806 100 : dst_rInfo = ExecGetTriggerResultRel(estate,
4807 : event->ate_dst_part,
4808 : rInfo);
4809 : }
4810 : else
4811 609514 : src_rInfo = dst_rInfo = rInfo;
4812 :
4813 : /*
4814 : * Fire it. Note that the AFTER_TRIGGER_IN_PROGRESS flag is
4815 : * still set, so recursive examinations of the event list
4816 : * won't try to re-fire it.
4817 : */
4818 609614 : AfterTriggerExecute(estate, event, rInfo,
4819 : src_rInfo, dst_rInfo,
4820 : trigdesc, finfo, instr,
4821 : per_tuple_context, slot1, slot2);
4822 :
4823 : /*
4824 : * Mark the event as done.
4825 : */
4826 609000 : event->ate_flags &= ~AFTER_TRIGGER_IN_PROGRESS;
4827 609000 : event->ate_flags |= AFTER_TRIGGER_DONE;
4828 : }
4829 1012 : else if (!(event->ate_flags & AFTER_TRIGGER_DONE))
4830 : {
4831 : /* something remains to be done */
4832 352 : all_fired = all_fired_in_chunk = false;
4833 : }
4834 : }
4835 :
4836 : /* Clear the chunk if delete_ok and nothing left of interest */
4837 4614 : if (delete_ok && all_fired_in_chunk)
4838 : {
4839 177 : chunk->freeptr = CHUNK_DATA_START(chunk);
4840 177 : chunk->endfree = chunk->endptr;
4841 :
4842 : /*
4843 : * If it's last chunk, must sync event list's tailfree too. Note
4844 : * that delete_ok must NOT be passed as true if there could be
4845 : * additional AfterTriggerEventList values pointing at this event
4846 : * list, since we'd fail to fix their copies of tailfree.
4847 : */
4848 177 : if (chunk == events->tail)
4849 177 : events->tailfree = chunk->freeptr;
4850 : }
4851 : }
4852 4560 : if (slot1 != NULL)
4853 : {
4854 19 : ExecDropSingleTupleTableSlot(slot1);
4855 19 : ExecDropSingleTupleTableSlot(slot2);
4856 : }
4857 :
4858 : /* Release working resources */
4859 4560 : MemoryContextDelete(per_tuple_context);
4860 :
4861 4560 : if (local_estate)
4862 : {
4863 179 : ExecCloseResultRelations(estate);
4864 179 : ExecResetTupleTable(estate->es_tupleTable, false);
4865 179 : FreeExecutorState(estate);
4866 : }
4867 :
4868 4560 : return all_fired;
4869 : }
4870 :
4871 :
4872 : /*
4873 : * GetAfterTriggersTableData
4874 : *
4875 : * Find or create an AfterTriggersTableData struct for the specified
4876 : * trigger event (relation + operation type). Ignore existing structs
4877 : * marked "closed"; we don't want to put any additional tuples into them,
4878 : * nor change their stmt-triggers-fired state.
4879 : *
4880 : * Note: the AfterTriggersTableData list is allocated in the current
4881 : * (sub)transaction's CurTransactionContext. This is OK because
4882 : * we don't need it to live past AfterTriggerEndQuery.
4883 : */
4884 : static AfterTriggersTableData *
4885 1633 : GetAfterTriggersTableData(Oid relid, CmdType cmdType)
4886 : {
4887 : AfterTriggersTableData *table;
4888 : AfterTriggersQueryData *qs;
4889 : MemoryContext oldcxt;
4890 : ListCell *lc;
4891 :
4892 : /* At this level, cmdType should not be, eg, CMD_MERGE */
4893 : Assert(cmdType == CMD_INSERT ||
4894 : cmdType == CMD_UPDATE ||
4895 : cmdType == CMD_DELETE);
4896 :
4897 : /* Caller should have ensured query_depth is OK. */
4898 : Assert(afterTriggers.query_depth >= 0 &&
4899 : afterTriggers.query_depth < afterTriggers.maxquerydepth);
4900 1633 : qs = &afterTriggers.query_stack[afterTriggers.query_depth];
4901 :
4902 1873 : foreach(lc, qs->tables)
4903 : {
4904 1084 : table = (AfterTriggersTableData *) lfirst(lc);
4905 1084 : if (table->relid == relid && table->cmdType == cmdType &&
4906 868 : !table->closed)
4907 844 : return table;
4908 : }
4909 :
4910 789 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4911 :
4912 789 : table = palloc0_object(AfterTriggersTableData);
4913 789 : table->relid = relid;
4914 789 : table->cmdType = cmdType;
4915 789 : qs->tables = lappend(qs->tables, table);
4916 :
4917 789 : MemoryContextSwitchTo(oldcxt);
4918 :
4919 789 : return table;
4920 : }
4921 :
4922 : /*
4923 : * Returns a TupleTableSlot suitable for holding the tuples to be put
4924 : * into AfterTriggersTableData's transition table tuplestores.
4925 : */
4926 : static TupleTableSlot *
4927 196 : GetAfterTriggersStoreSlot(AfterTriggersTableData *table,
4928 : TupleDesc tupdesc)
4929 : {
4930 : /* Create it if not already done. */
4931 196 : if (!table->storeslot)
4932 : {
4933 : MemoryContext oldcxt;
4934 :
4935 : /*
4936 : * We need this slot only until AfterTriggerEndQuery, but making it
4937 : * last till end-of-subxact is good enough. It'll be freed by
4938 : * AfterTriggerFreeQuery(). However, the passed-in tupdesc might have
4939 : * a different lifespan, so we'd better make a copy of that.
4940 : */
4941 56 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4942 56 : tupdesc = CreateTupleDescCopy(tupdesc);
4943 56 : table->storeslot = MakeSingleTupleTableSlot(tupdesc, &TTSOpsVirtual);
4944 56 : MemoryContextSwitchTo(oldcxt);
4945 : }
4946 :
4947 196 : return table->storeslot;
4948 : }
4949 :
4950 : /*
4951 : * MakeTransitionCaptureState
4952 : *
4953 : * Make a TransitionCaptureState object for the given TriggerDesc, target
4954 : * relation, and operation type. The TCS object holds all the state needed
4955 : * to decide whether to capture tuples in transition tables.
4956 : *
4957 : * If there are no triggers in 'trigdesc' that request relevant transition
4958 : * tables, then return NULL.
4959 : *
4960 : * The resulting object can be passed to the ExecAR* functions. When
4961 : * dealing with child tables, the caller can set tcs_original_insert_tuple
4962 : * to avoid having to reconstruct the original tuple in the root table's
4963 : * format.
4964 : *
4965 : * Note that we copy the flags from a parent table into this struct (rather
4966 : * than subsequently using the relation's TriggerDesc directly) so that we can
4967 : * use it to control collection of transition tuples from child tables.
4968 : *
4969 : * Per SQL spec, all operations of the same kind (INSERT/UPDATE/DELETE)
4970 : * on the same table during one query should share one transition table.
4971 : * Therefore, the Tuplestores are owned by an AfterTriggersTableData struct
4972 : * looked up using the table OID + CmdType, and are merely referenced by
4973 : * the TransitionCaptureState objects we hand out to callers.
4974 : */
4975 : TransitionCaptureState *
4976 74598 : MakeTransitionCaptureState(TriggerDesc *trigdesc, Oid relid, CmdType cmdType)
4977 : {
4978 : TransitionCaptureState *state;
4979 : bool need_old_upd,
4980 : need_new_upd,
4981 : need_old_del,
4982 : need_new_ins;
4983 : AfterTriggersTableData *ins_table;
4984 : AfterTriggersTableData *upd_table;
4985 : AfterTriggersTableData *del_table;
4986 : MemoryContext oldcxt;
4987 : ResourceOwner saveResourceOwner;
4988 :
4989 74598 : if (trigdesc == NULL)
4990 65877 : return NULL;
4991 :
4992 : /* Detect which table(s) we need. */
4993 8721 : switch (cmdType)
4994 : {
4995 4900 : case CMD_INSERT:
4996 4900 : need_old_upd = need_old_del = need_new_upd = false;
4997 4900 : need_new_ins = trigdesc->trig_insert_new_table;
4998 4900 : break;
4999 2581 : case CMD_UPDATE:
5000 2581 : need_old_upd = trigdesc->trig_update_old_table;
5001 2581 : need_new_upd = trigdesc->trig_update_new_table;
5002 2581 : need_old_del = need_new_ins = false;
5003 2581 : break;
5004 994 : case CMD_DELETE:
5005 994 : need_old_del = trigdesc->trig_delete_old_table;
5006 994 : need_old_upd = need_new_upd = need_new_ins = false;
5007 994 : break;
5008 246 : case CMD_MERGE:
5009 246 : need_old_upd = trigdesc->trig_update_old_table;
5010 246 : need_new_upd = trigdesc->trig_update_new_table;
5011 246 : need_old_del = trigdesc->trig_delete_old_table;
5012 246 : need_new_ins = trigdesc->trig_insert_new_table;
5013 246 : break;
5014 0 : default:
5015 0 : elog(ERROR, "unexpected CmdType: %d", (int) cmdType);
5016 : /* keep compiler quiet */
5017 : need_old_upd = need_new_upd = need_old_del = need_new_ins = false;
5018 : break;
5019 : }
5020 8721 : if (!need_old_upd && !need_new_upd && !need_new_ins && !need_old_del)
5021 8290 : return NULL;
5022 :
5023 : /* Check state, like AfterTriggerSaveEvent. */
5024 431 : if (afterTriggers.query_depth < 0)
5025 0 : elog(ERROR, "MakeTransitionCaptureState() called outside of query");
5026 :
5027 : /* Be sure we have enough space to record events at this query depth. */
5028 431 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
5029 315 : AfterTriggerEnlargeQueryState();
5030 :
5031 : /*
5032 : * Find or create AfterTriggersTableData struct(s) to hold the
5033 : * tuplestore(s). If there's a matching struct but it's marked closed,
5034 : * ignore it; we need a newer one.
5035 : *
5036 : * Note: MERGE must use the same AfterTriggersTableData structs as INSERT,
5037 : * UPDATE, and DELETE, so that any MERGE'd tuples are added to the same
5038 : * tuplestores as tuples from any INSERT, UPDATE, or DELETE commands
5039 : * running in the same top-level command (e.g., in a writable CTE).
5040 : *
5041 : * Note: the AfterTriggersTableData list, as well as the tuplestores, are
5042 : * allocated in the current (sub)transaction's CurTransactionContext, and
5043 : * the tuplestores are managed by the (sub)transaction's resource owner.
5044 : * This is sufficient lifespan because we do not allow triggers using
5045 : * transition tables to be deferrable; they will be fired during
5046 : * AfterTriggerEndQuery, after which it's okay to delete the data.
5047 : */
5048 431 : if (need_new_ins)
5049 180 : ins_table = GetAfterTriggersTableData(relid, CMD_INSERT);
5050 : else
5051 251 : ins_table = NULL;
5052 :
5053 431 : if (need_old_upd || need_new_upd)
5054 145 : upd_table = GetAfterTriggersTableData(relid, CMD_UPDATE);
5055 : else
5056 286 : upd_table = NULL;
5057 :
5058 431 : if (need_old_del)
5059 130 : del_table = GetAfterTriggersTableData(relid, CMD_DELETE);
5060 : else
5061 301 : del_table = NULL;
5062 :
5063 : /* Now create required tuplestore(s), if we don't have them already. */
5064 431 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
5065 431 : saveResourceOwner = CurrentResourceOwner;
5066 431 : CurrentResourceOwner = CurTransactionResourceOwner;
5067 :
5068 431 : if (need_old_upd && upd_table->old_tuplestore == NULL)
5069 121 : upd_table->old_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5070 431 : if (need_new_upd && upd_table->new_tuplestore == NULL)
5071 129 : upd_table->new_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5072 431 : if (need_old_del && del_table->old_tuplestore == NULL)
5073 98 : del_table->old_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5074 431 : if (need_new_ins && ins_table->new_tuplestore == NULL)
5075 172 : ins_table->new_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5076 :
5077 431 : CurrentResourceOwner = saveResourceOwner;
5078 431 : MemoryContextSwitchTo(oldcxt);
5079 :
5080 : /* Now build the TransitionCaptureState struct, in caller's context */
5081 431 : state = palloc0_object(TransitionCaptureState);
5082 431 : state->tcs_delete_old_table = need_old_del;
5083 431 : state->tcs_update_old_table = need_old_upd;
5084 431 : state->tcs_update_new_table = need_new_upd;
5085 431 : state->tcs_insert_new_table = need_new_ins;
5086 431 : state->tcs_insert_private = ins_table;
5087 431 : state->tcs_update_private = upd_table;
5088 431 : state->tcs_delete_private = del_table;
5089 :
5090 431 : return state;
5091 : }
5092 :
5093 :
5094 : /* ----------
5095 : * AfterTriggerBeginXact()
5096 : *
5097 : * Called at transaction start (either BEGIN or implicit for single
5098 : * statement outside of transaction block).
5099 : * ----------
5100 : */
5101 : void
5102 628809 : AfterTriggerBeginXact(void)
5103 : {
5104 : /*
5105 : * Initialize after-trigger state structure to empty
5106 : */
5107 628809 : afterTriggers.firing_counter = (CommandId) 1; /* mustn't be 0 */
5108 628809 : afterTriggers.query_depth = -1;
5109 628809 : afterTriggers.batch_callbacks = NIL;
5110 :
5111 : /*
5112 : * Verify that there is no leftover state remaining. If these assertions
5113 : * trip, it means that AfterTriggerEndXact wasn't called or didn't clean
5114 : * up properly.
5115 : */
5116 : Assert(afterTriggers.state == NULL);
5117 : Assert(afterTriggers.query_stack == NULL);
5118 : Assert(afterTriggers.maxquerydepth == 0);
5119 : Assert(afterTriggers.event_cxt == NULL);
5120 : Assert(afterTriggers.events.head == NULL);
5121 : Assert(afterTriggers.trans_stack == NULL);
5122 : Assert(afterTriggers.maxtransdepth == 0);
5123 : Assert(afterTriggerFiringDepth == 0);
5124 628809 : }
5125 :
5126 :
5127 : /* ----------
5128 : * AfterTriggerBeginQuery()
5129 : *
5130 : * Called just before we start processing a single query within a
5131 : * transaction (or subtransaction). Most of the real work gets deferred
5132 : * until somebody actually tries to queue a trigger event.
5133 : * ----------
5134 : */
5135 : void
5136 243891 : AfterTriggerBeginQuery(void)
5137 : {
5138 : /* Increase the query stack depth */
5139 243891 : afterTriggers.query_depth++;
5140 243891 : }
5141 :
5142 :
5143 : /* ----------
5144 : * AfterTriggerEndQuery()
5145 : *
5146 : * Called after one query has been completely processed. At this time
5147 : * we invoke all AFTER IMMEDIATE trigger events queued by the query, and
5148 : * transfer deferred trigger events to the global deferred-trigger list.
5149 : *
5150 : * Note that this must be called BEFORE closing down the executor
5151 : * with ExecutorEnd, because we make use of the EState's info about
5152 : * target relations. Normally it is called from ExecutorFinish.
5153 : * ----------
5154 : */
5155 : void
5156 240685 : AfterTriggerEndQuery(EState *estate)
5157 : {
5158 : AfterTriggersQueryData *qs;
5159 :
5160 : /* Must be inside a query, too */
5161 : Assert(afterTriggers.query_depth >= 0);
5162 :
5163 : /*
5164 : * If we never even got as far as initializing the event stack, there
5165 : * certainly won't be any events, so exit quickly.
5166 : */
5167 240685 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
5168 : {
5169 234714 : afterTriggers.query_depth--;
5170 234714 : return;
5171 : }
5172 :
5173 : /*
5174 : * Process all immediate-mode triggers queued by the query, and move the
5175 : * deferred ones to the main list of deferred events.
5176 : *
5177 : * Notice that we decide which ones will be fired, and put the deferred
5178 : * ones on the main list, before anything is actually fired. This ensures
5179 : * reasonably sane behavior if a trigger function does SET CONSTRAINTS ...
5180 : * IMMEDIATE: all events we have decided to defer will be available for it
5181 : * to fire.
5182 : *
5183 : * We loop in case a trigger queues more events at the same query level.
5184 : * Ordinary trigger functions, including all PL/pgSQL trigger functions,
5185 : * will instead fire any triggers in a dedicated query level. Foreign key
5186 : * enforcement triggers do add to the current query level, thanks to their
5187 : * passing fire_triggers = false to SPI_execute_snapshot(). Other
5188 : * C-language triggers might do likewise.
5189 : *
5190 : * If we find no firable events, we don't have to increment
5191 : * firing_counter.
5192 : */
5193 5971 : qs = &afterTriggers.query_stack[afterTriggers.query_depth];
5194 :
5195 5971 : afterTriggerFiringDepth++;
5196 : for (;;)
5197 : {
5198 6179 : if (afterTriggerMarkEvents(&qs->events, &afterTriggers.events, true))
5199 : {
5200 4922 : CommandId firing_id = afterTriggers.firing_counter++;
5201 4922 : AfterTriggerEventChunk *oldtail = qs->events.tail;
5202 :
5203 4922 : if (afterTriggerInvokeEvents(&qs->events, firing_id, estate, false))
5204 4173 : break; /* all fired */
5205 :
5206 : /*
5207 : * Firing a trigger could result in query_stack being repalloc'd,
5208 : * so we must recalculate qs after each afterTriggerInvokeEvents
5209 : * call. Furthermore, it's unsafe to pass delete_ok = true here,
5210 : * because that could cause afterTriggerInvokeEvents to try to
5211 : * access qs->events after the stack has been repalloc'd.
5212 : */
5213 208 : qs = &afterTriggers.query_stack[afterTriggers.query_depth];
5214 :
5215 : /*
5216 : * We'll need to scan the events list again. To reduce the cost
5217 : * of doing so, get rid of completely-fired chunks. We know that
5218 : * all events were marked IN_PROGRESS or DONE at the conclusion of
5219 : * afterTriggerMarkEvents, so any still-interesting events must
5220 : * have been added after that, and so must be in the chunk that
5221 : * was then the tail chunk, or in later chunks. So, zap all
5222 : * chunks before oldtail. This is approximately the same set of
5223 : * events we would have gotten rid of by passing delete_ok = true.
5224 : */
5225 : Assert(oldtail != NULL);
5226 208 : while (qs->events.head != oldtail)
5227 0 : afterTriggerDeleteHeadEventChunk(qs);
5228 : }
5229 : else
5230 1249 : break;
5231 : }
5232 :
5233 : /*
5234 : * Fire batch callbacks before releasing query-level storage and before
5235 : * decrementing query_depth. Callbacks may do real work (index probes,
5236 : * error reporting) and rely on query_depth still reflecting the current
5237 : * batch level so that nested calls from SPI inside AFTER triggers are
5238 : * correctly suppressed by FireAfterTriggerBatchCallbacks's depth guard.
5239 : */
5240 5422 : FireAfterTriggerBatchCallbacks();
5241 :
5242 : /* Release query-level-local storage, including tuplestores if any */
5243 5183 : AfterTriggerFreeQuery(&afterTriggers.query_stack[afterTriggers.query_depth]);
5244 :
5245 5183 : afterTriggers.query_depth--;
5246 5183 : afterTriggerFiringDepth--;
5247 : }
5248 :
5249 :
5250 : /*
5251 : * AfterTriggerFreeQuery
5252 : * Release subsidiary storage for a trigger query level.
5253 : * This includes closing down tuplestores.
5254 : * Note: it's important for this to be safe if interrupted by an error
5255 : * and then called again for the same query level.
5256 : */
5257 : static void
5258 5200 : AfterTriggerFreeQuery(AfterTriggersQueryData *qs)
5259 : {
5260 : Tuplestorestate *ts;
5261 : List *tables;
5262 : ListCell *lc;
5263 :
5264 : /* Drop the trigger events */
5265 5200 : afterTriggerFreeEventList(&qs->events);
5266 :
5267 : /* Drop FDW tuplestore if any */
5268 5200 : ts = qs->fdw_tuplestore;
5269 5200 : qs->fdw_tuplestore = NULL;
5270 5200 : if (ts)
5271 18 : tuplestore_end(ts);
5272 :
5273 : /* Release per-table subsidiary storage */
5274 5200 : tables = qs->tables;
5275 5943 : foreach(lc, tables)
5276 : {
5277 743 : AfterTriggersTableData *table = (AfterTriggersTableData *) lfirst(lc);
5278 :
5279 743 : ts = table->old_tuplestore;
5280 743 : table->old_tuplestore = NULL;
5281 743 : if (ts)
5282 199 : tuplestore_end(ts);
5283 743 : ts = table->new_tuplestore;
5284 743 : table->new_tuplestore = NULL;
5285 743 : if (ts)
5286 275 : tuplestore_end(ts);
5287 743 : if (table->storeslot)
5288 : {
5289 56 : TupleTableSlot *slot = table->storeslot;
5290 :
5291 56 : table->storeslot = NULL;
5292 56 : ExecDropSingleTupleTableSlot(slot);
5293 : }
5294 : }
5295 :
5296 : /*
5297 : * Now free the AfterTriggersTableData structs and list cells. Reset list
5298 : * pointer first; if list_free_deep somehow gets an error, better to leak
5299 : * that storage than have an infinite loop.
5300 : */
5301 5200 : qs->tables = NIL;
5302 5200 : list_free_deep(tables);
5303 5200 : }
5304 :
5305 :
5306 : /* ----------
5307 : * AfterTriggerFireDeferred()
5308 : *
5309 : * Called just before the current transaction is committed. At this
5310 : * time we invoke all pending DEFERRED triggers.
5311 : *
5312 : * It is possible for other modules to queue additional deferred triggers
5313 : * during pre-commit processing; therefore xact.c may have to call this
5314 : * multiple times.
5315 : * ----------
5316 : */
5317 : void
5318 599628 : AfterTriggerFireDeferred(void)
5319 : {
5320 : AfterTriggerEventList *events;
5321 599628 : bool snap_pushed = false;
5322 :
5323 : /* Must not be inside a query */
5324 : Assert(afterTriggers.query_depth == -1);
5325 :
5326 : /*
5327 : * If there are any triggers to fire, make sure we have set a snapshot for
5328 : * them to use. (Since PortalRunUtility doesn't set a snap for COMMIT, we
5329 : * can't assume ActiveSnapshot is valid on entry.)
5330 : */
5331 599628 : events = &afterTriggers.events;
5332 599628 : if (events->head != NULL)
5333 : {
5334 238 : PushActiveSnapshot(GetTransactionSnapshot());
5335 238 : snap_pushed = true;
5336 : }
5337 :
5338 : /*
5339 : * Run all the remaining triggers. Loop until they are all gone, in case
5340 : * some trigger queues more for us to do.
5341 : */
5342 599628 : afterTriggerFiringDepth++;
5343 599628 : while (afterTriggerMarkEvents(events, NULL, false))
5344 : {
5345 238 : CommandId firing_id = afterTriggers.firing_counter++;
5346 :
5347 238 : if (afterTriggerInvokeEvents(events, firing_id, NULL, true))
5348 169 : break; /* all fired */
5349 : }
5350 :
5351 : /* Flush any fast-path batches accumulated by the triggers just fired. */
5352 599559 : FireAfterTriggerBatchCallbacks();
5353 :
5354 599520 : afterTriggerFiringDepth--;
5355 :
5356 : /*
5357 : * We don't bother freeing the event list, since it will go away anyway
5358 : * (and more efficiently than via pfree) in AfterTriggerEndXact.
5359 : */
5360 :
5361 599520 : if (snap_pushed)
5362 130 : PopActiveSnapshot();
5363 599520 : }
5364 :
5365 :
5366 : /* ----------
5367 : * AfterTriggerEndXact()
5368 : *
5369 : * The current transaction is finishing.
5370 : *
5371 : * Any unfired triggers are canceled so we simply throw
5372 : * away anything we know.
5373 : *
5374 : * Note: it is possible for this to be called repeatedly in case of
5375 : * error during transaction abort; therefore, do not complain if
5376 : * already closed down.
5377 : * ----------
5378 : */
5379 : void
5380 629034 : AfterTriggerEndXact(bool isCommit)
5381 : {
5382 : /*
5383 : * Forget the pending-events list.
5384 : *
5385 : * Since all the info is in TopTransactionContext or children thereof, we
5386 : * don't really need to do anything to reclaim memory. However, the
5387 : * pending-events list could be large, and so it's useful to discard it as
5388 : * soon as possible --- especially if we are aborting because we ran out
5389 : * of memory for the list!
5390 : */
5391 629034 : if (afterTriggers.event_cxt)
5392 : {
5393 4455 : MemoryContextDelete(afterTriggers.event_cxt);
5394 4455 : afterTriggers.event_cxt = NULL;
5395 4455 : afterTriggers.events.head = NULL;
5396 4455 : afterTriggers.events.tail = NULL;
5397 4455 : afterTriggers.events.tailfree = NULL;
5398 : }
5399 :
5400 : /*
5401 : * Forget any subtransaction state as well. Since this can't be very
5402 : * large, we let the eventual reset of TopTransactionContext free the
5403 : * memory instead of doing it here.
5404 : */
5405 629034 : afterTriggers.trans_stack = NULL;
5406 629034 : afterTriggers.maxtransdepth = 0;
5407 :
5408 :
5409 : /*
5410 : * Forget the query stack and constraint-related state information. As
5411 : * with the subtransaction state information, we don't bother freeing the
5412 : * memory here.
5413 : */
5414 629034 : afterTriggers.query_stack = NULL;
5415 629034 : afterTriggers.maxquerydepth = 0;
5416 629034 : afterTriggers.state = NULL;
5417 :
5418 : /* No more afterTriggers manipulation until next transaction starts. */
5419 629034 : afterTriggers.query_depth = -1;
5420 :
5421 629034 : afterTriggerFiringDepth = 0;
5422 629034 : }
5423 :
5424 : /*
5425 : * AfterTriggerBeginSubXact()
5426 : *
5427 : * Start a subtransaction.
5428 : */
5429 : void
5430 11887 : AfterTriggerBeginSubXact(void)
5431 : {
5432 11887 : int my_level = GetCurrentTransactionNestLevel();
5433 :
5434 : /*
5435 : * Allocate more space in the trans_stack if needed. (Note: because the
5436 : * minimum nest level of a subtransaction is 2, we waste the first couple
5437 : * entries of the array; not worth the notational effort to avoid it.)
5438 : */
5439 13455 : while (my_level >= afterTriggers.maxtransdepth)
5440 : {
5441 1568 : if (afterTriggers.maxtransdepth == 0)
5442 : {
5443 : /* Arbitrarily initialize for max of 8 subtransaction levels */
5444 1524 : afterTriggers.trans_stack = (AfterTriggersTransData *)
5445 1524 : MemoryContextAlloc(TopTransactionContext,
5446 : 8 * sizeof(AfterTriggersTransData));
5447 1524 : afterTriggers.maxtransdepth = 8;
5448 : }
5449 : else
5450 : {
5451 : /* repalloc will keep the stack in the same context */
5452 44 : int new_alloc = afterTriggers.maxtransdepth * 2;
5453 :
5454 44 : afterTriggers.trans_stack = (AfterTriggersTransData *)
5455 44 : repalloc(afterTriggers.trans_stack,
5456 : new_alloc * sizeof(AfterTriggersTransData));
5457 44 : afterTriggers.maxtransdepth = new_alloc;
5458 : }
5459 : }
5460 :
5461 : /*
5462 : * Push the current information into the stack. The SET CONSTRAINTS state
5463 : * is not saved until/unless changed. Likewise, we don't make a
5464 : * per-subtransaction event context until needed.
5465 : */
5466 11887 : afterTriggers.trans_stack[my_level].state = NULL;
5467 11887 : afterTriggers.trans_stack[my_level].events = afterTriggers.events;
5468 11887 : afterTriggers.trans_stack[my_level].query_depth = afterTriggers.query_depth;
5469 11887 : afterTriggers.trans_stack[my_level].firing_counter = afterTriggers.firing_counter;
5470 11887 : }
5471 :
5472 : /*
5473 : * AfterTriggerEndSubXact()
5474 : *
5475 : * The current subtransaction is ending.
5476 : */
5477 : void
5478 11887 : AfterTriggerEndSubXact(bool isCommit)
5479 : {
5480 11887 : int my_level = GetCurrentTransactionNestLevel();
5481 : SetConstraintState state;
5482 : AfterTriggerEvent event;
5483 : AfterTriggerEventChunk *chunk;
5484 : CommandId subxact_firing_id;
5485 :
5486 : /*
5487 : * Pop the prior state if needed.
5488 : */
5489 11887 : if (isCommit)
5490 : {
5491 : Assert(my_level < afterTriggers.maxtransdepth);
5492 : /* If we saved a prior state, we don't need it anymore */
5493 5880 : state = afterTriggers.trans_stack[my_level].state;
5494 5880 : if (state != NULL)
5495 4 : pfree(state);
5496 : /* this avoids double pfree if error later: */
5497 5880 : afterTriggers.trans_stack[my_level].state = NULL;
5498 : Assert(afterTriggers.query_depth ==
5499 : afterTriggers.trans_stack[my_level].query_depth);
5500 : }
5501 : else
5502 : {
5503 : /*
5504 : * Aborting. It is possible subxact start failed before calling
5505 : * AfterTriggerBeginSubXact, in which case we mustn't risk touching
5506 : * trans_stack levels that aren't there.
5507 : */
5508 6007 : if (my_level >= afterTriggers.maxtransdepth)
5509 0 : return;
5510 :
5511 : /*
5512 : * Release query-level storage for queries being aborted, and restore
5513 : * query_depth to its pre-subxact value. This assumes that a
5514 : * subtransaction will not add events to query levels started in a
5515 : * earlier transaction state.
5516 : */
5517 6064 : while (afterTriggers.query_depth > afterTriggers.trans_stack[my_level].query_depth)
5518 : {
5519 57 : if (afterTriggers.query_depth < afterTriggers.maxquerydepth)
5520 17 : AfterTriggerFreeQuery(&afterTriggers.query_stack[afterTriggers.query_depth]);
5521 57 : afterTriggers.query_depth--;
5522 : }
5523 : Assert(afterTriggers.query_depth ==
5524 : afterTriggers.trans_stack[my_level].query_depth);
5525 :
5526 : /*
5527 : * Restore the global deferred-event list to its former length,
5528 : * discarding any events queued by the subxact.
5529 : */
5530 6007 : afterTriggerRestoreEventList(&afterTriggers.events,
5531 6007 : &afterTriggers.trans_stack[my_level].events);
5532 :
5533 : /*
5534 : * Restore the trigger state. If the saved state is NULL, then this
5535 : * subxact didn't save it, so it doesn't need restoring.
5536 : */
5537 6007 : state = afterTriggers.trans_stack[my_level].state;
5538 6007 : if (state != NULL)
5539 : {
5540 2 : pfree(afterTriggers.state);
5541 2 : afterTriggers.state = state;
5542 : }
5543 : /* this avoids double pfree if error later: */
5544 6007 : afterTriggers.trans_stack[my_level].state = NULL;
5545 :
5546 : /*
5547 : * Scan for any remaining deferred events that were marked DONE or IN
5548 : * PROGRESS by this subxact or a child, and un-mark them. We can
5549 : * recognize such events because they have a firing ID greater than or
5550 : * equal to the firing_counter value we saved at subtransaction start.
5551 : * (This essentially assumes that the current subxact includes all
5552 : * subxacts started after it.)
5553 : */
5554 6007 : subxact_firing_id = afterTriggers.trans_stack[my_level].firing_counter;
5555 6035 : for_each_event_chunk(event, chunk, afterTriggers.events)
5556 : {
5557 14 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
5558 :
5559 14 : if (event->ate_flags &
5560 : (AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS))
5561 : {
5562 2 : if (evtshared->ats_firing_id >= subxact_firing_id)
5563 2 : event->ate_flags &=
5564 : ~(AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS);
5565 : }
5566 : }
5567 : }
5568 : }
5569 :
5570 : /*
5571 : * Get the transition table for the given event and depending on whether we are
5572 : * processing the old or the new tuple.
5573 : */
5574 : static Tuplestorestate *
5575 44126 : GetAfterTriggersTransitionTable(int event,
5576 : TupleTableSlot *oldslot,
5577 : TupleTableSlot *newslot,
5578 : TransitionCaptureState *transition_capture)
5579 : {
5580 44126 : Tuplestorestate *tuplestore = NULL;
5581 44126 : bool delete_old_table = transition_capture->tcs_delete_old_table;
5582 44126 : bool update_old_table = transition_capture->tcs_update_old_table;
5583 44126 : bool update_new_table = transition_capture->tcs_update_new_table;
5584 44126 : bool insert_new_table = transition_capture->tcs_insert_new_table;
5585 :
5586 : /*
5587 : * For INSERT events NEW should be non-NULL, for DELETE events OLD should
5588 : * be non-NULL, whereas for UPDATE events normally both OLD and NEW are
5589 : * non-NULL. But for UPDATE events fired for capturing transition tuples
5590 : * during UPDATE partition-key row movement, OLD is NULL when the event is
5591 : * for a row being inserted, whereas NEW is NULL when the event is for a
5592 : * row being deleted.
5593 : */
5594 : Assert(!(event == TRIGGER_EVENT_DELETE && delete_old_table &&
5595 : TupIsNull(oldslot)));
5596 : Assert(!(event == TRIGGER_EVENT_INSERT && insert_new_table &&
5597 : TupIsNull(newslot)));
5598 :
5599 44126 : if (!TupIsNull(oldslot))
5600 : {
5601 : Assert(TupIsNull(newslot));
5602 3619 : if (event == TRIGGER_EVENT_DELETE && delete_old_table)
5603 3372 : tuplestore = transition_capture->tcs_delete_private->old_tuplestore;
5604 247 : else if (event == TRIGGER_EVENT_UPDATE && update_old_table)
5605 231 : tuplestore = transition_capture->tcs_update_private->old_tuplestore;
5606 : }
5607 40507 : else if (!TupIsNull(newslot))
5608 : {
5609 : Assert(TupIsNull(oldslot));
5610 40507 : if (event == TRIGGER_EVENT_INSERT && insert_new_table)
5611 40260 : tuplestore = transition_capture->tcs_insert_private->new_tuplestore;
5612 247 : else if (event == TRIGGER_EVENT_UPDATE && update_new_table)
5613 243 : tuplestore = transition_capture->tcs_update_private->new_tuplestore;
5614 : }
5615 :
5616 44126 : return tuplestore;
5617 : }
5618 :
5619 : /*
5620 : * Add the given heap tuple to the given tuplestore, applying the conversion
5621 : * map if necessary.
5622 : *
5623 : * If original_insert_tuple is given, we can add that tuple without conversion.
5624 : */
5625 : static void
5626 44126 : TransitionTableAddTuple(EState *estate,
5627 : int event,
5628 : TransitionCaptureState *transition_capture,
5629 : ResultRelInfo *relinfo,
5630 : TupleTableSlot *slot,
5631 : TupleTableSlot *original_insert_tuple,
5632 : Tuplestorestate *tuplestore)
5633 : {
5634 : TupleConversionMap *map;
5635 :
5636 : /*
5637 : * Nothing needs to be done if we don't have a tuplestore.
5638 : */
5639 44126 : if (tuplestore == NULL)
5640 20 : return;
5641 :
5642 44106 : if (original_insert_tuple)
5643 96 : tuplestore_puttupleslot(tuplestore, original_insert_tuple);
5644 44010 : else if ((map = ExecGetChildToRootMap(relinfo)) != NULL)
5645 : {
5646 : AfterTriggersTableData *table;
5647 : TupleTableSlot *storeslot;
5648 :
5649 196 : switch (event)
5650 : {
5651 8 : case TRIGGER_EVENT_INSERT:
5652 8 : table = transition_capture->tcs_insert_private;
5653 8 : break;
5654 164 : case TRIGGER_EVENT_UPDATE:
5655 164 : table = transition_capture->tcs_update_private;
5656 164 : break;
5657 24 : case TRIGGER_EVENT_DELETE:
5658 24 : table = transition_capture->tcs_delete_private;
5659 24 : break;
5660 0 : default:
5661 0 : elog(ERROR, "invalid after-trigger event code: %d", event);
5662 : table = NULL; /* keep compiler quiet */
5663 : break;
5664 : }
5665 :
5666 196 : storeslot = GetAfterTriggersStoreSlot(table, map->outdesc);
5667 196 : execute_attr_map_slot(map->attrMap, slot, storeslot);
5668 196 : tuplestore_puttupleslot(tuplestore, storeslot);
5669 : }
5670 : else
5671 43814 : tuplestore_puttupleslot(tuplestore, slot);
5672 : }
5673 :
5674 : /* ----------
5675 : * AfterTriggerEnlargeQueryState()
5676 : *
5677 : * Prepare the necessary state so that we can record AFTER trigger events
5678 : * queued by a query. It is allowed to have nested queries within a
5679 : * (sub)transaction, so we need to have separate state for each query
5680 : * nesting level.
5681 : * ----------
5682 : */
5683 : static void
5684 4688 : AfterTriggerEnlargeQueryState(void)
5685 : {
5686 4688 : int init_depth = afterTriggers.maxquerydepth;
5687 :
5688 : Assert(afterTriggers.query_depth >= afterTriggers.maxquerydepth);
5689 :
5690 4688 : if (afterTriggers.maxquerydepth == 0)
5691 : {
5692 4688 : int new_alloc = Max(afterTriggers.query_depth + 1, 8);
5693 :
5694 4688 : afterTriggers.query_stack = (AfterTriggersQueryData *)
5695 4688 : MemoryContextAlloc(TopTransactionContext,
5696 : new_alloc * sizeof(AfterTriggersQueryData));
5697 4688 : afterTriggers.maxquerydepth = new_alloc;
5698 : }
5699 : else
5700 : {
5701 : /* repalloc will keep the stack in the same context */
5702 0 : int old_alloc = afterTriggers.maxquerydepth;
5703 0 : int new_alloc = Max(afterTriggers.query_depth + 1,
5704 : old_alloc * 2);
5705 :
5706 0 : afterTriggers.query_stack = (AfterTriggersQueryData *)
5707 0 : repalloc(afterTriggers.query_stack,
5708 : new_alloc * sizeof(AfterTriggersQueryData));
5709 0 : afterTriggers.maxquerydepth = new_alloc;
5710 : }
5711 :
5712 : /* Initialize new array entries to empty */
5713 42192 : while (init_depth < afterTriggers.maxquerydepth)
5714 : {
5715 37504 : AfterTriggersQueryData *qs = &afterTriggers.query_stack[init_depth];
5716 :
5717 37504 : qs->events.head = NULL;
5718 37504 : qs->events.tail = NULL;
5719 37504 : qs->events.tailfree = NULL;
5720 37504 : qs->fdw_tuplestore = NULL;
5721 37504 : qs->tables = NIL;
5722 :
5723 37504 : ++init_depth;
5724 : }
5725 4688 : }
5726 :
5727 : /*
5728 : * Create an empty SetConstraintState with room for numalloc trigstates
5729 : */
5730 : static SetConstraintState
5731 67 : SetConstraintStateCreate(int numalloc)
5732 : {
5733 : SetConstraintState state;
5734 :
5735 : /* Behave sanely with numalloc == 0 */
5736 67 : if (numalloc <= 0)
5737 6 : numalloc = 1;
5738 :
5739 : /*
5740 : * We assume that zeroing will correctly initialize the state values.
5741 : */
5742 : state = (SetConstraintState)
5743 67 : MemoryContextAllocZero(TopTransactionContext,
5744 : offsetof(SetConstraintStateData, trigstates) +
5745 67 : numalloc * sizeof(SetConstraintTriggerData));
5746 :
5747 67 : state->numalloc = numalloc;
5748 :
5749 67 : return state;
5750 : }
5751 :
5752 : /*
5753 : * Copy a SetConstraintState
5754 : */
5755 : static SetConstraintState
5756 6 : SetConstraintStateCopy(SetConstraintState origstate)
5757 : {
5758 : SetConstraintState state;
5759 :
5760 6 : state = SetConstraintStateCreate(origstate->numstates);
5761 :
5762 6 : state->all_isset = origstate->all_isset;
5763 6 : state->all_isdeferred = origstate->all_isdeferred;
5764 6 : state->numstates = origstate->numstates;
5765 6 : memcpy(state->trigstates, origstate->trigstates,
5766 6 : origstate->numstates * sizeof(SetConstraintTriggerData));
5767 :
5768 6 : return state;
5769 : }
5770 :
5771 : /*
5772 : * Add a per-trigger item to a SetConstraintState. Returns possibly-changed
5773 : * pointer to the state object (it will change if we have to repalloc).
5774 : */
5775 : static SetConstraintState
5776 228 : SetConstraintStateAddItem(SetConstraintState state,
5777 : Oid tgoid, bool tgisdeferred)
5778 : {
5779 228 : if (state->numstates >= state->numalloc)
5780 : {
5781 20 : int newalloc = state->numalloc * 2;
5782 :
5783 20 : newalloc = Max(newalloc, 8); /* in case original has size 0 */
5784 : state = (SetConstraintState)
5785 20 : repalloc(state,
5786 : offsetof(SetConstraintStateData, trigstates) +
5787 20 : newalloc * sizeof(SetConstraintTriggerData));
5788 20 : state->numalloc = newalloc;
5789 : Assert(state->numstates < state->numalloc);
5790 : }
5791 :
5792 228 : state->trigstates[state->numstates].sct_tgoid = tgoid;
5793 228 : state->trigstates[state->numstates].sct_tgisdeferred = tgisdeferred;
5794 228 : state->numstates++;
5795 :
5796 228 : return state;
5797 : }
5798 :
5799 : /* ----------
5800 : * AfterTriggerSetState()
5801 : *
5802 : * Execute the SET CONSTRAINTS ... utility command.
5803 : * ----------
5804 : */
5805 : void
5806 71 : AfterTriggerSetState(ConstraintsSetStmt *stmt)
5807 : {
5808 71 : int my_level = GetCurrentTransactionNestLevel();
5809 :
5810 : /* If we haven't already done so, initialize our state. */
5811 71 : if (afterTriggers.state == NULL)
5812 61 : afterTriggers.state = SetConstraintStateCreate(8);
5813 :
5814 : /*
5815 : * If in a subtransaction, and we didn't save the current state already,
5816 : * save it so it can be restored if the subtransaction aborts.
5817 : */
5818 71 : if (my_level > 1 &&
5819 6 : afterTriggers.trans_stack[my_level].state == NULL)
5820 : {
5821 6 : afterTriggers.trans_stack[my_level].state =
5822 6 : SetConstraintStateCopy(afterTriggers.state);
5823 : }
5824 :
5825 : /*
5826 : * Handle SET CONSTRAINTS ALL ...
5827 : */
5828 71 : if (stmt->constraints == NIL)
5829 : {
5830 : /*
5831 : * Forget any previous SET CONSTRAINTS commands in this transaction.
5832 : */
5833 39 : afterTriggers.state->numstates = 0;
5834 :
5835 : /*
5836 : * Set the per-transaction ALL state to known.
5837 : */
5838 39 : afterTriggers.state->all_isset = true;
5839 39 : afterTriggers.state->all_isdeferred = stmt->deferred;
5840 : }
5841 : else
5842 : {
5843 : Relation conrel;
5844 : Relation tgrel;
5845 32 : List *conoidlist = NIL;
5846 32 : List *tgoidlist = NIL;
5847 : ListCell *lc;
5848 :
5849 : /*
5850 : * Handle SET CONSTRAINTS constraint-name [, ...]
5851 : *
5852 : * First, identify all the named constraints and make a list of their
5853 : * OIDs. Since, unlike the SQL spec, we allow multiple constraints of
5854 : * the same name within a schema, the specifications are not
5855 : * necessarily unique. Our strategy is to target all matching
5856 : * constraints within the first search-path schema that has any
5857 : * matches, but disregard matches in schemas beyond the first match.
5858 : * (This is a bit odd but it's the historical behavior.)
5859 : *
5860 : * A constraint in a partitioned table may have corresponding
5861 : * constraints in the partitions. Grab those too.
5862 : */
5863 32 : conrel = table_open(ConstraintRelationId, AccessShareLock);
5864 :
5865 64 : foreach(lc, stmt->constraints)
5866 : {
5867 32 : RangeVar *constraint = lfirst(lc);
5868 : bool found;
5869 : List *namespacelist;
5870 : ListCell *nslc;
5871 :
5872 32 : if (constraint->catalogname)
5873 : {
5874 0 : if (strcmp(constraint->catalogname, get_database_name(MyDatabaseId)) != 0)
5875 0 : ereport(ERROR,
5876 : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
5877 : errmsg("cross-database references are not implemented: \"%s.%s.%s\"",
5878 : constraint->catalogname, constraint->schemaname,
5879 : constraint->relname)));
5880 : }
5881 :
5882 : /*
5883 : * If we're given the schema name with the constraint, look only
5884 : * in that schema. If given a bare constraint name, use the
5885 : * search path to find the first matching constraint.
5886 : */
5887 32 : if (constraint->schemaname)
5888 : {
5889 8 : Oid namespaceId = LookupExplicitNamespace(constraint->schemaname,
5890 : false);
5891 :
5892 8 : namespacelist = list_make1_oid(namespaceId);
5893 : }
5894 : else
5895 : {
5896 24 : namespacelist = fetch_search_path(true);
5897 : }
5898 :
5899 32 : found = false;
5900 80 : foreach(nslc, namespacelist)
5901 : {
5902 80 : Oid namespaceId = lfirst_oid(nslc);
5903 : SysScanDesc conscan;
5904 : ScanKeyData skey[2];
5905 : HeapTuple tup;
5906 :
5907 80 : ScanKeyInit(&skey[0],
5908 : Anum_pg_constraint_conname,
5909 : BTEqualStrategyNumber, F_NAMEEQ,
5910 80 : CStringGetDatum(constraint->relname));
5911 80 : ScanKeyInit(&skey[1],
5912 : Anum_pg_constraint_connamespace,
5913 : BTEqualStrategyNumber, F_OIDEQ,
5914 : ObjectIdGetDatum(namespaceId));
5915 :
5916 80 : conscan = systable_beginscan(conrel, ConstraintNameNspIndexId,
5917 : true, NULL, 2, skey);
5918 :
5919 144 : while (HeapTupleIsValid(tup = systable_getnext(conscan)))
5920 : {
5921 64 : Form_pg_constraint con = (Form_pg_constraint) GETSTRUCT(tup);
5922 :
5923 64 : if (con->condeferrable)
5924 64 : conoidlist = lappend_oid(conoidlist, con->oid);
5925 0 : else if (stmt->deferred)
5926 0 : ereport(ERROR,
5927 : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
5928 : errmsg("constraint \"%s\" is not deferrable",
5929 : constraint->relname)));
5930 64 : found = true;
5931 : }
5932 :
5933 80 : systable_endscan(conscan);
5934 :
5935 : /*
5936 : * Once we've found a matching constraint we do not search
5937 : * later parts of the search path.
5938 : */
5939 80 : if (found)
5940 32 : break;
5941 : }
5942 :
5943 32 : list_free(namespacelist);
5944 :
5945 : /*
5946 : * Not found ?
5947 : */
5948 32 : if (!found)
5949 0 : ereport(ERROR,
5950 : (errcode(ERRCODE_UNDEFINED_OBJECT),
5951 : errmsg("constraint \"%s\" does not exist",
5952 : constraint->relname)));
5953 : }
5954 :
5955 : /*
5956 : * Scan for any possible descendants of the constraints. We append
5957 : * whatever we find to the same list that we're scanning; this has the
5958 : * effect that we create new scans for those, too, so if there are
5959 : * further descendents, we'll also catch them.
5960 : */
5961 172 : foreach(lc, conoidlist)
5962 : {
5963 140 : Oid parent = lfirst_oid(lc);
5964 : ScanKeyData key;
5965 : SysScanDesc scan;
5966 : HeapTuple tuple;
5967 :
5968 140 : ScanKeyInit(&key,
5969 : Anum_pg_constraint_conparentid,
5970 : BTEqualStrategyNumber, F_OIDEQ,
5971 : ObjectIdGetDatum(parent));
5972 :
5973 140 : scan = systable_beginscan(conrel, ConstraintParentIndexId, true, NULL, 1, &key);
5974 :
5975 216 : while (HeapTupleIsValid(tuple = systable_getnext(scan)))
5976 : {
5977 76 : Form_pg_constraint con = (Form_pg_constraint) GETSTRUCT(tuple);
5978 :
5979 76 : conoidlist = lappend_oid(conoidlist, con->oid);
5980 : }
5981 :
5982 140 : systable_endscan(scan);
5983 : }
5984 :
5985 32 : table_close(conrel, AccessShareLock);
5986 :
5987 : /*
5988 : * Now, locate the trigger(s) implementing each of these constraints,
5989 : * and make a list of their OIDs.
5990 : */
5991 32 : tgrel = table_open(TriggerRelationId, AccessShareLock);
5992 :
5993 172 : foreach(lc, conoidlist)
5994 : {
5995 140 : Oid conoid = lfirst_oid(lc);
5996 : ScanKeyData skey;
5997 : SysScanDesc tgscan;
5998 : HeapTuple htup;
5999 :
6000 140 : ScanKeyInit(&skey,
6001 : Anum_pg_trigger_tgconstraint,
6002 : BTEqualStrategyNumber, F_OIDEQ,
6003 : ObjectIdGetDatum(conoid));
6004 :
6005 140 : tgscan = systable_beginscan(tgrel, TriggerConstraintIndexId, true,
6006 : NULL, 1, &skey);
6007 :
6008 572 : while (HeapTupleIsValid(htup = systable_getnext(tgscan)))
6009 : {
6010 292 : Form_pg_trigger pg_trigger = (Form_pg_trigger) GETSTRUCT(htup);
6011 :
6012 : /*
6013 : * Silently skip triggers that are marked as non-deferrable in
6014 : * pg_trigger. This is not an error condition, since a
6015 : * deferrable RI constraint may have some non-deferrable
6016 : * actions.
6017 : */
6018 292 : if (pg_trigger->tgdeferrable)
6019 292 : tgoidlist = lappend_oid(tgoidlist, pg_trigger->oid);
6020 : }
6021 :
6022 140 : systable_endscan(tgscan);
6023 : }
6024 :
6025 32 : table_close(tgrel, AccessShareLock);
6026 :
6027 : /*
6028 : * Now we can set the trigger states of individual triggers for this
6029 : * xact.
6030 : */
6031 324 : foreach(lc, tgoidlist)
6032 : {
6033 292 : Oid tgoid = lfirst_oid(lc);
6034 292 : SetConstraintState state = afterTriggers.state;
6035 292 : bool found = false;
6036 : int i;
6037 :
6038 1632 : for (i = 0; i < state->numstates; i++)
6039 : {
6040 1404 : if (state->trigstates[i].sct_tgoid == tgoid)
6041 : {
6042 64 : state->trigstates[i].sct_tgisdeferred = stmt->deferred;
6043 64 : found = true;
6044 64 : break;
6045 : }
6046 : }
6047 292 : if (!found)
6048 : {
6049 228 : afterTriggers.state =
6050 228 : SetConstraintStateAddItem(state, tgoid, stmt->deferred);
6051 : }
6052 : }
6053 : }
6054 :
6055 : /*
6056 : * SQL99 requires that when a constraint is set to IMMEDIATE, any deferred
6057 : * checks against that constraint must be made when the SET CONSTRAINTS
6058 : * command is executed -- i.e. the effects of the SET CONSTRAINTS command
6059 : * apply retroactively. We've updated the constraints state, so scan the
6060 : * list of previously deferred events to fire any that have now become
6061 : * immediate.
6062 : *
6063 : * Obviously, if this was SET ... DEFERRED then it can't have converted
6064 : * any unfired events to immediate, so we need do nothing in that case.
6065 : */
6066 71 : if (!stmt->deferred)
6067 : {
6068 26 : AfterTriggerEventList *events = &afterTriggers.events;
6069 26 : bool snapshot_set = false;
6070 :
6071 26 : afterTriggerFiringDepth++;
6072 26 : while (afterTriggerMarkEvents(events, NULL, true))
6073 : {
6074 14 : CommandId firing_id = afterTriggers.firing_counter++;
6075 :
6076 : /*
6077 : * Make sure a snapshot has been established in case trigger
6078 : * functions need one. Note that we avoid setting a snapshot if
6079 : * we don't find at least one trigger that has to be fired now.
6080 : * This is so that BEGIN; SET CONSTRAINTS ...; SET TRANSACTION
6081 : * ISOLATION LEVEL SERIALIZABLE; ... works properly. (If we are
6082 : * at the start of a transaction it's not possible for any trigger
6083 : * events to be queued yet.)
6084 : */
6085 14 : if (!snapshot_set)
6086 : {
6087 14 : PushActiveSnapshot(GetTransactionSnapshot());
6088 14 : snapshot_set = true;
6089 : }
6090 :
6091 : /*
6092 : * We can delete fired events if we are at top transaction level,
6093 : * but we'd better not if inside a subtransaction, since the
6094 : * subtransaction could later get rolled back.
6095 : */
6096 10 : if (afterTriggerInvokeEvents(events, firing_id, NULL,
6097 14 : !IsSubTransaction()))
6098 10 : break; /* all fired */
6099 : }
6100 :
6101 : /*
6102 : * Flush any fast-path batches accumulated by the triggers just fired.
6103 : */
6104 22 : FireAfterTriggerBatchCallbacks();
6105 16 : afterTriggerFiringDepth--;
6106 :
6107 16 : if (snapshot_set)
6108 4 : PopActiveSnapshot();
6109 : }
6110 61 : }
6111 :
6112 : /* ----------
6113 : * AfterTriggerPendingOnRel()
6114 : * Test to see if there are any pending after-trigger events for rel.
6115 : *
6116 : * This is used by TRUNCATE, CLUSTER, ALTER TABLE, etc to detect whether
6117 : * it is unsafe to perform major surgery on a relation. Note that only
6118 : * local pending events are examined. We assume that having exclusive lock
6119 : * on a rel guarantees there are no unserviced events in other backends ---
6120 : * but having a lock does not prevent there being such events in our own.
6121 : *
6122 : * In some scenarios it'd be reasonable to remove pending events (more
6123 : * specifically, mark them DONE by the current subxact) but without a lot
6124 : * of knowledge of the trigger semantics we can't do this in general.
6125 : * ----------
6126 : */
6127 : bool
6128 92926 : AfterTriggerPendingOnRel(Oid relid)
6129 : {
6130 : AfterTriggerEvent event;
6131 : AfterTriggerEventChunk *chunk;
6132 : int depth;
6133 :
6134 : /* Scan queued events */
6135 92950 : for_each_event_chunk(event, chunk, afterTriggers.events)
6136 : {
6137 24 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
6138 :
6139 : /*
6140 : * We can ignore completed events. (Even if a DONE flag is rolled
6141 : * back by subxact abort, it's OK because the effects of the TRUNCATE
6142 : * or whatever must get rolled back too.)
6143 : */
6144 24 : if (event->ate_flags & AFTER_TRIGGER_DONE)
6145 0 : continue;
6146 :
6147 24 : if (evtshared->ats_relid == relid)
6148 12 : return true;
6149 : }
6150 :
6151 : /*
6152 : * Also scan events queued by incomplete queries. This could only matter
6153 : * if TRUNCATE/etc is executed by a function or trigger within an updating
6154 : * query on the same relation, which is pretty perverse, but let's check.
6155 : */
6156 92914 : for (depth = 0; depth <= afterTriggers.query_depth && depth < afterTriggers.maxquerydepth; depth++)
6157 : {
6158 0 : for_each_event_chunk(event, chunk, afterTriggers.query_stack[depth].events)
6159 : {
6160 0 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
6161 :
6162 0 : if (event->ate_flags & AFTER_TRIGGER_DONE)
6163 0 : continue;
6164 :
6165 0 : if (evtshared->ats_relid == relid)
6166 0 : return true;
6167 : }
6168 : }
6169 :
6170 92914 : return false;
6171 : }
6172 :
6173 : /* ----------
6174 : * AfterTriggerSaveEvent()
6175 : *
6176 : * Called by ExecA[RS]...Triggers() to queue up the triggers that should
6177 : * be fired for an event.
6178 : *
6179 : * NOTE: this is called whenever there are any triggers associated with
6180 : * the event (even if they are disabled). This function decides which
6181 : * triggers actually need to be queued. It is also called after each row,
6182 : * even if there are no triggers for that event, if there are any AFTER
6183 : * STATEMENT triggers for the statement which use transition tables, so that
6184 : * the transition tuplestores can be built. Furthermore, if the transition
6185 : * capture is happening for UPDATEd rows being moved to another partition due
6186 : * to the partition-key being changed, then this function is called once when
6187 : * the row is deleted (to capture OLD row), and once when the row is inserted
6188 : * into another partition (to capture NEW row). This is done separately because
6189 : * DELETE and INSERT happen on different tables.
6190 : *
6191 : * Transition tuplestores are built now, rather than when events are pulled
6192 : * off of the queue because AFTER ROW triggers are allowed to select from the
6193 : * transition tables for the statement.
6194 : *
6195 : * This contains special support to queue the update events for the case where
6196 : * a partitioned table undergoing a cross-partition update may have foreign
6197 : * keys pointing into it. Normally, a partitioned table's row triggers are
6198 : * not fired because the leaf partition(s) which are modified as a result of
6199 : * the operation on the partitioned table contain the same triggers which are
6200 : * fired instead. But that general scheme can cause problematic behavior with
6201 : * foreign key triggers during cross-partition updates, which are implemented
6202 : * as DELETE on the source partition followed by INSERT into the destination
6203 : * partition. Specifically, firing DELETE triggers would lead to the wrong
6204 : * foreign key action to be enforced considering that the original command is
6205 : * UPDATE; in this case, this function is called with relinfo as the
6206 : * partitioned table, and src_partinfo and dst_partinfo referring to the
6207 : * source and target leaf partitions, respectively.
6208 : *
6209 : * is_crosspart_update is true either when a DELETE event is fired on the
6210 : * source partition (which is to be ignored) or an UPDATE event is fired on
6211 : * the root partitioned table.
6212 : * ----------
6213 : */
6214 : static void
6215 452918 : AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
6216 : ResultRelInfo *src_partinfo,
6217 : ResultRelInfo *dst_partinfo,
6218 : int event, bool row_trigger,
6219 : TupleTableSlot *oldslot, TupleTableSlot *newslot,
6220 : List *recheckIndexes, Bitmapset *modifiedCols,
6221 : TransitionCaptureState *transition_capture,
6222 : bool is_crosspart_update)
6223 : {
6224 452918 : Relation rel = relinfo->ri_RelationDesc;
6225 452918 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
6226 : AfterTriggerEventData new_event;
6227 : AfterTriggerSharedData new_shared;
6228 452918 : char relkind = rel->rd_rel->relkind;
6229 : int tgtype_event;
6230 : int tgtype_level;
6231 : int i;
6232 452918 : Tuplestorestate *fdw_tuplestore = NULL;
6233 :
6234 : /*
6235 : * Check state. We use a normal test not Assert because it is possible to
6236 : * reach here in the wrong state given misconfigured RI triggers, in
6237 : * particular deferring a cascade action trigger.
6238 : */
6239 452918 : if (afterTriggers.query_depth < 0)
6240 0 : elog(ERROR, "AfterTriggerSaveEvent() called outside of query");
6241 :
6242 : /* Be sure we have enough space to record events at this query depth. */
6243 452918 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
6244 4146 : AfterTriggerEnlargeQueryState();
6245 :
6246 : /*
6247 : * If the directly named relation has any triggers with transition tables,
6248 : * then we need to capture transition tuples.
6249 : */
6250 452918 : if (row_trigger && transition_capture != NULL)
6251 : {
6252 43911 : TupleTableSlot *original_insert_tuple = transition_capture->tcs_original_insert_tuple;
6253 :
6254 : /*
6255 : * Capture the old tuple in the appropriate transition table based on
6256 : * the event.
6257 : */
6258 43911 : if (!TupIsNull(oldslot))
6259 : {
6260 : Tuplestorestate *old_tuplestore;
6261 :
6262 3619 : old_tuplestore = GetAfterTriggersTransitionTable(event,
6263 : oldslot,
6264 : NULL,
6265 : transition_capture);
6266 3619 : TransitionTableAddTuple(estate, event, transition_capture, relinfo,
6267 : oldslot, NULL, old_tuplestore);
6268 : }
6269 :
6270 : /*
6271 : * Capture the new tuple in the appropriate transition table based on
6272 : * the event.
6273 : */
6274 43911 : if (!TupIsNull(newslot))
6275 : {
6276 : Tuplestorestate *new_tuplestore;
6277 :
6278 40507 : new_tuplestore = GetAfterTriggersTransitionTable(event,
6279 : NULL,
6280 : newslot,
6281 : transition_capture);
6282 40507 : TransitionTableAddTuple(estate, event, transition_capture, relinfo,
6283 : newslot, original_insert_tuple, new_tuplestore);
6284 : }
6285 :
6286 : /*
6287 : * If transition tables are the only reason we're here, return. As
6288 : * mentioned above, we can also be here during update tuple routing in
6289 : * presence of transition tables, in which case this function is
6290 : * called separately for OLD and NEW, so we expect exactly one of them
6291 : * to be NULL.
6292 : */
6293 43911 : if (trigdesc == NULL ||
6294 43751 : (event == TRIGGER_EVENT_DELETE && !trigdesc->trig_delete_after_row) ||
6295 40431 : (event == TRIGGER_EVENT_INSERT && !trigdesc->trig_insert_after_row) ||
6296 263 : (event == TRIGGER_EVENT_UPDATE && !trigdesc->trig_update_after_row) ||
6297 32 : (event == TRIGGER_EVENT_UPDATE && (TupIsNull(oldslot) ^ TupIsNull(newslot))))
6298 43807 : return;
6299 : }
6300 :
6301 : /*
6302 : * We normally don't see partitioned tables here for row level triggers
6303 : * except in the special case of a cross-partition update. In that case,
6304 : * nodeModifyTable.c:ExecCrossPartitionUpdateForeignKey() calls here to
6305 : * queue an update event on the root target partitioned table, also
6306 : * passing the source and destination partitions and their tuples.
6307 : */
6308 : Assert(!row_trigger ||
6309 : rel->rd_rel->relkind != RELKIND_PARTITIONED_TABLE ||
6310 : (is_crosspart_update &&
6311 : TRIGGER_FIRED_BY_UPDATE(event) &&
6312 : src_partinfo != NULL && dst_partinfo != NULL));
6313 :
6314 : /*
6315 : * Validate the event code and collect the associated tuple CTIDs.
6316 : *
6317 : * The event code will be used both as a bitmask and an array offset, so
6318 : * validation is important to make sure we don't walk off the edge of our
6319 : * arrays.
6320 : *
6321 : * Also, if we're considering statement-level triggers, check whether we
6322 : * already queued a set of them for this event, and cancel the prior set
6323 : * if so. This preserves the behavior that statement-level triggers fire
6324 : * just once per statement and fire after row-level triggers.
6325 : */
6326 409111 : switch (event)
6327 : {
6328 405657 : case TRIGGER_EVENT_INSERT:
6329 405657 : tgtype_event = TRIGGER_TYPE_INSERT;
6330 405657 : if (row_trigger)
6331 : {
6332 : Assert(oldslot == NULL);
6333 : Assert(newslot != NULL);
6334 405308 : ItemPointerCopy(&(newslot->tts_tid), &(new_event.ate_ctid1));
6335 405308 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6336 : }
6337 : else
6338 : {
6339 : Assert(oldslot == NULL);
6340 : Assert(newslot == NULL);
6341 349 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6342 349 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6343 349 : cancel_prior_stmt_triggers(RelationGetRelid(rel),
6344 : CMD_INSERT, event);
6345 : }
6346 405657 : break;
6347 975 : case TRIGGER_EVENT_DELETE:
6348 975 : tgtype_event = TRIGGER_TYPE_DELETE;
6349 975 : if (row_trigger)
6350 : {
6351 : Assert(oldslot != NULL);
6352 : Assert(newslot == NULL);
6353 810 : ItemPointerCopy(&(oldslot->tts_tid), &(new_event.ate_ctid1));
6354 810 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6355 : }
6356 : else
6357 : {
6358 : Assert(oldslot == NULL);
6359 : Assert(newslot == NULL);
6360 165 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6361 165 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6362 165 : cancel_prior_stmt_triggers(RelationGetRelid(rel),
6363 : CMD_DELETE, event);
6364 : }
6365 975 : break;
6366 2474 : case TRIGGER_EVENT_UPDATE:
6367 2474 : tgtype_event = TRIGGER_TYPE_UPDATE;
6368 2474 : if (row_trigger)
6369 : {
6370 : Assert(oldslot != NULL);
6371 : Assert(newslot != NULL);
6372 2197 : ItemPointerCopy(&(oldslot->tts_tid), &(new_event.ate_ctid1));
6373 2197 : ItemPointerCopy(&(newslot->tts_tid), &(new_event.ate_ctid2));
6374 :
6375 : /*
6376 : * Also remember the OIDs of partitions to fetch these tuples
6377 : * out of later in AfterTriggerExecute().
6378 : */
6379 2197 : if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6380 : {
6381 : Assert(src_partinfo != NULL && dst_partinfo != NULL);
6382 192 : new_event.ate_src_part =
6383 192 : RelationGetRelid(src_partinfo->ri_RelationDesc);
6384 192 : new_event.ate_dst_part =
6385 192 : RelationGetRelid(dst_partinfo->ri_RelationDesc);
6386 : }
6387 : }
6388 : else
6389 : {
6390 : Assert(oldslot == NULL);
6391 : Assert(newslot == NULL);
6392 277 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6393 277 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6394 277 : cancel_prior_stmt_triggers(RelationGetRelid(rel),
6395 : CMD_UPDATE, event);
6396 : }
6397 2474 : break;
6398 5 : case TRIGGER_EVENT_TRUNCATE:
6399 5 : tgtype_event = TRIGGER_TYPE_TRUNCATE;
6400 : Assert(oldslot == NULL);
6401 : Assert(newslot == NULL);
6402 5 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6403 5 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6404 5 : break;
6405 0 : default:
6406 0 : elog(ERROR, "invalid after-trigger event code: %d", event);
6407 : tgtype_event = 0; /* keep compiler quiet */
6408 : break;
6409 : }
6410 :
6411 : /* Determine flags */
6412 409111 : if (!(relkind == RELKIND_FOREIGN_TABLE && row_trigger))
6413 : {
6414 409083 : if (row_trigger && event == TRIGGER_EVENT_UPDATE)
6415 : {
6416 2187 : if (relkind == RELKIND_PARTITIONED_TABLE)
6417 192 : new_event.ate_flags = AFTER_TRIGGER_CP_UPDATE;
6418 : else
6419 1995 : new_event.ate_flags = AFTER_TRIGGER_2CTID;
6420 : }
6421 : else
6422 406896 : new_event.ate_flags = AFTER_TRIGGER_1CTID;
6423 : }
6424 :
6425 : /* else, we'll initialize ate_flags for each trigger */
6426 :
6427 409111 : tgtype_level = (row_trigger ? TRIGGER_TYPE_ROW : TRIGGER_TYPE_STATEMENT);
6428 :
6429 : /*
6430 : * Must convert/copy the source and destination partition tuples into the
6431 : * root partitioned table's format/slot, because the processing in the
6432 : * loop below expects both oldslot and newslot tuples to be in that form.
6433 : */
6434 409111 : if (row_trigger && rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6435 : {
6436 : TupleTableSlot *rootslot;
6437 : TupleConversionMap *map;
6438 :
6439 192 : rootslot = ExecGetTriggerOldSlot(estate, relinfo);
6440 192 : map = ExecGetChildToRootMap(src_partinfo);
6441 192 : if (map)
6442 24 : oldslot = execute_attr_map_slot(map->attrMap,
6443 : oldslot,
6444 : rootslot);
6445 : else
6446 168 : oldslot = ExecCopySlot(rootslot, oldslot);
6447 :
6448 192 : rootslot = ExecGetTriggerNewSlot(estate, relinfo);
6449 192 : map = ExecGetChildToRootMap(dst_partinfo);
6450 192 : if (map)
6451 28 : newslot = execute_attr_map_slot(map->attrMap,
6452 : newslot,
6453 : rootslot);
6454 : else
6455 164 : newslot = ExecCopySlot(rootslot, newslot);
6456 : }
6457 :
6458 1639376 : for (i = 0; i < trigdesc->numtriggers; i++)
6459 : {
6460 1230265 : Trigger *trigger = &trigdesc->triggers[i];
6461 :
6462 1230265 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
6463 : tgtype_level,
6464 : TRIGGER_TYPE_AFTER,
6465 : tgtype_event))
6466 619182 : continue;
6467 611083 : if (!TriggerEnabled(estate, relinfo, trigger, event,
6468 : modifiedCols, oldslot, newslot))
6469 273 : continue;
6470 :
6471 610810 : if (relkind == RELKIND_FOREIGN_TABLE && row_trigger)
6472 : {
6473 29 : if (fdw_tuplestore == NULL)
6474 : {
6475 25 : fdw_tuplestore = GetCurrentFDWTuplestore();
6476 25 : new_event.ate_flags = AFTER_TRIGGER_FDW_FETCH;
6477 : }
6478 : else
6479 : /* subsequent event for the same tuple */
6480 4 : new_event.ate_flags = AFTER_TRIGGER_FDW_REUSE;
6481 : }
6482 :
6483 : /*
6484 : * If the trigger is a foreign key enforcement trigger, there are
6485 : * certain cases where we can skip queueing the event because we can
6486 : * tell by inspection that the FK constraint will still pass. There
6487 : * are also some cases during cross-partition updates of a partitioned
6488 : * table where queuing the event can be skipped.
6489 : */
6490 610810 : if (TRIGGER_FIRED_BY_UPDATE(event) || TRIGGER_FIRED_BY_DELETE(event))
6491 : {
6492 4353 : switch (RI_FKey_trigger_type(trigger->tgfoid))
6493 : {
6494 1700 : case RI_TRIGGER_PK:
6495 :
6496 : /*
6497 : * For cross-partitioned updates of partitioned PK table,
6498 : * skip the event fired by the component delete on the
6499 : * source leaf partition unless the constraint originates
6500 : * in the partition itself (!tgisclone), because the
6501 : * update event that will be fired on the root
6502 : * (partitioned) target table will be used to perform the
6503 : * necessary foreign key enforcement action.
6504 : */
6505 1700 : if (is_crosspart_update &&
6506 338 : TRIGGER_FIRED_BY_DELETE(event) &&
6507 178 : trigger->tgisclone)
6508 168 : continue;
6509 :
6510 : /* Update or delete on trigger's PK table */
6511 1532 : if (!RI_FKey_pk_upd_check_required(trigger, rel,
6512 : oldslot, newslot))
6513 : {
6514 : /* skip queuing this event */
6515 292 : continue;
6516 : }
6517 1240 : break;
6518 :
6519 748 : case RI_TRIGGER_FK:
6520 :
6521 : /*
6522 : * Update on trigger's FK table. We can skip the update
6523 : * event fired on a partitioned table during a
6524 : * cross-partition of that table, because the insert event
6525 : * that is fired on the destination leaf partition would
6526 : * suffice to perform the necessary foreign key check.
6527 : * Moreover, RI_FKey_fk_upd_check_required() expects to be
6528 : * passed a tuple that contains system attributes, most of
6529 : * which are not present in the virtual slot belonging to
6530 : * a partitioned table.
6531 : */
6532 748 : if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE ||
6533 668 : !RI_FKey_fk_upd_check_required(trigger, rel,
6534 : oldslot, newslot))
6535 : {
6536 : /* skip queuing this event */
6537 442 : continue;
6538 : }
6539 306 : break;
6540 :
6541 1905 : case RI_TRIGGER_NONE:
6542 :
6543 : /*
6544 : * Not an FK trigger. No need to queue the update event
6545 : * fired during a cross-partitioned update of a
6546 : * partitioned table, because the same row trigger must be
6547 : * present in the leaf partition(s) that are affected as
6548 : * part of this update and the events fired on them are
6549 : * queued instead.
6550 : */
6551 1905 : if (row_trigger &&
6552 1439 : rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6553 20 : continue;
6554 1885 : break;
6555 : }
6556 : }
6557 :
6558 : /*
6559 : * If the trigger is a deferred unique constraint check trigger, only
6560 : * queue it if the unique constraint was potentially violated, which
6561 : * we know from index insertion time.
6562 : */
6563 609888 : if (trigger->tgfoid == F_UNIQUE_KEY_RECHECK)
6564 : {
6565 139 : if (!list_member_oid(recheckIndexes, trigger->tgconstrindid))
6566 58 : continue; /* Uniqueness definitely not violated */
6567 : }
6568 :
6569 : /*
6570 : * Fill in event structure and add it to the current query's queue.
6571 : * Note we set ats_table to NULL whenever this trigger doesn't use
6572 : * transition tables, to improve sharability of the shared event data.
6573 : */
6574 609830 : new_shared.ats_event =
6575 1219660 : (event & TRIGGER_EVENT_OPMASK) |
6576 609830 : (row_trigger ? TRIGGER_EVENT_ROW : 0) |
6577 609830 : (trigger->tgdeferrable ? AFTER_TRIGGER_DEFERRABLE : 0) |
6578 609830 : (trigger->tginitdeferred ? AFTER_TRIGGER_INITDEFERRED : 0);
6579 609830 : new_shared.ats_tgoid = trigger->tgoid;
6580 609830 : new_shared.ats_relid = RelationGetRelid(rel);
6581 609830 : new_shared.ats_rolid = GetUserId();
6582 609830 : new_shared.ats_firing_id = 0;
6583 609830 : if ((trigger->tgoldtable || trigger->tgnewtable) &&
6584 : transition_capture != NULL)
6585 : {
6586 487 : switch (event)
6587 : {
6588 192 : case TRIGGER_EVENT_INSERT:
6589 192 : new_shared.ats_table = transition_capture->tcs_insert_private;
6590 192 : break;
6591 155 : case TRIGGER_EVENT_UPDATE:
6592 155 : new_shared.ats_table = transition_capture->tcs_update_private;
6593 155 : break;
6594 140 : case TRIGGER_EVENT_DELETE:
6595 140 : new_shared.ats_table = transition_capture->tcs_delete_private;
6596 140 : break;
6597 0 : default:
6598 : /* Must be TRUNCATE, see switch above */
6599 0 : new_shared.ats_table = NULL;
6600 0 : break;
6601 : }
6602 : }
6603 : else
6604 609343 : new_shared.ats_table = NULL;
6605 609830 : new_shared.ats_modifiedcols = modifiedCols;
6606 :
6607 609830 : afterTriggerAddEvent(&afterTriggers.query_stack[afterTriggers.query_depth].events,
6608 : &new_event, &new_shared);
6609 : }
6610 :
6611 : /*
6612 : * Finally, spool any foreign tuple(s). The tuplestore squashes them to
6613 : * minimal tuples, so this loses any system columns. The executor lost
6614 : * those columns before us, for an unrelated reason, so this is fine.
6615 : */
6616 409111 : if (fdw_tuplestore)
6617 : {
6618 25 : if (oldslot != NULL)
6619 16 : tuplestore_puttupleslot(fdw_tuplestore, oldslot);
6620 25 : if (newslot != NULL)
6621 18 : tuplestore_puttupleslot(fdw_tuplestore, newslot);
6622 : }
6623 : }
6624 :
6625 : /*
6626 : * Detect whether we already queued BEFORE STATEMENT triggers for the given
6627 : * relation + operation, and set the flag so the next call will report "true".
6628 : */
6629 : static bool
6630 387 : before_stmt_triggers_fired(Oid relid, CmdType cmdType)
6631 : {
6632 : bool result;
6633 : AfterTriggersTableData *table;
6634 :
6635 : /* Check state, like AfterTriggerSaveEvent. */
6636 387 : if (afterTriggers.query_depth < 0)
6637 0 : elog(ERROR, "before_stmt_triggers_fired() called outside of query");
6638 :
6639 : /* Be sure we have enough space to record events at this query depth. */
6640 387 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
6641 227 : AfterTriggerEnlargeQueryState();
6642 :
6643 : /*
6644 : * We keep this state in the AfterTriggersTableData that also holds
6645 : * transition tables for the relation + operation. In this way, if we are
6646 : * forced to make a new set of transition tables because more tuples get
6647 : * entered after we've already fired triggers, we will allow a new set of
6648 : * statement triggers to get queued.
6649 : */
6650 387 : table = GetAfterTriggersTableData(relid, cmdType);
6651 387 : result = table->before_trig_done;
6652 387 : table->before_trig_done = true;
6653 387 : return result;
6654 : }
6655 :
6656 : /*
6657 : * If we previously queued a set of AFTER STATEMENT triggers for the given
6658 : * relation + operation, and they've not been fired yet, cancel them. The
6659 : * caller will queue a fresh set that's after any row-level triggers that may
6660 : * have been queued by the current sub-statement, preserving (as much as
6661 : * possible) the property that AFTER ROW triggers fire before AFTER STATEMENT
6662 : * triggers, and that the latter only fire once. This deals with the
6663 : * situation where several FK enforcement triggers sequentially queue triggers
6664 : * for the same table into the same trigger query level. We can't fully
6665 : * prevent odd behavior though: if there are AFTER ROW triggers taking
6666 : * transition tables, we don't want to change the transition tables once the
6667 : * first such trigger has seen them. In such a case, any additional events
6668 : * will result in creating new transition tables and allowing new firings of
6669 : * statement triggers.
6670 : *
6671 : * This also saves the current event list location so that a later invocation
6672 : * of this function can cheaply find the triggers we're about to queue and
6673 : * cancel them.
6674 : */
6675 : static void
6676 791 : cancel_prior_stmt_triggers(Oid relid, CmdType cmdType, int tgevent)
6677 : {
6678 : AfterTriggersTableData *table;
6679 791 : AfterTriggersQueryData *qs = &afterTriggers.query_stack[afterTriggers.query_depth];
6680 :
6681 : /*
6682 : * We keep this state in the AfterTriggersTableData that also holds
6683 : * transition tables for the relation + operation. In this way, if we are
6684 : * forced to make a new set of transition tables because more tuples get
6685 : * entered after we've already fired triggers, we will allow a new set of
6686 : * statement triggers to get queued without canceling the old ones.
6687 : */
6688 791 : table = GetAfterTriggersTableData(relid, cmdType);
6689 :
6690 791 : if (table->after_trig_done)
6691 : {
6692 : /*
6693 : * We want to start scanning from the tail location that existed just
6694 : * before we inserted any statement triggers. But the events list
6695 : * might've been entirely empty then, in which case scan from the
6696 : * current head.
6697 : */
6698 : AfterTriggerEvent event;
6699 : AfterTriggerEventChunk *chunk;
6700 :
6701 48 : if (table->after_trig_events.tail)
6702 : {
6703 40 : chunk = table->after_trig_events.tail;
6704 40 : event = (AfterTriggerEvent) table->after_trig_events.tailfree;
6705 : }
6706 : else
6707 : {
6708 8 : chunk = qs->events.head;
6709 8 : event = NULL;
6710 : }
6711 :
6712 72 : for_each_chunk_from(chunk)
6713 : {
6714 48 : if (event == NULL)
6715 8 : event = (AfterTriggerEvent) CHUNK_DATA_START(chunk);
6716 100 : for_each_event_from(event, chunk)
6717 : {
6718 76 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
6719 :
6720 : /*
6721 : * Exit loop when we reach events that aren't AS triggers for
6722 : * the target relation.
6723 : */
6724 76 : if (evtshared->ats_relid != relid)
6725 0 : goto done;
6726 76 : if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) != tgevent)
6727 0 : goto done;
6728 76 : if (!TRIGGER_FIRED_FOR_STATEMENT(evtshared->ats_event))
6729 24 : goto done;
6730 52 : if (!TRIGGER_FIRED_AFTER(evtshared->ats_event))
6731 0 : goto done;
6732 : /* OK, mark it DONE */
6733 52 : event->ate_flags &= ~AFTER_TRIGGER_IN_PROGRESS;
6734 52 : event->ate_flags |= AFTER_TRIGGER_DONE;
6735 : }
6736 : /* signal we must reinitialize event ptr for next chunk */
6737 24 : event = NULL;
6738 : }
6739 : }
6740 767 : done:
6741 :
6742 : /* In any case, save current insertion point for next time */
6743 791 : table->after_trig_done = true;
6744 791 : table->after_trig_events = qs->events;
6745 791 : }
6746 :
6747 : /*
6748 : * GUC assign_hook for session_replication_role
6749 : */
6750 : void
6751 1919 : assign_session_replication_role(int newval, void *extra)
6752 : {
6753 : /*
6754 : * Must flush the plan cache when changing replication role; but don't
6755 : * flush unnecessarily.
6756 : */
6757 1919 : if (SessionReplicationRole != newval)
6758 644 : ResetPlanCache();
6759 1919 : }
6760 :
6761 : /*
6762 : * SQL function pg_trigger_depth()
6763 : */
6764 : Datum
6765 84 : pg_trigger_depth(PG_FUNCTION_ARGS)
6766 : {
6767 84 : PG_RETURN_INT32(MyTriggerDepth);
6768 : }
6769 :
6770 : /*
6771 : * Check whether a trigger modified a virtual generated column and replace the
6772 : * value with null if so.
6773 : *
6774 : * We need to check this so that we don't end up storing a non-null value in a
6775 : * virtual generated column.
6776 : *
6777 : * We don't need to check for stored generated columns, since those will be
6778 : * overwritten later anyway.
6779 : */
6780 : static HeapTuple
6781 1268 : check_modified_virtual_generated(TupleDesc tupdesc, HeapTuple tuple)
6782 : {
6783 1268 : if (!(tupdesc->constr && tupdesc->constr->has_generated_virtual))
6784 1257 : return tuple;
6785 :
6786 39 : for (int i = 0; i < tupdesc->natts; i++)
6787 : {
6788 28 : if (TupleDescAttr(tupdesc, i)->attgenerated == ATTRIBUTE_GENERATED_VIRTUAL)
6789 : {
6790 11 : if (!heap_attisnull(tuple, i + 1, tupdesc))
6791 : {
6792 8 : int replCol = i + 1;
6793 8 : Datum replValue = 0;
6794 8 : bool replIsnull = true;
6795 :
6796 8 : tuple = heap_modify_tuple_by_cols(tuple, tupdesc, 1, &replCol, &replValue, &replIsnull);
6797 : }
6798 : }
6799 : }
6800 :
6801 11 : return tuple;
6802 : }
6803 :
6804 : /*
6805 : * RegisterAfterTriggerBatchCallback
6806 : * Register a function to be called when the current trigger-firing
6807 : * batch completes.
6808 : *
6809 : * Must be called from within a trigger function's execution context
6810 : * (i.e., while afterTriggers state is active).
6811 : *
6812 : * The callback list is cleared after invocation, so the caller must
6813 : * re-register for each new batch if needed.
6814 : */
6815 : void
6816 1476 : RegisterAfterTriggerBatchCallback(AfterTriggerBatchCallback callback,
6817 : void *arg)
6818 : {
6819 : AfterTriggerCallbackItem *item;
6820 : MemoryContext oldcxt;
6821 :
6822 : /*
6823 : * Allocate in TopTransactionContext so the item survives for the duration
6824 : * of the batch, which may span multiple trigger invocations.
6825 : *
6826 : * Must be called while afterTriggers is active; callbacks registered
6827 : * outside a trigger-firing context would never fire.
6828 : */
6829 : Assert(afterTriggerFiringDepth > 0);
6830 1476 : oldcxt = MemoryContextSwitchTo(TopTransactionContext);
6831 1476 : item = palloc(sizeof(AfterTriggerCallbackItem));
6832 1476 : item->callback = callback;
6833 1476 : item->arg = arg;
6834 1476 : afterTriggers.batch_callbacks =
6835 1476 : lappend(afterTriggers.batch_callbacks, item);
6836 1476 : MemoryContextSwitchTo(oldcxt);
6837 1476 : }
6838 :
6839 : /*
6840 : * FireAfterTriggerBatchCallbacks
6841 : * Invoke and clear all registered batch callbacks.
6842 : *
6843 : * Only fires at the outermost query level (query_depth == 0) or from
6844 : * top-level operations (query_depth == -1, e.g. AfterTriggerFireDeferred
6845 : * at COMMIT). Nested queries from SPI inside AFTER triggers run at
6846 : * depth > 0 and must not tear down resources the outer batch still needs.
6847 : */
6848 : static void
6849 605003 : FireAfterTriggerBatchCallbacks(void)
6850 : {
6851 : ListCell *lc;
6852 :
6853 605003 : if (afterTriggers.query_depth > 0)
6854 841 : return;
6855 :
6856 : Assert(afterTriggerFiringDepth > 0);
6857 605355 : foreach(lc, afterTriggers.batch_callbacks)
6858 : {
6859 1477 : AfterTriggerCallbackItem *item = lfirst(lc);
6860 :
6861 1477 : item->callback(item->arg);
6862 : }
6863 :
6864 603878 : list_free_deep(afterTriggers.batch_callbacks);
6865 603878 : afterTriggers.batch_callbacks = NIL;
6866 : }
6867 :
6868 : /*
6869 : * AfterTriggerIsActive
6870 : * Returns true if we're inside the after-trigger framework where
6871 : * registered batch callbacks will actually be invoked.
6872 : *
6873 : * This is false during validateForeignKeyConstraint(), which calls
6874 : * RI trigger functions directly outside the after-trigger framework.
6875 : */
6876 : bool
6877 603674 : AfterTriggerIsActive(void)
6878 : {
6879 603674 : return afterTriggerFiringDepth > 0;
6880 : }
|