Age Owner Branch data TLA Line data Source code
1 : : /*-------------------------------------------------------------------------
2 : : *
3 : : * trigger.c
4 : : * PostgreSQL TRIGGERs support code.
5 : : *
6 : : * Portions Copyright (c) 1996-2026, PostgreSQL Global Development Group
7 : : * Portions Copyright (c) 1994, Regents of the University of California
8 : : *
9 : : * IDENTIFICATION
10 : : * src/backend/commands/trigger.c
11 : : *
12 : : *-------------------------------------------------------------------------
13 : : */
14 : : #include "postgres.h"
15 : :
16 : : #include "access/genam.h"
17 : : #include "access/htup_details.h"
18 : : #include "access/relation.h"
19 : : #include "access/sysattr.h"
20 : : #include "access/table.h"
21 : : #include "access/tableam.h"
22 : : #include "access/tupconvert.h"
23 : : #include "access/xact.h"
24 : : #include "catalog/catalog.h"
25 : : #include "catalog/dependency.h"
26 : : #include "catalog/indexing.h"
27 : : #include "catalog/objectaccess.h"
28 : : #include "catalog/partition.h"
29 : : #include "catalog/pg_constraint.h"
30 : : #include "catalog/pg_inherits.h"
31 : : #include "catalog/pg_proc.h"
32 : : #include "catalog/pg_trigger.h"
33 : : #include "catalog/pg_type.h"
34 : : #include "commands/trigger.h"
35 : : #include "executor/executor.h"
36 : : #include "executor/instrument.h"
37 : : #include "miscadmin.h"
38 : : #include "nodes/bitmapset.h"
39 : : #include "nodes/makefuncs.h"
40 : : #include "optimizer/optimizer.h"
41 : : #include "parser/parse_clause.h"
42 : : #include "parser/parse_collate.h"
43 : : #include "parser/parse_func.h"
44 : : #include "parser/parse_relation.h"
45 : : #include "partitioning/partdesc.h"
46 : : #include "pgstat.h"
47 : : #include "rewrite/rewriteHandler.h"
48 : : #include "rewrite/rewriteManip.h"
49 : : #include "storage/lmgr.h"
50 : : #include "utils/acl.h"
51 : : #include "utils/builtins.h"
52 : : #include "utils/fmgroids.h"
53 : : #include "utils/guc_hooks.h"
54 : : #include "utils/inval.h"
55 : : #include "utils/lsyscache.h"
56 : : #include "utils/memutils.h"
57 : : #include "utils/plancache.h"
58 : : #include "utils/rel.h"
59 : : #include "utils/snapmgr.h"
60 : : #include "utils/syscache.h"
61 : : #include "utils/tuplestore.h"
62 : :
63 : :
64 : : /* GUC variables */
65 : : int SessionReplicationRole = SESSION_REPLICATION_ROLE_ORIGIN;
66 : :
67 : : /* How many levels deep into trigger execution are we? */
68 : : static int MyTriggerDepth = 0;
69 : :
70 : : /* Local function prototypes */
71 : : static void renametrig_internal(Relation tgrel, Relation targetrel,
72 : : HeapTuple trigtup, const char *newname,
73 : : const char *expected_name);
74 : : static void renametrig_partition(Relation tgrel, Oid partitionId,
75 : : Oid parentTriggerOid, const char *newname,
76 : : const char *expected_name);
77 : : static void SetTriggerFlags(TriggerDesc *trigdesc, Trigger *trigger);
78 : : static bool GetTupleForTrigger(EState *estate,
79 : : EPQState *epqstate,
80 : : ResultRelInfo *relinfo,
81 : : ItemPointer tid,
82 : : LockTupleMode lockmode,
83 : : TupleTableSlot *oldslot,
84 : : bool do_epq_recheck,
85 : : TupleTableSlot **epqslot,
86 : : TM_Result *tmresultp,
87 : : TM_FailureData *tmfdp);
88 : : static bool TriggerEnabled(EState *estate, ResultRelInfo *relinfo,
89 : : Trigger *trigger, TriggerEvent event,
90 : : Bitmapset *modifiedCols,
91 : : TupleTableSlot *oldslot, TupleTableSlot *newslot);
92 : : static HeapTuple ExecCallTriggerFunc(TriggerData *trigdata,
93 : : int tgindx,
94 : : FmgrInfo *finfo,
95 : : TriggerInstrumentation *instr,
96 : : MemoryContext per_tuple_context);
97 : : static void AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
98 : : ResultRelInfo *src_partinfo,
99 : : ResultRelInfo *dst_partinfo,
100 : : int event, bool row_trigger,
101 : : TupleTableSlot *oldslot, TupleTableSlot *newslot,
102 : : List *recheckIndexes, Bitmapset *modifiedCols,
103 : : TransitionCaptureState *transition_capture,
104 : : bool is_crosspart_update);
105 : : static void AfterTriggerEnlargeQueryState(void);
106 : : static bool before_stmt_triggers_fired(Oid relid, CmdType cmdType);
107 : : static HeapTuple check_modified_virtual_generated(TupleDesc tupdesc, HeapTuple tuple);
108 : :
109 : :
110 : : /*
111 : : * Create a trigger. Returns the address of the created trigger.
112 : : *
113 : : * queryString is the source text of the CREATE TRIGGER command.
114 : : * This must be supplied if a whenClause is specified, else it can be NULL.
115 : : *
116 : : * relOid, if nonzero, is the relation on which the trigger should be
117 : : * created. If zero, the name provided in the statement will be looked up.
118 : : *
119 : : * refRelOid, if nonzero, is the relation to which the constraint trigger
120 : : * refers. If zero, the constraint relation name provided in the statement
121 : : * will be looked up as needed.
122 : : *
123 : : * constraintOid, if nonzero, says that this trigger is being created
124 : : * internally to implement that constraint. A suitable pg_depend entry will
125 : : * be made to link the trigger to that constraint. constraintOid is zero when
126 : : * executing a user-entered CREATE TRIGGER command. (For CREATE CONSTRAINT
127 : : * TRIGGER, we build a pg_constraint entry internally.)
128 : : *
129 : : * indexOid, if nonzero, is the OID of an index associated with the constraint.
130 : : * We do nothing with this except store it into pg_trigger.tgconstrindid;
131 : : * but when creating a trigger for a deferrable unique constraint on a
132 : : * partitioned table, its children are looked up. Note we don't cope with
133 : : * invalid indexes in that case.
134 : : *
135 : : * funcoid, if nonzero, is the OID of the function to invoke. When this is
136 : : * given, stmt->funcname is ignored.
137 : : *
138 : : * parentTriggerOid, if nonzero, is a trigger that begets this one; so that
139 : : * if that trigger is dropped, this one should be too. There are two cases
140 : : * when a nonzero value is passed for this: 1) when this function recurses to
141 : : * create the trigger on partitions, 2) when creating child foreign key
142 : : * triggers; see CreateFKCheckTrigger() and createForeignKeyActionTriggers().
143 : : *
144 : : * If whenClause is passed, it is an already-transformed expression for
145 : : * WHEN. In this case, we ignore any that may come in stmt->whenClause.
146 : : *
147 : : * If isInternal is true then this is an internally-generated trigger.
148 : : * This argument sets the tgisinternal field of the pg_trigger entry, and
149 : : * if true causes us to modify the given trigger name to ensure uniqueness.
150 : : *
151 : : * When isInternal is not true we require ACL_TRIGGER permissions on the
152 : : * relation, as well as ACL_EXECUTE on the trigger function. For internal
153 : : * triggers the caller must apply any required permission checks.
154 : : *
155 : : * When called on partitioned tables, this function recurses to create the
156 : : * trigger on all the partitions, except if isInternal is true, in which
157 : : * case caller is expected to execute recursion on its own. in_partition
158 : : * indicates such a recursive call; outside callers should pass "false"
159 : : * (but see CloneRowTriggersToPartition).
160 : : */
161 : : ObjectAddress
47 peter@eisentraut.org 162 :GNC 10995 : CreateTrigger(const CreateTrigStmt *stmt, const char *queryString,
163 : : Oid relOid, Oid refRelOid, Oid constraintOid, Oid indexOid,
164 : : Oid funcoid, Oid parentTriggerOid, Node *whenClause,
165 : : bool isInternal, bool in_partition)
166 : : {
167 : : return
1754 alvherre@alvh.no-ip. 168 :CBC 10995 : CreateTriggerFiringOn(stmt, queryString, relOid, refRelOid,
169 : : constraintOid, indexOid, funcoid,
170 : : parentTriggerOid, whenClause, isInternal,
171 : : in_partition, TRIGGER_FIRES_ON_ORIGIN);
172 : : }
173 : :
174 : : /*
175 : : * Like the above; additionally the firing condition
176 : : * (always/origin/replica/disabled) can be specified.
177 : : */
178 : : ObjectAddress
47 peter@eisentraut.org 179 :GNC 11598 : CreateTriggerFiringOn(const CreateTrigStmt *stmt, const char *queryString,
180 : : Oid relOid, Oid refRelOid, Oid constraintOid,
181 : : Oid indexOid, Oid funcoid, Oid parentTriggerOid,
182 : : Node *whenClause, bool isInternal, bool in_partition,
183 : : char trigger_fires_when)
184 : : {
185 : : int16 tgtype;
186 : : int ncolumns;
187 : : int16 *columns;
188 : : int2vector *tgattr;
189 : : List *whenRtable;
190 : : char *qual;
191 : : Datum values[Natts_pg_trigger];
192 : : bool nulls[Natts_pg_trigger];
193 : : Relation rel;
194 : : AclResult aclresult;
195 : : Relation tgrel;
196 : : Relation pgrel;
1998 tgl@sss.pgh.pa.us 197 :CBC 11598 : HeapTuple tuple = NULL;
198 : : Oid funcrettype;
199 : 11598 : Oid trigoid = InvalidOid;
200 : : char internaltrigname[NAMEDATALEN];
201 : : char *trigname;
8615 202 : 11598 : Oid constrrelid = InvalidOid;
203 : : ObjectAddress myself,
204 : : referenced;
3469 kgrittn@postgresql.o 205 : 11598 : char *oldtablename = NULL;
206 : 11598 : char *newtablename = NULL;
207 : : bool partition_recurse;
1998 tgl@sss.pgh.pa.us 208 : 11598 : bool trigger_exists = false;
209 : 11598 : Oid existing_constraint_oid = InvalidOid;
210 : 11598 : bool existing_isInternal = false;
1581 alvherre@alvh.no-ip. 211 : 11598 : bool existing_isClone = false;
212 : :
4460 rhaas@postgresql.org 213 [ + + ]: 11598 : if (OidIsValid(relOid))
2661 andres@anarazel.de 214 : 9483 : rel = table_open(relOid, ShareRowExclusiveLock);
215 : : else
216 : 2115 : rel = table_openrv(stmt->relation, ShareRowExclusiveLock);
217 : :
218 : : /*
219 : : * Triggers must be on tables or views, and there are additional
220 : : * relation-type-specific restrictions.
221 : : */
2965 alvherre@alvh.no-ip. 222 [ + + ]: 11598 : if (rel->rd_rel->relkind == RELKIND_RELATION)
223 : : {
224 : : /* Tables can't have INSTEAD OF triggers */
5686 tgl@sss.pgh.pa.us 225 [ + + ]: 9530 : if (stmt->timing != TRIGGER_TYPE_BEFORE &&
226 [ + + ]: 8669 : stmt->timing != TRIGGER_TYPE_AFTER)
227 [ + - ]: 12 : ereport(ERROR,
228 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
229 : : errmsg("\"%s\" is a table",
230 : : RelationGetRelationName(rel)),
231 : : errdetail("Tables cannot have INSTEAD OF triggers.")));
232 : : }
2965 alvherre@alvh.no-ip. 233 [ + + ]: 2068 : else if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
234 : : {
235 : : /* Partitioned tables can't have INSTEAD OF triggers */
236 [ + + ]: 1871 : if (stmt->timing != TRIGGER_TYPE_BEFORE &&
237 [ + + ]: 1795 : stmt->timing != TRIGGER_TYPE_AFTER)
3436 rhaas@postgresql.org 238 [ + - ]: 4 : ereport(ERROR,
239 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
240 : : errmsg("\"%s\" is a table",
241 : : RelationGetRelationName(rel)),
242 : : errdetail("Tables cannot have INSTEAD OF triggers.")));
243 : :
244 : : /*
245 : : * FOR EACH ROW triggers have further restrictions
246 : : */
2965 alvherre@alvh.no-ip. 247 [ + + ]: 1867 : if (stmt->row)
248 : : {
249 : : /*
250 : : * Disallow use of transition tables.
251 : : *
252 : : * Note that we have another restriction about transition tables
253 : : * in partitions; search for 'has_superclass' below for an
254 : : * explanation. The check here is just to protect from the fact
255 : : * that if we allowed it here, the creation would succeed for a
256 : : * partitioned table with no partitions, but would be blocked by
257 : : * the other restriction when the first partition was created,
258 : : * which is very unfriendly behavior.
259 : : */
260 [ + + ]: 1703 : if (stmt->transitionRels != NIL)
261 [ + - ]: 4 : ereport(ERROR,
262 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
263 : : errmsg("\"%s\" is a partitioned table",
264 : : RelationGetRelationName(rel)),
265 : : errdetail("ROW triggers with transition tables are not supported on partitioned tables.")));
266 : : }
267 : : }
5686 tgl@sss.pgh.pa.us 268 [ + + ]: 197 : else if (rel->rd_rel->relkind == RELKIND_VIEW)
269 : : {
270 : : /*
271 : : * Views can have INSTEAD OF triggers (which we check below are
272 : : * row-level), or statement-level BEFORE/AFTER triggers.
273 : : */
274 [ + + + + ]: 139 : if (stmt->timing != TRIGGER_TYPE_INSTEAD && stmt->row)
275 [ + - ]: 24 : ereport(ERROR,
276 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
277 : : errmsg("\"%s\" is a view",
278 : : RelationGetRelationName(rel)),
279 : : errdetail("Views cannot have row-level BEFORE or AFTER triggers.")));
280 : : /* Disallow TRUNCATE triggers on VIEWs */
281 [ + + ]: 115 : if (TRIGGER_FOR_TRUNCATE(stmt->events))
282 [ + - ]: 8 : ereport(ERROR,
283 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
284 : : errmsg("\"%s\" is a view",
285 : : RelationGetRelationName(rel)),
286 : : errdetail("Views cannot have TRUNCATE triggers.")));
287 : : }
4426 noah@leadboat.com 288 [ + - ]: 58 : else if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
289 : : {
290 [ + + ]: 58 : if (stmt->timing != TRIGGER_TYPE_BEFORE &&
291 [ - + ]: 31 : stmt->timing != TRIGGER_TYPE_AFTER)
4426 noah@leadboat.com 292 [ # # ]:UBC 0 : ereport(ERROR,
293 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
294 : : errmsg("\"%s\" is a foreign table",
295 : : RelationGetRelationName(rel)),
296 : : errdetail("Foreign tables cannot have INSTEAD OF triggers.")));
297 : :
298 : : /*
299 : : * We disallow constraint triggers to protect the assumption that
300 : : * triggers on FKs can't be deferred. See notes with AfterTriggers
301 : : * data structures, below.
302 : : */
4426 noah@leadboat.com 303 [ + + ]:CBC 58 : if (stmt->isconstraint)
304 [ + - ]: 4 : ereport(ERROR,
305 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
306 : : errmsg("\"%s\" is a foreign table",
307 : : RelationGetRelationName(rel)),
308 : : errdetail("Foreign tables cannot have constraint triggers.")));
309 : : }
310 : : else
8325 tgl@sss.pgh.pa.us 311 [ # # ]:UBC 0 : ereport(ERROR,
312 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
313 : : errmsg("relation \"%s\" cannot have triggers",
314 : : RelationGetRelationName(rel)),
315 : : errdetail_relkind_not_supported(rel->rd_rel->relkind)));
316 : :
8789 tgl@sss.pgh.pa.us 317 [ + + + + ]:CBC 11542 : if (!allowSystemTableMods && IsSystemRelation(rel))
8325 318 [ + - ]: 1 : ereport(ERROR,
319 : : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
320 : : errmsg("permission denied: \"%s\" is a system catalog",
321 : : RelationGetRelationName(rel))));
322 : :
4460 rhaas@postgresql.org 323 [ + + ]: 11541 : if (stmt->isconstraint)
324 : : {
325 : : /*
326 : : * We must take a lock on the target relation to protect against
327 : : * concurrent drop. It's not clear that AccessShareLock is strong
328 : : * enough, but we certainly need at least that much... otherwise, we
329 : : * might end up creating a pg_constraint entry referencing a
330 : : * nonexistent table.
331 : : */
332 [ + + ]: 8997 : if (OidIsValid(refRelOid))
333 : : {
334 : 8802 : LockRelationOid(refRelOid, AccessShareLock);
335 : 8802 : constrrelid = refRelOid;
336 : : }
337 [ + + ]: 195 : else if (stmt->constrrel != NULL)
338 : 16 : constrrelid = RangeVarGetRelid(stmt->constrrel, AccessShareLock,
339 : : false);
340 : : }
341 : :
342 : : /* permission checks */
5952 tgl@sss.pgh.pa.us 343 [ + + ]: 11541 : if (!isInternal)
344 : : {
8564 bruce@momjian.us 345 : 2661 : aclresult = pg_class_aclcheck(RelationGetRelid(rel), GetUserId(),
346 : : ACL_TRIGGER);
8661 peter_e@gmx.net 347 [ - + ]: 2661 : if (aclresult != ACLCHECK_OK)
3076 peter_e@gmx.net 348 :UBC 0 : aclcheck_error(aclresult, get_relkind_objtype(rel->rd_rel->relkind),
8313 tgl@sss.pgh.pa.us 349 : 0 : RelationGetRelationName(rel));
350 : :
6312 tgl@sss.pgh.pa.us 351 [ + + ]:CBC 2661 : if (OidIsValid(constrrelid))
352 : : {
8564 bruce@momjian.us 353 : 28 : aclresult = pg_class_aclcheck(constrrelid, GetUserId(),
354 : : ACL_TRIGGER);
8661 peter_e@gmx.net 355 [ - + ]: 28 : if (aclresult != ACLCHECK_OK)
3076 peter_e@gmx.net 356 :UBC 0 : aclcheck_error(aclresult, get_relkind_objtype(get_rel_relkind(constrrelid)),
8313 tgl@sss.pgh.pa.us 357 : 0 : get_rel_name(constrrelid));
358 : : }
359 : : }
360 : :
361 : : /*
362 : : * When called on a partitioned table to create a FOR EACH ROW trigger
363 : : * that's not internal, we create one trigger for each partition, too.
364 : : *
365 : : * For that, we'd better hold lock on all of them ahead of time.
366 : : */
2965 alvherre@alvh.no-ip. 367 [ + + + + ]:CBC 13513 : partition_recurse = !isInternal && stmt->row &&
368 [ + + ]: 1972 : rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE;
369 [ + + ]: 11541 : if (partition_recurse)
370 : 279 : list_free(find_all_inheritors(RelationGetRelid(rel),
371 : : ShareRowExclusiveLock, NULL));
372 : :
373 : : /* Compute tgtype */
10467 bruce@momjian.us 374 : 11541 : TRIGGER_CLEAR_TYPE(tgtype);
375 [ + + ]: 11541 : if (stmt->row)
376 : 10852 : TRIGGER_SETT_ROW(tgtype);
5686 tgl@sss.pgh.pa.us 377 : 11541 : tgtype |= stmt->timing;
6165 378 : 11541 : tgtype |= stmt->events;
379 : :
380 : : /* Disallow ROW-level TRUNCATE triggers */
381 [ + + - + ]: 11541 : if (TRIGGER_FOR_ROW(tgtype) && TRIGGER_FOR_TRUNCATE(tgtype))
6165 tgl@sss.pgh.pa.us 382 [ # # ]:UBC 0 : ereport(ERROR,
383 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
384 : : errmsg("TRUNCATE FOR EACH ROW triggers are not supported")));
385 : :
386 : : /* INSTEAD triggers must be row-level, and can't have WHEN or columns */
5686 tgl@sss.pgh.pa.us 387 [ + + ]:CBC 11541 : if (TRIGGER_FOR_INSTEAD(tgtype))
388 : : {
389 [ + + ]: 79 : if (!TRIGGER_FOR_ROW(tgtype))
390 [ + - ]: 4 : ereport(ERROR,
391 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
392 : : errmsg("INSTEAD OF triggers must be FOR EACH ROW")));
393 [ + + ]: 75 : if (stmt->whenClause)
394 [ + - ]: 4 : ereport(ERROR,
395 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
396 : : errmsg("INSTEAD OF triggers cannot have WHEN conditions")));
397 [ + + ]: 71 : if (stmt->columns != NIL)
398 [ + - ]: 4 : ereport(ERROR,
399 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
400 : : errmsg("INSTEAD OF triggers cannot have column lists")));
401 : : }
402 : :
403 : : /*
404 : : * We don't yet support naming ROW transition variables, but the parser
405 : : * recognizes the syntax so we can give a nicer message here.
406 : : *
407 : : * Per standard, REFERENCING TABLE names are only allowed on AFTER
408 : : * triggers. Per standard, REFERENCING ROW names are not allowed with FOR
409 : : * EACH STATEMENT. Per standard, each OLD/NEW, ROW/TABLE permutation is
410 : : * only allowed once. Per standard, OLD may not be specified when
411 : : * creating a trigger only for INSERT, and NEW may not be specified when
412 : : * creating a trigger only for DELETE.
413 : : *
414 : : * Notice that the standard allows an AFTER ... FOR EACH ROW trigger to
415 : : * reference both ROW and TABLE transition data.
416 : : */
3469 kgrittn@postgresql.o 417 [ + + ]: 11529 : if (stmt->transitionRels != NIL)
418 : : {
419 : 324 : List *varList = stmt->transitionRels;
420 : : ListCell *lc;
421 : :
422 [ + - + + : 708 : foreach(lc, varList)
+ + ]
423 : : {
3275 bruce@momjian.us 424 : 416 : TriggerTransition *tt = lfirst_node(TriggerTransition, lc);
425 : :
3469 kgrittn@postgresql.o 426 [ - + ]: 416 : if (!(tt->isTable))
3469 kgrittn@postgresql.o 427 [ # # ]:UBC 0 : ereport(ERROR,
428 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
429 : : errmsg("ROW variable naming in the REFERENCING clause is not supported"),
430 : : errhint("Use OLD TABLE or NEW TABLE for naming transition tables.")));
431 : :
432 : : /*
433 : : * Because of the above test, we omit further ROW-related testing
434 : : * below. If we later allow naming OLD and NEW ROW variables,
435 : : * adjustments will be needed below.
436 : : */
437 : :
3283 rhaas@postgresql.org 438 [ + + ]:CBC 416 : if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
439 [ + - ]: 4 : ereport(ERROR,
440 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
441 : : errmsg("\"%s\" is a foreign table",
442 : : RelationGetRelationName(rel)),
443 : : errdetail("Triggers on foreign tables cannot have transition tables.")));
444 : :
445 [ + + ]: 412 : if (rel->rd_rel->relkind == RELKIND_VIEW)
446 [ + - ]: 4 : ereport(ERROR,
447 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
448 : : errmsg("\"%s\" is a view",
449 : : RelationGetRelationName(rel)),
450 : : errdetail("Triggers on views cannot have transition tables.")));
451 : :
452 : : /*
453 : : * We currently don't allow row-level triggers with transition
454 : : * tables on partition or inheritance children. Such triggers
455 : : * would somehow need to see tuples converted to the format of the
456 : : * table they're attached to, and it's not clear which subset of
457 : : * tuples each child should see. See also the prohibitions in
458 : : * ATExecAttachPartition() and ATExecAddInherit().
459 : : */
3233 rhodiumtoad@postgres 460 [ + + + + ]: 408 : if (TRIGGER_FOR_ROW(tgtype) && has_superclass(rel->rd_id))
461 : : {
462 : : /* Use appropriate error message. */
463 [ + + ]: 8 : if (rel->rd_rel->relispartition)
464 [ + - ]: 4 : ereport(ERROR,
465 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
466 : : errmsg("ROW triggers with transition tables are not supported on partitions")));
467 : : else
468 [ + - ]: 4 : ereport(ERROR,
469 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
470 : : errmsg("ROW triggers with transition tables are not supported on inheritance children")));
471 : : }
472 : :
3469 kgrittn@postgresql.o 473 [ - + ]: 400 : if (stmt->timing != TRIGGER_TYPE_AFTER)
3469 kgrittn@postgresql.o 474 [ # # ]:UBC 0 : ereport(ERROR,
475 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
476 : : errmsg("transition table name can only be specified for an AFTER trigger")));
477 : :
3283 rhaas@postgresql.org 478 [ + + ]:CBC 400 : if (TRIGGER_FOR_TRUNCATE(tgtype))
479 [ + - ]: 4 : ereport(ERROR,
480 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
481 : : errmsg("TRUNCATE triggers with transition tables are not supported")));
482 : :
483 : : /*
484 : : * We currently don't allow multi-event triggers ("INSERT OR
485 : : * UPDATE") with transition tables, because it's not clear how to
486 : : * handle INSERT ... ON CONFLICT statements which can fire both
487 : : * INSERT and UPDATE triggers. We show the inserted tuples to
488 : : * INSERT triggers and the updated tuples to UPDATE triggers, but
489 : : * it's not yet clear what INSERT OR UPDATE trigger should see.
490 : : * This restriction could be lifted if we can decide on the right
491 : : * semantics in a later release.
492 : : */
3233 rhodiumtoad@postgres 493 : 396 : if (((TRIGGER_FOR_INSERT(tgtype) ? 1 : 0) +
494 : 396 : (TRIGGER_FOR_UPDATE(tgtype) ? 1 : 0) +
495 [ + + ]: 396 : (TRIGGER_FOR_DELETE(tgtype) ? 1 : 0)) != 1)
496 [ + - ]: 4 : ereport(ERROR,
497 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
498 : : errmsg("transition tables cannot be specified for triggers with more than one event")));
499 : :
500 : : /*
501 : : * We currently don't allow column-specific triggers with
502 : : * transition tables. Per spec, that seems to require
503 : : * accumulating separate transition tables for each combination of
504 : : * columns, which is a lot of work for a rather marginal feature.
505 : : */
3153 tgl@sss.pgh.pa.us 506 [ + + ]: 392 : if (stmt->columns != NIL)
507 [ + - ]: 4 : ereport(ERROR,
508 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
509 : : errmsg("transition tables cannot be specified for triggers with column lists")));
510 : :
511 : : /*
512 : : * We disallow constraint triggers with transition tables, to
513 : : * protect the assumption that such triggers can't be deferred.
514 : : * See notes with AfterTriggers data structures, below.
515 : : *
516 : : * Currently this is enforced by the grammar, so just Assert here.
517 : : */
518 [ - + ]: 388 : Assert(!stmt->isconstraint);
519 : :
3469 kgrittn@postgresql.o 520 [ + + ]: 388 : if (tt->isNew)
521 : : {
522 [ + + ]: 204 : if (!(TRIGGER_FOR_INSERT(tgtype) ||
523 [ - + ]: 109 : TRIGGER_FOR_UPDATE(tgtype)))
3469 kgrittn@postgresql.o 524 [ # # ]:UBC 0 : ereport(ERROR,
525 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
526 : : errmsg("NEW TABLE can only be specified for an INSERT or UPDATE trigger")));
527 : :
3469 kgrittn@postgresql.o 528 [ - + ]:CBC 204 : if (newtablename != NULL)
3469 kgrittn@postgresql.o 529 [ # # ]:UBC 0 : ereport(ERROR,
530 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
531 : : errmsg("NEW TABLE cannot be specified multiple times")));
532 : :
3469 kgrittn@postgresql.o 533 :CBC 204 : newtablename = tt->name;
534 : : }
535 : : else
536 : : {
537 [ + + ]: 184 : if (!(TRIGGER_FOR_DELETE(tgtype) ||
538 [ + + ]: 105 : TRIGGER_FOR_UPDATE(tgtype)))
539 [ + - ]: 4 : ereport(ERROR,
540 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
541 : : errmsg("OLD TABLE can only be specified for a DELETE or UPDATE trigger")));
542 : :
543 [ - + ]: 180 : if (oldtablename != NULL)
3469 kgrittn@postgresql.o 544 [ # # ]:UBC 0 : ereport(ERROR,
545 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
546 : : errmsg("OLD TABLE cannot be specified multiple times")));
547 : :
3469 kgrittn@postgresql.o 548 :CBC 180 : oldtablename = tt->name;
549 : : }
550 : : }
551 : :
552 [ + + + + ]: 292 : if (newtablename != NULL && oldtablename != NULL &&
553 [ - + ]: 92 : strcmp(newtablename, oldtablename) == 0)
3469 kgrittn@postgresql.o 554 [ # # ]:UBC 0 : ereport(ERROR,
555 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
556 : : errmsg("OLD TABLE name and NEW TABLE name cannot be the same")));
557 : : }
558 : :
559 : : /*
560 : : * Parse the WHEN clause, if any and we weren't passed an already
561 : : * transformed one.
562 : : *
563 : : * Note that as a side effect, we fill whenRtable when parsing. If we got
564 : : * an already parsed clause, this does not occur, which is what we want --
565 : : * no point in adding redundant dependencies below.
566 : : */
2965 alvherre@alvh.no-ip. 567 [ + + + + ]:CBC 11497 : if (!whenClause && stmt->whenClause)
6010 tgl@sss.pgh.pa.us 568 : 88 : {
569 : : ParseState *pstate;
570 : : ParseNamespaceItem *nsitem;
571 : : List *varList;
572 : : ListCell *lc;
573 : :
574 : : /* Set up a pstate to parse with */
575 : 120 : pstate = make_parsestate(NULL);
576 : 120 : pstate->p_sourcetext = queryString;
577 : :
578 : : /*
579 : : * Set up nsitems for OLD and NEW references.
580 : : *
581 : : * 'OLD' must always have varno equal to 1 and 'NEW' equal to 2.
582 : : */
2315 583 : 120 : nsitem = addRangeTableEntryForRelation(pstate, rel,
584 : : AccessShareLock,
585 : : makeAlias("old", NIL),
586 : : false, false);
587 : 120 : addNSItemToQuery(pstate, nsitem, false, true, true);
588 : 120 : nsitem = addRangeTableEntryForRelation(pstate, rel,
589 : : AccessShareLock,
590 : : makeAlias("new", NIL),
591 : : false, false);
592 : 120 : addNSItemToQuery(pstate, nsitem, false, true, true);
593 : :
594 : : /* Transform expression. Copy to be sure we don't modify original */
6010 595 : 120 : whenClause = transformWhereClause(pstate,
596 : 120 : copyObject(stmt->whenClause),
597 : : EXPR_KIND_TRIGGER_WHEN,
598 : : "WHEN");
599 : : /* we have to fix its collations too */
5507 600 : 120 : assign_expr_collations(pstate, whenClause);
601 : :
602 : : /*
603 : : * Check for disallowed references to OLD/NEW.
604 : : *
605 : : * NB: pull_var_clause is okay here only because we don't allow
606 : : * subselects in WHEN clauses; it would fail to examine the contents
607 : : * of subselects.
608 : : */
3708 609 : 120 : varList = pull_var_clause(whenClause, 0);
6010 610 [ + + + + : 237 : foreach(lc, varList)
+ + ]
611 : : {
612 : 149 : Var *var = (Var *) lfirst(lc);
613 : :
614 [ + + - ]: 149 : switch (var->varno)
615 : : {
616 : 56 : case PRS2_OLD_VARNO:
617 [ + + ]: 56 : if (!TRIGGER_FOR_ROW(tgtype))
618 [ + - ]: 4 : ereport(ERROR,
619 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
620 : : errmsg("statement trigger's WHEN condition cannot reference column values"),
621 : : parser_errposition(pstate, var->location)));
622 [ + + ]: 52 : if (TRIGGER_FOR_INSERT(tgtype))
623 [ + - ]: 4 : ereport(ERROR,
624 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
625 : : errmsg("INSERT trigger's WHEN condition cannot reference OLD values"),
626 : : parser_errposition(pstate, var->location)));
627 : : /* system columns are okay here */
628 : 48 : break;
629 : 93 : case PRS2_NEW_VARNO:
630 [ - + ]: 93 : if (!TRIGGER_FOR_ROW(tgtype))
6010 tgl@sss.pgh.pa.us 631 [ # # ]:UBC 0 : ereport(ERROR,
632 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
633 : : errmsg("statement trigger's WHEN condition cannot reference column values"),
634 : : parser_errposition(pstate, var->location)));
6010 tgl@sss.pgh.pa.us 635 [ + + ]:CBC 93 : if (TRIGGER_FOR_DELETE(tgtype))
636 [ + - ]: 4 : ereport(ERROR,
637 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
638 : : errmsg("DELETE trigger's WHEN condition cannot reference NEW values"),
639 : : parser_errposition(pstate, var->location)));
640 [ + + + + ]: 89 : if (var->varattno < 0 && TRIGGER_FOR_BEFORE(tgtype))
641 [ + - ]: 4 : ereport(ERROR,
642 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
643 : : errmsg("BEFORE trigger's WHEN condition cannot reference NEW system columns"),
644 : : parser_errposition(pstate, var->location)));
2593 peter@eisentraut.org 645 [ + + ]: 85 : if (TRIGGER_FOR_BEFORE(tgtype) &&
646 [ + + ]: 34 : var->varattno == 0 &&
647 [ + + ]: 12 : RelationGetDescr(rel)->constr &&
452 648 [ + + ]: 8 : (RelationGetDescr(rel)->constr->has_generated_stored ||
649 [ + - ]: 4 : RelationGetDescr(rel)->constr->has_generated_virtual))
2593 650 [ + - ]: 8 : ereport(ERROR,
651 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
652 : : errmsg("BEFORE trigger's WHEN condition cannot reference NEW generated columns"),
653 : : errdetail("A whole-row reference is used and the table contains generated columns."),
654 : : parser_errposition(pstate, var->location)));
655 [ + + ]: 77 : if (TRIGGER_FOR_BEFORE(tgtype) &&
656 [ + + ]: 26 : var->varattno > 0 &&
657 [ + + ]: 22 : TupleDescAttr(RelationGetDescr(rel), var->varattno - 1)->attgenerated)
658 [ + - ]: 8 : ereport(ERROR,
659 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
660 : : errmsg("BEFORE trigger's WHEN condition cannot reference NEW generated columns"),
661 : : errdetail("Column \"%s\" is a generated column.",
662 : : NameStr(TupleDescAttr(RelationGetDescr(rel), var->varattno - 1)->attname)),
663 : : parser_errposition(pstate, var->location)));
6010 tgl@sss.pgh.pa.us 664 : 69 : break;
6010 tgl@sss.pgh.pa.us 665 :UBC 0 : default:
666 : : /* can't happen without add_missing_from, so just elog */
667 [ # # ]: 0 : elog(ERROR, "trigger WHEN condition cannot contain references to other relations");
668 : : break;
669 : : }
670 : : }
671 : :
672 : : /* we'll need the rtable for recordDependencyOnExpr */
6010 tgl@sss.pgh.pa.us 673 :CBC 88 : whenRtable = pstate->p_rtable;
674 : :
675 : 88 : qual = nodeToString(whenClause);
676 : :
677 : 88 : free_parsestate(pstate);
678 : : }
2965 alvherre@alvh.no-ip. 679 [ + + ]: 11377 : else if (!whenClause)
680 : : {
6010 tgl@sss.pgh.pa.us 681 : 11349 : whenClause = NULL;
682 : 11349 : whenRtable = NIL;
683 : 11349 : qual = NULL;
684 : : }
685 : : else
686 : : {
2965 alvherre@alvh.no-ip. 687 : 28 : qual = nodeToString(whenClause);
688 : 28 : whenRtable = NIL;
689 : : }
690 : :
691 : : /*
692 : : * Find and validate the trigger function.
693 : : */
694 [ + + ]: 11465 : if (!OidIsValid(funcoid))
2366 695 : 10862 : funcoid = LookupFuncName(stmt->funcname, 0, NULL, false);
5185 tgl@sss.pgh.pa.us 696 [ + + ]: 11465 : if (!isInternal)
697 : : {
1269 peter@eisentraut.org 698 : 2585 : aclresult = object_aclcheck(ProcedureRelationId, funcoid, GetUserId(), ACL_EXECUTE);
5185 tgl@sss.pgh.pa.us 699 [ - + ]: 2585 : if (aclresult != ACLCHECK_OK)
3076 peter_e@gmx.net 700 :UBC 0 : aclcheck_error(aclresult, OBJECT_FUNCTION,
5185 tgl@sss.pgh.pa.us 701 : 0 : NameListToString(stmt->funcname));
702 : : }
6757 tgl@sss.pgh.pa.us 703 :CBC 11465 : funcrettype = get_func_rettype(funcoid);
704 [ - + ]: 11465 : if (funcrettype != TRIGGEROID)
2252 tgl@sss.pgh.pa.us 705 [ # # ]:UBC 0 : ereport(ERROR,
706 : : (errcode(ERRCODE_INVALID_OBJECT_DEFINITION),
707 : : errmsg("function %s must return type %s",
708 : : NameListToString(stmt->funcname), "trigger")));
709 : :
710 : : /*
711 : : * Scan pg_trigger to see if there is already a trigger of the same name.
712 : : * Skip this for internally generated triggers, since we'll modify the
713 : : * name to be unique below.
714 : : *
715 : : * NOTE that this is cool only because we have ShareRowExclusiveLock on
716 : : * the relation, so the trigger set won't be changing underneath us.
717 : : */
1998 tgl@sss.pgh.pa.us 718 :CBC 11465 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
719 [ + + ]: 11465 : if (!isInternal)
720 : : {
721 : : ScanKeyData skeys[2];
722 : : SysScanDesc tgscan;
723 : :
724 : 2585 : ScanKeyInit(&skeys[0],
725 : : Anum_pg_trigger_tgrelid,
726 : : BTEqualStrategyNumber, F_OIDEQ,
727 : : ObjectIdGetDatum(RelationGetRelid(rel)));
728 : :
729 : 2585 : ScanKeyInit(&skeys[1],
730 : : Anum_pg_trigger_tgname,
731 : : BTEqualStrategyNumber, F_NAMEEQ,
732 : 2585 : CStringGetDatum(stmt->trigname));
733 : :
734 : 2585 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
735 : : NULL, 2, skeys);
736 : :
737 : : /* There should be at most one matching tuple */
738 [ + + ]: 2585 : if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
739 : : {
740 : 68 : Form_pg_trigger oldtrigger = (Form_pg_trigger) GETSTRUCT(tuple);
741 : :
742 : 68 : trigoid = oldtrigger->oid;
743 : 68 : existing_constraint_oid = oldtrigger->tgconstraint;
744 : 68 : existing_isInternal = oldtrigger->tgisinternal;
1581 alvherre@alvh.no-ip. 745 : 68 : existing_isClone = OidIsValid(oldtrigger->tgparentid);
1998 tgl@sss.pgh.pa.us 746 : 68 : trigger_exists = true;
747 : : /* copy the tuple to use in CatalogTupleUpdate() */
748 : 68 : tuple = heap_copytuple(tuple);
749 : : }
750 : 2585 : systable_endscan(tgscan);
751 : : }
752 : :
753 [ + + ]: 11465 : if (!trigger_exists)
754 : : {
755 : : /* Generate the OID for the new trigger. */
756 : 11397 : trigoid = GetNewOidWithIndex(tgrel, TriggerOidIndexId,
757 : : Anum_pg_trigger_oid);
758 : : }
759 : : else
760 : : {
761 : : /*
762 : : * If OR REPLACE was specified, we'll replace the old trigger;
763 : : * otherwise complain about the duplicate name.
764 : : */
765 [ + + ]: 68 : if (!stmt->replace)
766 [ + - ]: 12 : ereport(ERROR,
767 : : (errcode(ERRCODE_DUPLICATE_OBJECT),
768 : : errmsg("trigger \"%s\" for relation \"%s\" already exists",
769 : : stmt->trigname, RelationGetRelationName(rel))));
770 : :
771 : : /*
772 : : * An internal trigger or a child trigger (isClone) cannot be replaced
773 : : * by a user-defined trigger. However, skip this test when
774 : : * in_partition, because then we're recursing from a partitioned table
775 : : * and the check was made at the parent level.
776 : : */
1581 alvherre@alvh.no-ip. 777 [ + - + + ]: 56 : if ((existing_isInternal || existing_isClone) &&
778 [ + - + + ]: 40 : !isInternal && !in_partition)
1998 tgl@sss.pgh.pa.us 779 [ + - ]: 4 : ereport(ERROR,
780 : : (errcode(ERRCODE_DUPLICATE_OBJECT),
781 : : errmsg("trigger \"%s\" for relation \"%s\" is an internal or a child trigger",
782 : : stmt->trigname, RelationGetRelationName(rel))));
783 : :
784 : : /*
785 : : * It is not allowed to replace with a constraint trigger; gram.y
786 : : * should have enforced this already.
787 : : */
788 [ - + ]: 52 : Assert(!stmt->isconstraint);
789 : :
790 : : /*
791 : : * It is not allowed to replace an existing constraint trigger,
792 : : * either. (The reason for these restrictions is partly that it seems
793 : : * difficult to deal with pending trigger events in such cases, and
794 : : * partly that the command might imply changing the constraint's
795 : : * properties as well, which doesn't seem nice.)
796 : : */
797 [ - + ]: 52 : if (OidIsValid(existing_constraint_oid))
1998 tgl@sss.pgh.pa.us 798 [ # # ]:UBC 0 : ereport(ERROR,
799 : : (errcode(ERRCODE_DUPLICATE_OBJECT),
800 : : errmsg("trigger \"%s\" for relation \"%s\" is a constraint trigger",
801 : : stmt->trigname, RelationGetRelationName(rel))));
802 : : }
803 : :
804 : : /*
805 : : * If it's a user-entered CREATE CONSTRAINT TRIGGER command, make a
806 : : * corresponding pg_constraint entry.
807 : : */
5952 tgl@sss.pgh.pa.us 808 [ + + + + ]:CBC 11449 : if (stmt->isconstraint && !OidIsValid(constraintOid))
809 : : {
810 : : /* Internal callers should have made their own constraints */
811 [ - + ]: 117 : Assert(!isInternal);
812 : 117 : constraintOid = CreateConstraintEntry(stmt->trigname,
813 : 117 : RelationGetNamespace(rel),
814 : : CONSTRAINT_TRIGGER,
815 : 117 : stmt->deferrable,
816 : 117 : stmt->initdeferred,
817 : : true, /* Is Enforced */
818 : : true,
819 : : InvalidOid, /* no parent */
820 : : RelationGetRelid(rel),
821 : : NULL, /* no conkey */
822 : : 0,
823 : : 0,
824 : : InvalidOid, /* no domain */
825 : : InvalidOid, /* no index */
826 : : InvalidOid, /* no foreign key */
827 : : NULL,
828 : : NULL,
829 : : NULL,
830 : : NULL,
831 : : 0,
832 : : ' ',
833 : : ' ',
834 : : NULL,
835 : : 0,
836 : : ' ',
837 : : NULL, /* no exclusion */
838 : : NULL, /* no check constraint */
839 : : NULL,
840 : : true, /* islocal */
841 : : 0, /* inhcount */
842 : : true, /* noinherit */
843 : : false, /* conperiod */
844 : : isInternal); /* is_internal */
845 : : }
846 : :
847 : : /*
848 : : * If trigger is internally generated, modify the provided trigger name to
849 : : * ensure uniqueness by appending the trigger OID. (Callers will usually
850 : : * supply a simple constant trigger name in these cases.)
851 : : */
852 [ + + ]: 11449 : if (isInternal)
853 : : {
854 : 8880 : snprintf(internaltrigname, sizeof(internaltrigname),
5952 tgl@sss.pgh.pa.us 855 :GIC 8880 : "%s_%u", stmt->trigname, trigoid);
5952 tgl@sss.pgh.pa.us 856 :CBC 8880 : trigname = internaltrigname;
857 : : }
858 : : else
859 : : {
860 : : /* user-defined trigger; use the specified trigger name as-is */
6757 861 : 2569 : trigname = stmt->trigname;
862 : : }
863 : :
864 : : /*
865 : : * Build the new pg_trigger tuple.
866 : : */
6393 867 : 11449 : memset(nulls, false, sizeof(nulls));
868 : :
2723 andres@anarazel.de 869 : 11449 : values[Anum_pg_trigger_oid - 1] = ObjectIdGetDatum(trigoid);
10121 bruce@momjian.us 870 : 11449 : values[Anum_pg_trigger_tgrelid - 1] = ObjectIdGetDatum(RelationGetRelid(rel));
2259 alvherre@alvh.no-ip. 871 : 11449 : values[Anum_pg_trigger_tgparentid - 1] = ObjectIdGetDatum(parentTriggerOid);
9406 tgl@sss.pgh.pa.us 872 : 11449 : values[Anum_pg_trigger_tgname - 1] = DirectFunctionCall1(namein,
873 : : CStringGetDatum(trigname));
9437 874 : 11449 : values[Anum_pg_trigger_tgfoid - 1] = ObjectIdGetDatum(funcoid);
10467 bruce@momjian.us 875 : 11449 : values[Anum_pg_trigger_tgtype - 1] = Int16GetDatum(tgtype);
270 peter@eisentraut.org 876 :GNC 11449 : values[Anum_pg_trigger_tgenabled - 1] = CharGetDatum(trigger_fires_when);
1581 alvherre@alvh.no-ip. 877 :CBC 11449 : values[Anum_pg_trigger_tgisinternal - 1] = BoolGetDatum(isInternal);
9437 tgl@sss.pgh.pa.us 878 : 11449 : values[Anum_pg_trigger_tgconstrrelid - 1] = ObjectIdGetDatum(constrrelid);
6125 879 : 11449 : values[Anum_pg_trigger_tgconstrindid - 1] = ObjectIdGetDatum(indexOid);
7020 880 : 11449 : values[Anum_pg_trigger_tgconstraint - 1] = ObjectIdGetDatum(constraintOid);
9437 881 : 11449 : values[Anum_pg_trigger_tgdeferrable - 1] = BoolGetDatum(stmt->deferrable);
882 : 11449 : values[Anum_pg_trigger_tginitdeferred - 1] = BoolGetDatum(stmt->initdeferred);
883 : :
10467 bruce@momjian.us 884 [ + + ]: 11449 : if (stmt->args)
885 : : {
886 : : ListCell *le;
887 : : char *args;
8014 neilc@samurai.com 888 : 457 : int16 nargs = list_length(stmt->args);
10466 bruce@momjian.us 889 : 457 : int len = 0;
890 : :
10467 891 [ + - + + : 1088 : foreach(le, stmt->args)
+ + ]
892 : : {
8654 tgl@sss.pgh.pa.us 893 : 631 : char *ar = strVal(lfirst(le));
894 : :
9437 895 : 631 : len += strlen(ar) + 4;
10416 vadim4o@yahoo.com 896 [ + + ]: 4992 : for (; *ar; ar++)
897 : : {
10442 898 [ - + ]: 4361 : if (*ar == '\\')
10442 vadim4o@yahoo.com 899 :UBC 0 : len++;
900 : : }
901 : : }
10467 bruce@momjian.us 902 :CBC 457 : args = (char *) palloc(len + 1);
9398 tgl@sss.pgh.pa.us 903 : 457 : args[0] = '\0';
10467 bruce@momjian.us 904 [ + - + + : 1088 : foreach(le, stmt->args)
+ + ]
905 : : {
8654 tgl@sss.pgh.pa.us 906 : 631 : char *s = strVal(lfirst(le));
10416 vadim4o@yahoo.com 907 : 631 : char *d = args + strlen(args);
908 : :
10442 909 [ + + ]: 4992 : while (*s)
910 : : {
911 [ - + ]: 4361 : if (*s == '\\')
10442 vadim4o@yahoo.com 912 :UBC 0 : *d++ = '\\';
10442 vadim4o@yahoo.com 913 :CBC 4361 : *d++ = *s++;
914 : : }
9398 tgl@sss.pgh.pa.us 915 : 631 : strcpy(d, "\\000");
916 : : }
10467 bruce@momjian.us 917 : 457 : values[Anum_pg_trigger_tgnargs - 1] = Int16GetDatum(nargs);
9411 tgl@sss.pgh.pa.us 918 : 457 : values[Anum_pg_trigger_tgargs - 1] = DirectFunctionCall1(byteain,
919 : : CStringGetDatum(args));
920 : : }
921 : : else
922 : : {
10467 bruce@momjian.us 923 : 10992 : values[Anum_pg_trigger_tgnargs - 1] = Int16GetDatum(0);
9411 tgl@sss.pgh.pa.us 924 : 10992 : values[Anum_pg_trigger_tgargs - 1] = DirectFunctionCall1(byteain,
925 : : CStringGetDatum(""));
926 : : }
927 : :
928 : : /* build column number array if it's a column-specific trigger */
6047 929 : 11449 : ncolumns = list_length(stmt->columns);
930 [ + + ]: 11449 : if (ncolumns == 0)
931 : 11380 : columns = NULL;
932 : : else
933 : : {
934 : : ListCell *cell;
935 : 69 : int i = 0;
936 : :
5062 peter_e@gmx.net 937 : 69 : columns = (int16 *) palloc(ncolumns * sizeof(int16));
6047 tgl@sss.pgh.pa.us 938 [ + - + + : 143 : foreach(cell, stmt->columns)
+ + ]
939 : : {
5912 bruce@momjian.us 940 : 78 : char *name = strVal(lfirst(cell));
941 : : int16 attnum;
942 : : int j;
943 : :
944 : : /* Lookup column name. System columns are not allowed */
6047 tgl@sss.pgh.pa.us 945 : 78 : attnum = attnameAttNum(rel, name, false);
946 [ - + ]: 78 : if (attnum == InvalidAttrNumber)
6047 tgl@sss.pgh.pa.us 947 [ # # ]:UBC 0 : ereport(ERROR,
948 : : (errcode(ERRCODE_UNDEFINED_COLUMN),
949 : : errmsg("column \"%s\" of relation \"%s\" does not exist",
950 : : name, RelationGetRelationName(rel))));
951 : :
952 : : /* Check for duplicates */
6047 tgl@sss.pgh.pa.us 953 [ + + ]:CBC 83 : for (j = i - 1; j >= 0; j--)
954 : : {
955 [ + + ]: 9 : if (columns[j] == attnum)
956 [ + - ]: 4 : ereport(ERROR,
957 : : (errcode(ERRCODE_DUPLICATE_COLUMN),
958 : : errmsg("column \"%s\" specified more than once",
959 : : name)));
960 : : }
961 : :
962 : 74 : columns[i++] = attnum;
963 : : }
964 : : }
965 : 11445 : tgattr = buildint2vector(columns, ncolumns);
10467 bruce@momjian.us 966 : 11445 : values[Anum_pg_trigger_tgattr - 1] = PointerGetDatum(tgattr);
967 : :
968 : : /* set tgqual if trigger has WHEN clause */
6010 tgl@sss.pgh.pa.us 969 [ + + ]: 11445 : if (qual)
970 : 116 : values[Anum_pg_trigger_tgqual - 1] = CStringGetTextDatum(qual);
971 : : else
972 : 11329 : nulls[Anum_pg_trigger_tgqual - 1] = true;
973 : :
3469 kgrittn@postgresql.o 974 [ + + ]: 11445 : if (oldtablename)
975 : 180 : values[Anum_pg_trigger_tgoldtable - 1] = DirectFunctionCall1(namein,
976 : : CStringGetDatum(oldtablename));
977 : : else
978 : 11265 : nulls[Anum_pg_trigger_tgoldtable - 1] = true;
979 [ + + ]: 11445 : if (newtablename)
980 : 204 : values[Anum_pg_trigger_tgnewtable - 1] = DirectFunctionCall1(namein,
981 : : CStringGetDatum(newtablename));
982 : : else
983 : 11241 : nulls[Anum_pg_trigger_tgnewtable - 1] = true;
984 : :
985 : : /*
986 : : * Insert or replace tuple in pg_trigger.
987 : : */
1998 tgl@sss.pgh.pa.us 988 [ + + ]: 11445 : if (!trigger_exists)
989 : : {
990 : 11393 : tuple = heap_form_tuple(tgrel->rd_att, values, nulls);
991 : 11393 : CatalogTupleInsert(tgrel, tuple);
992 : : }
993 : : else
994 : : {
995 : : HeapTuple newtup;
996 : :
997 : 52 : newtup = heap_form_tuple(tgrel->rd_att, values, nulls);
998 : 52 : CatalogTupleUpdate(tgrel, &tuple->t_self, newtup);
999 : 52 : heap_freetuple(newtup);
1000 : : }
1001 : :
1002 : 11445 : heap_freetuple(tuple); /* free either original or new tuple */
2661 andres@anarazel.de 1003 : 11445 : table_close(tgrel, RowExclusiveLock);
1004 : :
10467 bruce@momjian.us 1005 : 11445 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgname - 1]));
1006 : 11445 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgargs - 1]));
6047 tgl@sss.pgh.pa.us 1007 : 11445 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgattr - 1]));
3469 kgrittn@postgresql.o 1008 [ + + ]: 11445 : if (oldtablename)
1009 : 180 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgoldtable - 1]));
1010 [ + + ]: 11445 : if (newtablename)
1011 : 204 : pfree(DatumGetPointer(values[Anum_pg_trigger_tgnewtable - 1]));
1012 : :
1013 : : /*
1014 : : * Update relation's pg_class entry; if necessary; and if not, send an SI
1015 : : * message to make other backends (and this one) rebuild relcache entries.
1016 : : */
2661 andres@anarazel.de 1017 : 11445 : pgrel = table_open(RelationRelationId, RowExclusiveLock);
5924 rhaas@postgresql.org 1018 : 11445 : tuple = SearchSysCacheCopy1(RELOID,
1019 : : ObjectIdGetDatum(RelationGetRelid(rel)));
10121 bruce@momjian.us 1020 [ - + ]: 11445 : if (!HeapTupleIsValid(tuple))
8325 tgl@sss.pgh.pa.us 1021 [ # # ]:UBC 0 : elog(ERROR, "cache lookup failed for relation %u",
1022 : : RelationGetRelid(rel));
2965 alvherre@alvh.no-ip. 1023 [ + + ]:CBC 11445 : if (!((Form_pg_class) GETSTRUCT(tuple))->relhastriggers)
1024 : : {
1025 : 4258 : ((Form_pg_class) GETSTRUCT(tuple))->relhastriggers = true;
1026 : :
1027 : 4258 : CatalogTupleUpdate(pgrel, &tuple->t_self, tuple);
1028 : :
1029 : 4258 : CommandCounterIncrement();
1030 : : }
1031 : : else
1032 : 7187 : CacheInvalidateRelcacheByTuple(tuple);
1033 : :
9637 JanWieck@Yahoo.com 1034 : 11445 : heap_freetuple(tuple);
2661 andres@anarazel.de 1035 : 11445 : table_close(pgrel, RowExclusiveLock);
1036 : :
1037 : : /*
1038 : : * If we're replacing a trigger, flush all the old dependencies before
1039 : : * recording new ones.
1040 : : */
1998 tgl@sss.pgh.pa.us 1041 [ + + ]: 11445 : if (trigger_exists)
1042 : 52 : deleteDependencyRecordsFor(TriggerRelationId, trigoid, true);
1043 : :
1044 : : /*
1045 : : * Record dependencies for trigger. Always place a normal dependency on
1046 : : * the function.
1047 : : */
7020 1048 : 11445 : myself.classId = TriggerRelationId;
1049 : 11445 : myself.objectId = trigoid;
1050 : 11445 : myself.objectSubId = 0;
1051 : :
7691 1052 : 11445 : referenced.classId = ProcedureRelationId;
8698 1053 : 11445 : referenced.objectId = funcoid;
1054 : 11445 : referenced.objectSubId = 0;
1055 : 11445 : recordDependencyOn(&myself, &referenced, DEPENDENCY_NORMAL);
1056 : :
5952 1057 [ + + + - ]: 11445 : if (isInternal && OidIsValid(constraintOid))
1058 : : {
1059 : : /*
1060 : : * Internally-generated trigger for a constraint, so make it an
1061 : : * internal dependency of the constraint. We can skip depending on
1062 : : * the relation(s), as there'll be an indirect dependency via the
1063 : : * constraint.
1064 : : */
7020 1065 : 8880 : referenced.classId = ConstraintRelationId;
1066 : 8880 : referenced.objectId = constraintOid;
1067 : 8880 : referenced.objectSubId = 0;
1068 : 8880 : recordDependencyOn(&myself, &referenced, DEPENDENCY_INTERNAL);
1069 : : }
1070 : : else
1071 : : {
1072 : : /*
1073 : : * User CREATE TRIGGER, so place dependencies. We make trigger be
1074 : : * auto-dropped if its relation is dropped or if the FK relation is
1075 : : * dropped. (Auto drop is compatible with our pre-7.3 behavior.)
1076 : : */
7691 1077 : 2565 : referenced.classId = RelationRelationId;
8698 1078 : 2565 : referenced.objectId = RelationGetRelid(rel);
1079 : 2565 : referenced.objectSubId = 0;
2640 1080 : 2565 : recordDependencyOn(&myself, &referenced, DEPENDENCY_AUTO);
1081 : :
6125 1082 [ + + ]: 2565 : if (OidIsValid(constrrelid))
1083 : : {
7691 1084 : 28 : referenced.classId = RelationRelationId;
8698 1085 : 28 : referenced.objectId = constrrelid;
1086 : 28 : referenced.objectSubId = 0;
1087 : 28 : recordDependencyOn(&myself, &referenced, DEPENDENCY_AUTO);
1088 : : }
1089 : : /* Not possible to have an index dependency in this case */
6125 1090 [ - + ]: 2565 : Assert(!OidIsValid(indexOid));
1091 : :
1092 : : /*
1093 : : * If it's a user-specified constraint trigger, make the constraint
1094 : : * internally dependent on the trigger instead of vice versa.
1095 : : */
5952 1096 [ + + ]: 2565 : if (OidIsValid(constraintOid))
1097 : : {
1098 : 117 : referenced.classId = ConstraintRelationId;
1099 : 117 : referenced.objectId = constraintOid;
1100 : 117 : referenced.objectSubId = 0;
1101 : 117 : recordDependencyOn(&referenced, &myself, DEPENDENCY_INTERNAL);
1102 : : }
1103 : :
1104 : : /*
1105 : : * If it's a partition trigger, create the partition dependencies.
1106 : : */
2965 alvherre@alvh.no-ip. 1107 [ + + ]: 2565 : if (OidIsValid(parentTriggerOid))
1108 : : {
1109 : 595 : ObjectAddressSet(referenced, TriggerRelationId, parentTriggerOid);
2640 tgl@sss.pgh.pa.us 1110 : 595 : recordDependencyOn(&myself, &referenced, DEPENDENCY_PARTITION_PRI);
1111 : 595 : ObjectAddressSet(referenced, RelationRelationId, RelationGetRelid(rel));
1112 : 595 : recordDependencyOn(&myself, &referenced, DEPENDENCY_PARTITION_SEC);
1113 : : }
1114 : : }
1115 : :
1116 : : /* If column-specific trigger, add normal dependencies on columns */
6047 1117 [ + + ]: 11445 : if (columns != NULL)
1118 : : {
1119 : : int i;
1120 : :
1121 : 65 : referenced.classId = RelationRelationId;
1122 : 65 : referenced.objectId = RelationGetRelid(rel);
1123 [ + + ]: 135 : for (i = 0; i < ncolumns; i++)
1124 : : {
1125 : 70 : referenced.objectSubId = columns[i];
1126 : 70 : recordDependencyOn(&myself, &referenced, DEPENDENCY_NORMAL);
1127 : : }
1128 : : }
1129 : :
1130 : : /*
1131 : : * If it has a WHEN clause, add dependencies on objects mentioned in the
1132 : : * expression (eg, functions, as well as any columns used).
1133 : : */
2965 alvherre@alvh.no-ip. 1134 [ + + ]: 11445 : if (whenRtable != NIL)
6010 tgl@sss.pgh.pa.us 1135 : 88 : recordDependencyOnExpr(&myself, whenClause, whenRtable,
1136 : : DEPENDENCY_NORMAL);
1137 : :
1138 : : /* Post creation hook for new trigger */
4797 rhaas@postgresql.org 1139 [ + + ]: 11445 : InvokeObjectPostCreateHookArg(TriggerRelationId, trigoid, 0,
1140 : : isInternal);
1141 : :
1142 : : /*
1143 : : * Lastly, create the trigger on child relations, if needed.
1144 : : */
2965 alvherre@alvh.no-ip. 1145 [ + + ]: 11445 : if (partition_recurse)
1146 : : {
1839 1147 : 271 : PartitionDesc partdesc = RelationGetPartitionDesc(rel, true);
1148 : : int i;
1149 : : MemoryContext oldcxt,
1150 : : perChildCxt;
1151 : :
2965 1152 : 271 : perChildCxt = AllocSetContextCreate(CurrentMemoryContext,
1153 : : "part trig clone",
1154 : : ALLOCSET_SMALL_SIZES);
1155 : :
1156 : : /*
1157 : : * We don't currently expect to be called with a valid indexOid. If
1158 : : * that ever changes then we'll need to write code here to find the
1159 : : * corresponding child index.
1160 : : */
1337 drowley@postgresql.o 1161 [ - + ]: 271 : Assert(!OidIsValid(indexOid));
1162 : :
2965 alvherre@alvh.no-ip. 1163 : 271 : oldcxt = MemoryContextSwitchTo(perChildCxt);
1164 : :
1165 : : /* Iterate to create the trigger on each existing partition */
1166 [ + + ]: 739 : for (i = 0; i < partdesc->nparts; i++)
1167 : : {
1168 : : CreateTrigStmt *childStmt;
1169 : : Relation childTbl;
1170 : : Node *qual;
1171 : :
2661 andres@anarazel.de 1172 : 472 : childTbl = table_open(partdesc->oids[i], ShareRowExclusiveLock);
1173 : :
1174 : : /*
1175 : : * Initialize our fabricated parse node by copying the original
1176 : : * one, then resetting fields that we pass separately.
1177 : : */
565 peter@eisentraut.org 1178 : 472 : childStmt = copyObject(stmt);
2965 alvherre@alvh.no-ip. 1179 : 472 : childStmt->funcname = NIL;
1180 : 472 : childStmt->whenClause = NULL;
1181 : :
1182 : : /* If there is a WHEN clause, create a modified copy of it */
1183 : 472 : qual = copyObject(whenClause);
1184 : : qual = (Node *)
1185 : 472 : map_partition_varattnos((List *) qual, PRS2_OLD_VARNO,
1186 : : childTbl, rel);
1187 : : qual = (Node *)
1188 : 472 : map_partition_varattnos((List *) qual, PRS2_NEW_VARNO,
1189 : : childTbl, rel);
1190 : :
1754 1191 : 472 : CreateTriggerFiringOn(childStmt, queryString,
1192 : 472 : partdesc->oids[i], refRelOid,
1193 : : InvalidOid, InvalidOid,
1194 : : funcoid, trigoid, qual,
1195 : : isInternal, true, trigger_fires_when);
1196 : :
2661 andres@anarazel.de 1197 : 468 : table_close(childTbl, NoLock);
1198 : :
2965 alvherre@alvh.no-ip. 1199 : 468 : MemoryContextReset(perChildCxt);
1200 : : }
1201 : :
1202 : 267 : MemoryContextSwitchTo(oldcxt);
1203 : 267 : MemoryContextDelete(perChildCxt);
1204 : : }
1205 : :
1206 : : /* Keep lock on target rel until end of xact */
2661 andres@anarazel.de 1207 : 11441 : table_close(rel, NoLock);
1208 : :
4081 alvherre@alvh.no-ip. 1209 : 11441 : return myself;
1210 : : }
1211 : :
1212 : : /*
1213 : : * TriggerSetParentTrigger
1214 : : * Set a partition's trigger as child of its parent trigger,
1215 : : * or remove the linkage if parentTrigId is InvalidOid.
1216 : : *
1217 : : * This updates the constraint's pg_trigger row to show it as inherited, and
1218 : : * adds PARTITION dependencies to prevent the trigger from being deleted
1219 : : * on its own. Alternatively, reverse that.
1220 : : */
1221 : : void
1581 1222 : 344 : TriggerSetParentTrigger(Relation trigRel,
1223 : : Oid childTrigId,
1224 : : Oid parentTrigId,
1225 : : Oid childTableId)
1226 : : {
1227 : : SysScanDesc tgscan;
1228 : : ScanKeyData skey[1];
1229 : : Form_pg_trigger trigForm;
1230 : : HeapTuple tuple,
1231 : : newtup;
1232 : : ObjectAddress depender;
1233 : : ObjectAddress referenced;
1234 : :
1235 : : /*
1236 : : * Find the trigger to delete.
1237 : : */
1238 : 344 : ScanKeyInit(&skey[0],
1239 : : Anum_pg_trigger_oid,
1240 : : BTEqualStrategyNumber, F_OIDEQ,
1241 : : ObjectIdGetDatum(childTrigId));
1242 : :
1243 : 344 : tgscan = systable_beginscan(trigRel, TriggerOidIndexId, true,
1244 : : NULL, 1, skey);
1245 : :
1246 : 344 : tuple = systable_getnext(tgscan);
1247 [ - + ]: 344 : if (!HeapTupleIsValid(tuple))
1581 alvherre@alvh.no-ip. 1248 [ # # ]:UBC 0 : elog(ERROR, "could not find tuple for trigger %u", childTrigId);
1581 alvherre@alvh.no-ip. 1249 :CBC 344 : newtup = heap_copytuple(tuple);
1250 : 344 : trigForm = (Form_pg_trigger) GETSTRUCT(newtup);
1251 [ + + ]: 344 : if (OidIsValid(parentTrigId))
1252 : : {
1253 : : /* don't allow setting parent for a constraint that already has one */
1254 [ - + ]: 200 : if (OidIsValid(trigForm->tgparentid))
1581 alvherre@alvh.no-ip. 1255 [ # # ]:UBC 0 : elog(ERROR, "trigger %u already has a parent trigger",
1256 : : childTrigId);
1257 : :
1581 alvherre@alvh.no-ip. 1258 :CBC 200 : trigForm->tgparentid = parentTrigId;
1259 : :
1260 : 200 : CatalogTupleUpdate(trigRel, &tuple->t_self, newtup);
1261 : :
1262 : 200 : ObjectAddressSet(depender, TriggerRelationId, childTrigId);
1263 : :
1264 : 200 : ObjectAddressSet(referenced, TriggerRelationId, parentTrigId);
1265 : 200 : recordDependencyOn(&depender, &referenced, DEPENDENCY_PARTITION_PRI);
1266 : :
1267 : 200 : ObjectAddressSet(referenced, RelationRelationId, childTableId);
1268 : 200 : recordDependencyOn(&depender, &referenced, DEPENDENCY_PARTITION_SEC);
1269 : : }
1270 : : else
1271 : : {
1272 : 144 : trigForm->tgparentid = InvalidOid;
1273 : :
1274 : 144 : CatalogTupleUpdate(trigRel, &tuple->t_self, newtup);
1275 : :
1276 : 144 : deleteDependencyRecordsForClass(TriggerRelationId, childTrigId,
1277 : : TriggerRelationId,
1278 : : DEPENDENCY_PARTITION_PRI);
1279 : 144 : deleteDependencyRecordsForClass(TriggerRelationId, childTrigId,
1280 : : RelationRelationId,
1281 : : DEPENDENCY_PARTITION_SEC);
1282 : : }
1283 : :
1284 : 344 : heap_freetuple(newtup);
1285 : 344 : systable_endscan(tgscan);
1286 : 344 : }
1287 : :
1288 : :
1289 : : /*
1290 : : * Guts of trigger deletion.
1291 : : */
1292 : : void
8698 tgl@sss.pgh.pa.us 1293 : 9752 : RemoveTriggerById(Oid trigOid)
1294 : : {
1295 : : Relation tgrel;
1296 : : SysScanDesc tgscan;
1297 : : ScanKeyData skey[1];
1298 : : HeapTuple tup;
1299 : : Oid relid;
1300 : : Relation rel;
1301 : :
2661 andres@anarazel.de 1302 : 9752 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1303 : :
1304 : : /*
1305 : : * Find the trigger to delete.
1306 : : */
8210 tgl@sss.pgh.pa.us 1307 : 9752 : ScanKeyInit(&skey[0],
1308 : : Anum_pg_trigger_oid,
1309 : : BTEqualStrategyNumber, F_OIDEQ,
1310 : : ObjectIdGetDatum(trigOid));
1311 : :
7691 1312 : 9752 : tgscan = systable_beginscan(tgrel, TriggerOidIndexId, true,
1313 : : NULL, 1, skey);
1314 : :
8698 1315 : 9752 : tup = systable_getnext(tgscan);
1316 [ - + ]: 9752 : if (!HeapTupleIsValid(tup))
8325 tgl@sss.pgh.pa.us 1317 [ # # ]:UBC 0 : elog(ERROR, "could not find tuple for trigger %u", trigOid);
1318 : :
1319 : : /*
1320 : : * Open and exclusive-lock the relation the trigger belongs to.
1321 : : */
8698 tgl@sss.pgh.pa.us 1322 :CBC 9752 : relid = ((Form_pg_trigger) GETSTRUCT(tup))->tgrelid;
1323 : :
2661 andres@anarazel.de 1324 : 9752 : rel = table_open(relid, AccessExclusiveLock);
1325 : :
5686 tgl@sss.pgh.pa.us 1326 [ + + ]: 9752 : if (rel->rd_rel->relkind != RELKIND_RELATION &&
4426 noah@leadboat.com 1327 [ + + ]: 1763 : rel->rd_rel->relkind != RELKIND_VIEW &&
3436 rhaas@postgresql.org 1328 [ + + ]: 1673 : rel->rd_rel->relkind != RELKIND_FOREIGN_TABLE &&
1329 [ - + ]: 1623 : rel->rd_rel->relkind != RELKIND_PARTITIONED_TABLE)
8325 tgl@sss.pgh.pa.us 1330 [ # # ]:UBC 0 : ereport(ERROR,
1331 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
1332 : : errmsg("relation \"%s\" cannot have triggers",
1333 : : RelationGetRelationName(rel)),
1334 : : errdetail_relkind_not_supported(rel->rd_rel->relkind)));
1335 : :
8789 tgl@sss.pgh.pa.us 1336 [ + + - + ]:CBC 9752 : if (!allowSystemTableMods && IsSystemRelation(rel))
8325 tgl@sss.pgh.pa.us 1337 [ # # ]:UBC 0 : ereport(ERROR,
1338 : : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1339 : : errmsg("permission denied: \"%s\" is a system catalog",
1340 : : RelationGetRelationName(rel))));
1341 : :
1342 : : /*
1343 : : * Delete the pg_trigger tuple.
1344 : : */
3380 tgl@sss.pgh.pa.us 1345 :CBC 9752 : CatalogTupleDelete(tgrel, &tup->t_self);
1346 : :
8841 1347 : 9752 : systable_endscan(tgscan);
2661 andres@anarazel.de 1348 : 9752 : table_close(tgrel, RowExclusiveLock);
1349 : :
1350 : : /*
1351 : : * We do not bother to try to determine whether any other triggers remain,
1352 : : * which would be needed in order to decide whether it's safe to clear the
1353 : : * relation's relhastriggers. (In any case, there might be a concurrent
1354 : : * process adding new triggers.) Instead, just force a relcache inval to
1355 : : * make other backends (and this one too!) rebuild their relcache entries.
1356 : : * There's no great harm in leaving relhastriggers true even if there are
1357 : : * no triggers left.
1358 : : */
6386 tgl@sss.pgh.pa.us 1359 : 9752 : CacheInvalidateRelcache(rel);
1360 : :
1361 : : /* Keep lock on trigger's rel until end of xact */
2661 andres@anarazel.de 1362 : 9752 : table_close(rel, NoLock);
10474 vadim4o@yahoo.com 1363 : 9752 : }
1364 : :
1365 : : /*
1366 : : * get_trigger_oid - Look up a trigger by name to find its OID.
1367 : : *
1368 : : * If missing_ok is false, throw an error if trigger not found. If
1369 : : * true, just return InvalidOid.
1370 : : */
1371 : : Oid
5752 rhaas@postgresql.org 1372 : 511 : get_trigger_oid(Oid relid, const char *trigname, bool missing_ok)
1373 : : {
1374 : : Relation tgrel;
1375 : : ScanKeyData skey[2];
1376 : : SysScanDesc tgscan;
1377 : : HeapTuple tup;
1378 : : Oid oid;
1379 : :
1380 : : /*
1381 : : * Find the trigger, verify permissions, set up object address
1382 : : */
2661 andres@anarazel.de 1383 : 511 : tgrel = table_open(TriggerRelationId, AccessShareLock);
1384 : :
5752 rhaas@postgresql.org 1385 : 511 : ScanKeyInit(&skey[0],
1386 : : Anum_pg_trigger_tgrelid,
1387 : : BTEqualStrategyNumber, F_OIDEQ,
1388 : : ObjectIdGetDatum(relid));
1389 : 511 : ScanKeyInit(&skey[1],
1390 : : Anum_pg_trigger_tgname,
1391 : : BTEqualStrategyNumber, F_NAMEEQ,
1392 : : CStringGetDatum(trigname));
1393 : :
1394 : 511 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1395 : : NULL, 2, skey);
1396 : :
1397 : 511 : tup = systable_getnext(tgscan);
1398 : :
1399 [ + + ]: 511 : if (!HeapTupleIsValid(tup))
1400 : : {
1401 [ + + ]: 20 : if (!missing_ok)
1402 [ + - ]: 16 : ereport(ERROR,
1403 : : (errcode(ERRCODE_UNDEFINED_OBJECT),
1404 : : errmsg("trigger \"%s\" for table \"%s\" does not exist",
1405 : : trigname, get_rel_name(relid))));
1406 : 4 : oid = InvalidOid;
1407 : : }
1408 : : else
1409 : : {
2723 andres@anarazel.de 1410 : 491 : oid = ((Form_pg_trigger) GETSTRUCT(tup))->oid;
1411 : : }
1412 : :
5752 rhaas@postgresql.org 1413 : 495 : systable_endscan(tgscan);
2661 andres@anarazel.de 1414 : 495 : table_close(tgrel, AccessShareLock);
5752 rhaas@postgresql.org 1415 : 495 : return oid;
1416 : : }
1417 : :
1418 : : /*
1419 : : * Perform permissions and integrity checks before acquiring a relation lock.
1420 : : */
1421 : : static void
5255 1422 : 27 : RangeVarCallbackForRenameTrigger(const RangeVar *rv, Oid relid, Oid oldrelid,
1423 : : void *arg)
1424 : : {
1425 : : HeapTuple tuple;
1426 : : Form_pg_class form;
1427 : :
1428 : 27 : tuple = SearchSysCache1(RELOID, ObjectIdGetDatum(relid));
1429 [ - + ]: 27 : if (!HeapTupleIsValid(tuple))
5077 bruce@momjian.us 1430 :UBC 0 : return; /* concurrently dropped */
5255 rhaas@postgresql.org 1431 :CBC 27 : form = (Form_pg_class) GETSTRUCT(tuple);
1432 : :
1433 : : /* only tables and views can have triggers */
4426 noah@leadboat.com 1434 [ + + + - ]: 27 : if (form->relkind != RELKIND_RELATION && form->relkind != RELKIND_VIEW &&
3436 rhaas@postgresql.org 1435 [ + - ]: 16 : form->relkind != RELKIND_FOREIGN_TABLE &&
1436 [ - + ]: 16 : form->relkind != RELKIND_PARTITIONED_TABLE)
5077 bruce@momjian.us 1437 [ # # ]:UBC 0 : ereport(ERROR,
1438 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
1439 : : errmsg("relation \"%s\" cannot have triggers",
1440 : : rv->relname),
1441 : : errdetail_relkind_not_supported(form->relkind)));
1442 : :
1443 : : /* you must own the table to rename one of its triggers */
1269 peter@eisentraut.org 1444 [ - + ]:CBC 27 : if (!object_ownercheck(RelationRelationId, relid, GetUserId()))
3076 peter_e@gmx.net 1445 :UBC 0 : aclcheck_error(ACLCHECK_NOT_OWNER, get_relkind_objtype(get_rel_relkind(relid)), rv->relname);
4541 rhaas@postgresql.org 1446 [ + + + + ]:CBC 27 : if (!allowSystemTableMods && IsSystemClass(relid, form))
5077 bruce@momjian.us 1447 [ + - ]: 1 : ereport(ERROR,
1448 : : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1449 : : errmsg("permission denied: \"%s\" is a system catalog",
1450 : : rv->relname)));
1451 : :
5255 rhaas@postgresql.org 1452 : 26 : ReleaseSysCache(tuple);
1453 : : }
1454 : :
1455 : : /*
1456 : : * renametrig - changes the name of a trigger on a relation
1457 : : *
1458 : : * trigger name is changed in trigger catalog.
1459 : : * No record of the previous name is kept.
1460 : : *
1461 : : * get proper relrelation from relation catalog (if not arg)
1462 : : * scan trigger catalog
1463 : : * for name conflict (within rel)
1464 : : * for original trigger (if not arg)
1465 : : * modify tgname in trigger tuple
1466 : : * update row in catalog
1467 : : */
1468 : : ObjectAddress
1469 : 26 : renametrig(RenameStmt *stmt)
1470 : : {
1471 : : Oid tgoid;
1472 : : Relation targetrel;
1473 : : Relation tgrel;
1474 : : HeapTuple tuple;
1475 : : SysScanDesc tgscan;
1476 : : ScanKeyData key[2];
1477 : : Oid relid;
1478 : : ObjectAddress address;
1479 : :
1480 : : /*
1481 : : * Look up name, check permissions, and acquire lock (which we will NOT
1482 : : * release until end of transaction).
1483 : : */
1484 : 26 : relid = RangeVarGetRelidExtended(stmt->relation, AccessExclusiveLock,
1485 : : 0,
1486 : : RangeVarCallbackForRenameTrigger,
1487 : : NULL);
1488 : :
1489 : : /* Have lock already, so just need to build relcache entry. */
1490 : 25 : targetrel = relation_open(relid, NoLock);
1491 : :
1492 : : /*
1493 : : * On partitioned tables, this operation recurses to partitions. Lock all
1494 : : * tables upfront.
1495 : : */
1748 alvherre@alvh.no-ip. 1496 [ + + ]: 25 : if (targetrel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1497 : 16 : (void) find_all_inheritors(relid, AccessExclusiveLock, NULL);
1498 : :
1499 : 25 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1500 : :
1501 : : /*
1502 : : * Search for the trigger to modify.
1503 : : */
8210 tgl@sss.pgh.pa.us 1504 : 25 : ScanKeyInit(&key[0],
1505 : : Anum_pg_trigger_tgrelid,
1506 : : BTEqualStrategyNumber, F_OIDEQ,
1507 : : ObjectIdGetDatum(relid));
1508 : 25 : ScanKeyInit(&key[1],
1509 : : Anum_pg_trigger_tgname,
1510 : : BTEqualStrategyNumber, F_NAMEEQ,
5255 rhaas@postgresql.org 1511 : 25 : PointerGetDatum(stmt->subname));
7691 tgl@sss.pgh.pa.us 1512 : 25 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1513 : : NULL, 2, key);
8771 1514 [ + - ]: 25 : if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1515 : : {
1516 : : Form_pg_trigger trigform;
1517 : :
2661 andres@anarazel.de 1518 : 25 : trigform = (Form_pg_trigger) GETSTRUCT(tuple);
1519 : 25 : tgoid = trigform->oid;
1520 : :
1521 : : /*
1522 : : * If the trigger descends from a trigger on a parent partitioned
1523 : : * table, reject the rename. We don't allow a trigger in a partition
1524 : : * to differ in name from that of its parent: that would lead to an
1525 : : * inconsistency that pg_dump would not reproduce.
1526 : : */
1748 alvherre@alvh.no-ip. 1527 [ + + ]: 25 : if (OidIsValid(trigform->tgparentid))
1528 [ + - ]: 4 : ereport(ERROR,
1529 : : errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
1530 : : errmsg("cannot rename trigger \"%s\" on table \"%s\"",
1531 : : stmt->subname, RelationGetRelationName(targetrel)),
1532 : : errhint("Rename the trigger on the partitioned table \"%s\" instead.",
1533 : : get_rel_name(get_partition_parent(relid, false))));
1534 : :
1535 : :
1536 : : /* Rename the trigger on this relation ... */
1537 : 21 : renametrig_internal(tgrel, targetrel, tuple, stmt->newname,
1538 : 21 : stmt->subname);
1539 : :
1540 : : /* ... and if it is partitioned, recurse to its partitions */
1541 [ + + ]: 21 : if (targetrel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1542 : : {
1543 : 12 : PartitionDesc partdesc = RelationGetPartitionDesc(targetrel, true);
1544 : :
1545 [ + + ]: 20 : for (int i = 0; i < partdesc->nparts; i++)
1546 : : {
1547 : 12 : Oid partitionId = partdesc->oids[i];
1548 : :
1549 : 12 : renametrig_partition(tgrel, partitionId, trigform->oid,
1550 : 12 : stmt->newname, stmt->subname);
1551 : : }
1552 : : }
1553 : : }
1554 : : else
1555 : : {
8325 tgl@sss.pgh.pa.us 1556 [ # # ]:UBC 0 : ereport(ERROR,
1557 : : (errcode(ERRCODE_UNDEFINED_OBJECT),
1558 : : errmsg("trigger \"%s\" for table \"%s\" does not exist",
1559 : : stmt->subname, RelationGetRelationName(targetrel))));
1560 : : }
1561 : :
4081 alvherre@alvh.no-ip. 1562 :CBC 17 : ObjectAddressSet(address, TriggerRelationId, tgoid);
1563 : :
8771 tgl@sss.pgh.pa.us 1564 : 17 : systable_endscan(tgscan);
1565 : :
2661 andres@anarazel.de 1566 : 17 : table_close(tgrel, RowExclusiveLock);
1567 : :
1568 : : /*
1569 : : * Close rel, but keep exclusive lock!
1570 : : */
5255 rhaas@postgresql.org 1571 : 17 : relation_close(targetrel, NoLock);
1572 : :
4081 alvherre@alvh.no-ip. 1573 : 17 : return address;
1574 : : }
1575 : :
1576 : : /*
1577 : : * Subroutine for renametrig -- perform the actual work of renaming one
1578 : : * trigger on one table.
1579 : : *
1580 : : * If the trigger has a name different from the expected one, raise a
1581 : : * NOTICE about it.
1582 : : */
1583 : : static void
1748 1584 : 37 : renametrig_internal(Relation tgrel, Relation targetrel, HeapTuple trigtup,
1585 : : const char *newname, const char *expected_name)
1586 : : {
1587 : : HeapTuple tuple;
1588 : : Form_pg_trigger tgform;
1589 : : ScanKeyData key[2];
1590 : : SysScanDesc tgscan;
1591 : :
1592 : : /* If the trigger already has the new name, nothing to do. */
1593 : 37 : tgform = (Form_pg_trigger) GETSTRUCT(trigtup);
1594 [ - + ]: 37 : if (strcmp(NameStr(tgform->tgname), newname) == 0)
1748 alvherre@alvh.no-ip. 1595 :UBC 0 : return;
1596 : :
1597 : : /*
1598 : : * Before actually trying the rename, search for triggers with the same
1599 : : * name. The update would fail with an ugly message in that case, and it
1600 : : * is better to throw a nicer error.
1601 : : */
1748 alvherre@alvh.no-ip. 1602 :CBC 37 : ScanKeyInit(&key[0],
1603 : : Anum_pg_trigger_tgrelid,
1604 : : BTEqualStrategyNumber, F_OIDEQ,
1605 : : ObjectIdGetDatum(RelationGetRelid(targetrel)));
1606 : 37 : ScanKeyInit(&key[1],
1607 : : Anum_pg_trigger_tgname,
1608 : : BTEqualStrategyNumber, F_NAMEEQ,
1609 : : PointerGetDatum(newname));
1610 : 37 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1611 : : NULL, 2, key);
1612 [ + + ]: 37 : if (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1613 [ + - ]: 4 : ereport(ERROR,
1614 : : (errcode(ERRCODE_DUPLICATE_OBJECT),
1615 : : errmsg("trigger \"%s\" for relation \"%s\" already exists",
1616 : : newname, RelationGetRelationName(targetrel))));
1617 : 33 : systable_endscan(tgscan);
1618 : :
1619 : : /*
1620 : : * The target name is free; update the existing pg_trigger tuple with it.
1621 : : */
1622 : 33 : tuple = heap_copytuple(trigtup); /* need a modifiable copy */
1623 : 33 : tgform = (Form_pg_trigger) GETSTRUCT(tuple);
1624 : :
1625 : : /*
1626 : : * If the trigger has a name different from what we expected, let the user
1627 : : * know. (We can proceed anyway, since we must have reached here following
1628 : : * a tgparentid link.)
1629 : : */
1630 [ - + ]: 33 : if (strcmp(NameStr(tgform->tgname), expected_name) != 0)
1748 alvherre@alvh.no-ip. 1631 [ # # ]:UBC 0 : ereport(NOTICE,
1632 : : errmsg("renamed trigger \"%s\" on relation \"%s\"",
1633 : : NameStr(tgform->tgname),
1634 : : RelationGetRelationName(targetrel)));
1635 : :
1748 alvherre@alvh.no-ip. 1636 :CBC 33 : namestrcpy(&tgform->tgname, newname);
1637 : :
1638 : 33 : CatalogTupleUpdate(tgrel, &tuple->t_self, tuple);
1639 : :
1640 [ - + ]: 33 : InvokeObjectPostAlterHook(TriggerRelationId, tgform->oid, 0);
1641 : :
1642 : : /*
1643 : : * Invalidate relation's relcache entry so that other backends (and this
1644 : : * one too!) are sent SI message to make them rebuild relcache entries.
1645 : : * (Ideally this should happen automatically...)
1646 : : */
1647 : 33 : CacheInvalidateRelcache(targetrel);
1648 : : }
1649 : :
1650 : : /*
1651 : : * Subroutine for renametrig -- Helper for recursing to partitions when
1652 : : * renaming triggers on a partitioned table.
1653 : : */
1654 : : static void
1655 : 20 : renametrig_partition(Relation tgrel, Oid partitionId, Oid parentTriggerOid,
1656 : : const char *newname, const char *expected_name)
1657 : : {
1658 : : SysScanDesc tgscan;
1659 : : ScanKeyData key;
1660 : : HeapTuple tuple;
1661 : :
1662 : : /*
1663 : : * Given a relation and the OID of a trigger on parent relation, find the
1664 : : * corresponding trigger in the child and rename that trigger to the given
1665 : : * name.
1666 : : */
1667 : 20 : ScanKeyInit(&key,
1668 : : Anum_pg_trigger_tgrelid,
1669 : : BTEqualStrategyNumber, F_OIDEQ,
1670 : : ObjectIdGetDatum(partitionId));
1671 : 20 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1672 : : NULL, 1, &key);
1673 [ + + ]: 32 : while (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1674 : : {
1675 : 28 : Form_pg_trigger tgform = (Form_pg_trigger) GETSTRUCT(tuple);
1676 : : Relation partitionRel;
1677 : :
1678 [ + + ]: 28 : if (tgform->tgparentid != parentTriggerOid)
1679 : 12 : continue; /* not our trigger */
1680 : :
1681 : 16 : partitionRel = table_open(partitionId, NoLock);
1682 : :
1683 : : /* Rename the trigger on this partition */
1684 : 16 : renametrig_internal(tgrel, partitionRel, tuple, newname, expected_name);
1685 : :
1686 : : /* And if this relation is partitioned, recurse to its partitions */
1687 [ + + ]: 12 : if (partitionRel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
1688 : : {
1689 : 4 : PartitionDesc partdesc = RelationGetPartitionDesc(partitionRel,
1690 : : true);
1691 : :
1692 [ + + ]: 12 : for (int i = 0; i < partdesc->nparts; i++)
1693 : : {
1308 drowley@postgresql.o 1694 : 8 : Oid partoid = partdesc->oids[i];
1695 : :
1696 : 8 : renametrig_partition(tgrel, partoid, tgform->oid, newname,
1748 alvherre@alvh.no-ip. 1697 : 8 : NameStr(tgform->tgname));
1698 : : }
1699 : : }
1700 : 12 : table_close(partitionRel, NoLock);
1701 : :
1702 : : /* There should be at most one matching tuple */
1744 1703 : 12 : break;
1704 : : }
1748 1705 : 16 : systable_endscan(tgscan);
1706 : 16 : }
1707 : :
1708 : : /*
1709 : : * EnableDisableTrigger()
1710 : : *
1711 : : * Called by ALTER TABLE ENABLE/DISABLE [ REPLICA | ALWAYS ] TRIGGER
1712 : : * to change 'tgenabled' field for the specified trigger(s)
1713 : : *
1714 : : * rel: relation to process (caller must hold suitable lock on it)
1715 : : * tgname: name of trigger to process, or NULL to scan all triggers
1716 : : * tgparent: if not zero, process only triggers with this tgparentid
1717 : : * fires_when: new value for tgenabled field. In addition to generic
1718 : : * enablement/disablement, this also defines when the trigger
1719 : : * should be fired in session replication roles.
1720 : : * skip_system: if true, skip "system" triggers (constraint triggers)
1721 : : * recurse: if true, recurse to partitions
1722 : : *
1723 : : * Caller should have checked permissions for the table; here we also
1724 : : * enforce that superuser privilege is required to alter the state of
1725 : : * system triggers
1726 : : */
1727 : : void
1158 tgl@sss.pgh.pa.us 1728 : 264 : EnableDisableTrigger(Relation rel, const char *tgname, Oid tgparent,
1729 : : char fires_when, bool skip_system, bool recurse,
1730 : : LOCKMODE lockmode)
1731 : : {
1732 : : Relation tgrel;
1733 : : int nkeys;
1734 : : ScanKeyData keys[2];
1735 : : SysScanDesc tgscan;
1736 : : HeapTuple tuple;
1737 : : bool found;
1738 : : bool changed;
1739 : :
1740 : : /* Scan the relevant entries in pg_triggers */
2661 andres@anarazel.de 1741 : 264 : tgrel = table_open(TriggerRelationId, RowExclusiveLock);
1742 : :
7560 tgl@sss.pgh.pa.us 1743 : 264 : ScanKeyInit(&keys[0],
1744 : : Anum_pg_trigger_tgrelid,
1745 : : BTEqualStrategyNumber, F_OIDEQ,
1746 : : ObjectIdGetDatum(RelationGetRelid(rel)));
1747 [ + + ]: 264 : if (tgname)
1748 : : {
1749 : 175 : ScanKeyInit(&keys[1],
1750 : : Anum_pg_trigger_tgname,
1751 : : BTEqualStrategyNumber, F_NAMEEQ,
1752 : : CStringGetDatum(tgname));
1753 : 175 : nkeys = 2;
1754 : : }
1755 : : else
1756 : 89 : nkeys = 1;
1757 : :
1758 : 264 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1759 : : NULL, nkeys, keys);
1760 : :
1761 : 264 : found = changed = false;
1762 : :
1763 [ + + ]: 712 : while (HeapTupleIsValid(tuple = systable_getnext(tgscan)))
1764 : : {
1765 : 448 : Form_pg_trigger oldtrig = (Form_pg_trigger) GETSTRUCT(tuple);
1766 : :
1158 1767 [ + + + + ]: 448 : if (OidIsValid(tgparent) && tgparent != oldtrig->tgparentid)
1768 : 128 : continue;
1769 : :
5952 1770 [ + + ]: 320 : if (oldtrig->tgisinternal)
1771 : : {
1772 : : /* system trigger ... ok to process? */
7560 1773 [ + + ]: 48 : if (skip_system)
1774 : 8 : continue;
1775 [ - + ]: 40 : if (!superuser())
7560 tgl@sss.pgh.pa.us 1776 [ # # ]:UBC 0 : ereport(ERROR,
1777 : : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
1778 : : errmsg("permission denied: \"%s\" is a system trigger",
1779 : : NameStr(oldtrig->tgname))));
1780 : : }
1781 : :
7560 tgl@sss.pgh.pa.us 1782 :CBC 312 : found = true;
1783 : :
6987 JanWieck@Yahoo.com 1784 [ + + ]: 312 : if (oldtrig->tgenabled != fires_when)
1785 : : {
1786 : : /* need to change this one ... make a copy to scribble on */
7507 bruce@momjian.us 1787 : 292 : HeapTuple newtup = heap_copytuple(tuple);
7560 tgl@sss.pgh.pa.us 1788 : 292 : Form_pg_trigger newtrig = (Form_pg_trigger) GETSTRUCT(newtup);
1789 : :
6987 JanWieck@Yahoo.com 1790 : 292 : newtrig->tgenabled = fires_when;
1791 : :
3381 alvherre@alvh.no-ip. 1792 : 292 : CatalogTupleUpdate(tgrel, &newtup->t_self, newtup);
1793 : :
7560 tgl@sss.pgh.pa.us 1794 : 292 : heap_freetuple(newtup);
1795 : :
1796 : 292 : changed = true;
1797 : : }
1798 : :
1799 : : /*
1800 : : * When altering FOR EACH ROW triggers on a partitioned table, do the
1801 : : * same on the partitions as well, unless ONLY is specified.
1802 : : *
1803 : : * Note that we recurse even if we didn't change the trigger above,
1804 : : * because the partitions' copy of the trigger may have a different
1805 : : * value of tgenabled than the parent's trigger and thus might need to
1806 : : * be changed.
1807 : : */
1370 alvherre@alvh.no-ip. 1808 [ + + ]: 312 : if (recurse &&
1809 [ + + ]: 295 : rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE &&
1810 [ + + ]: 57 : (TRIGGER_FOR_ROW(oldtrig->tgtype)))
1811 : : {
1812 : 49 : PartitionDesc partdesc = RelationGetPartitionDesc(rel, true);
1813 : : int i;
1814 : :
1815 [ + + ]: 122 : for (i = 0; i < partdesc->nparts; i++)
1816 : : {
1817 : : Relation part;
1818 : :
1819 : 73 : part = relation_open(partdesc->oids[i], lockmode);
1820 : : /* Match on child triggers' tgparentid, not their name */
1158 tgl@sss.pgh.pa.us 1821 : 73 : EnableDisableTrigger(part, NULL, oldtrig->oid,
1822 : : fires_when, skip_system, recurse,
1823 : : lockmode);
1370 alvherre@alvh.no-ip. 1824 : 73 : table_close(part, NoLock); /* keep lock till commit */
1825 : : }
1826 : : }
1827 : :
4797 rhaas@postgresql.org 1828 [ + + ]: 312 : InvokeObjectPostAlterHook(TriggerRelationId,
1829 : : oldtrig->oid, 0);
1830 : : }
1831 : :
7560 tgl@sss.pgh.pa.us 1832 : 264 : systable_endscan(tgscan);
1833 : :
2661 andres@anarazel.de 1834 : 264 : table_close(tgrel, RowExclusiveLock);
1835 : :
7560 tgl@sss.pgh.pa.us 1836 [ + + - + ]: 264 : if (tgname && !found)
7560 tgl@sss.pgh.pa.us 1837 [ # # ]:UBC 0 : ereport(ERROR,
1838 : : (errcode(ERRCODE_UNDEFINED_OBJECT),
1839 : : errmsg("trigger \"%s\" for table \"%s\" does not exist",
1840 : : tgname, RelationGetRelationName(rel))));
1841 : :
1842 : : /*
1843 : : * If we changed anything, broadcast a SI inval message to force each
1844 : : * backend (including our own!) to rebuild relation's relcache entry.
1845 : : * Otherwise they will fail to apply the change promptly.
1846 : : */
7560 tgl@sss.pgh.pa.us 1847 [ + + ]:CBC 264 : if (changed)
1848 : 260 : CacheInvalidateRelcache(rel);
1849 : 264 : }
1850 : :
1851 : :
1852 : : /*
1853 : : * Build trigger data to attach to the given relcache entry.
1854 : : *
1855 : : * Note that trigger data attached to a relcache entry must be stored in
1856 : : * CacheMemoryContext to ensure it survives as long as the relcache entry.
1857 : : * But we should be running in a less long-lived working context. To avoid
1858 : : * leaking cache memory if this routine fails partway through, we build a
1859 : : * temporary TriggerDesc in working memory and then copy the completed
1860 : : * structure into cache memory.
1861 : : */
1862 : : void
10467 bruce@momjian.us 1863 : 42333 : RelationBuildTriggers(Relation relation)
1864 : : {
1865 : : TriggerDesc *trigdesc;
1866 : : int numtrigs;
1867 : : int maxtrigs;
1868 : : Trigger *triggers;
1869 : : Relation tgrel;
1870 : : ScanKeyData skey;
1871 : : SysScanDesc tgscan;
1872 : : HeapTuple htup;
1873 : : MemoryContext oldContext;
1874 : : int i;
1875 : :
1876 : : /*
1877 : : * Allocate a working array to hold the triggers (the array is extended if
1878 : : * necessary)
1879 : : */
6386 tgl@sss.pgh.pa.us 1880 : 42333 : maxtrigs = 16;
1881 : 42333 : triggers = (Trigger *) palloc(maxtrigs * sizeof(Trigger));
1882 : 42333 : numtrigs = 0;
1883 : :
1884 : : /*
1885 : : * Note: since we scan the triggers using TriggerRelidNameIndexId, we will
1886 : : * be reading the triggers in name order, except possibly during
1887 : : * emergency-recovery operations (ie, IgnoreSystemIndexes). This in turn
1888 : : * ensures that triggers will be fired in name order.
1889 : : */
8210 1890 : 42333 : ScanKeyInit(&skey,
1891 : : Anum_pg_trigger_tgrelid,
1892 : : BTEqualStrategyNumber, F_OIDEQ,
1893 : : ObjectIdGetDatum(RelationGetRelid(relation)));
1894 : :
2661 andres@anarazel.de 1895 : 42333 : tgrel = table_open(TriggerRelationId, AccessShareLock);
7691 tgl@sss.pgh.pa.us 1896 : 42333 : tgscan = systable_beginscan(tgrel, TriggerRelidNameIndexId, true,
1897 : : NULL, 1, &skey);
1898 : :
8841 1899 [ + + ]: 118854 : while (HeapTupleIsValid(htup = systable_getnext(tgscan)))
1900 : : {
1901 : 76521 : Form_pg_trigger pg_trigger = (Form_pg_trigger) GETSTRUCT(htup);
1902 : : Trigger *build;
1903 : : Datum datum;
1904 : : bool isnull;
1905 : :
6386 1906 [ + + ]: 76521 : if (numtrigs >= maxtrigs)
1907 : : {
1908 : 32 : maxtrigs *= 2;
1909 : 32 : triggers = (Trigger *) repalloc(triggers, maxtrigs * sizeof(Trigger));
1910 : : }
1911 : 76521 : build = &(triggers[numtrigs]);
1912 : :
2723 andres@anarazel.de 1913 : 76521 : build->tgoid = pg_trigger->oid;
8604 tgl@sss.pgh.pa.us 1914 : 76521 : build->tgname = DatumGetCString(DirectFunctionCall1(nameout,
1915 : : NameGetDatum(&pg_trigger->tgname)));
10467 bruce@momjian.us 1916 : 76521 : build->tgfoid = pg_trigger->tgfoid;
1917 : 76521 : build->tgtype = pg_trigger->tgtype;
9715 JanWieck@Yahoo.com 1918 : 76521 : build->tgenabled = pg_trigger->tgenabled;
5952 tgl@sss.pgh.pa.us 1919 : 76521 : build->tgisinternal = pg_trigger->tgisinternal;
2239 alvherre@alvh.no-ip. 1920 : 76521 : build->tgisclone = OidIsValid(pg_trigger->tgparentid);
8800 tgl@sss.pgh.pa.us 1921 : 76521 : build->tgconstrrelid = pg_trigger->tgconstrrelid;
6125 1922 : 76521 : build->tgconstrindid = pg_trigger->tgconstrindid;
7020 1923 : 76521 : build->tgconstraint = pg_trigger->tgconstraint;
9715 JanWieck@Yahoo.com 1924 : 76521 : build->tgdeferrable = pg_trigger->tgdeferrable;
1925 : 76521 : build->tginitdeferred = pg_trigger->tginitdeferred;
10467 bruce@momjian.us 1926 : 76521 : build->tgnargs = pg_trigger->tgnargs;
1927 : : /* tgattr is first var-width field, so OK to access directly */
7707 tgl@sss.pgh.pa.us 1928 : 76521 : build->tgnattr = pg_trigger->tgattr.dim1;
1929 [ + + ]: 76521 : if (build->tgnattr > 0)
1930 : : {
5062 peter_e@gmx.net 1931 : 346 : build->tgattr = (int16 *) palloc(build->tgnattr * sizeof(int16));
7707 tgl@sss.pgh.pa.us 1932 : 346 : memcpy(build->tgattr, &(pg_trigger->tgattr.values),
5062 peter_e@gmx.net 1933 : 346 : build->tgnattr * sizeof(int16));
1934 : : }
1935 : : else
7707 tgl@sss.pgh.pa.us 1936 : 76175 : build->tgattr = NULL;
10467 bruce@momjian.us 1937 [ + + ]: 76521 : if (build->tgnargs > 0)
1938 : : {
1939 : : bytea *val;
1940 : : char *p;
1941 : :
3341 noah@leadboat.com 1942 : 2834 : val = DatumGetByteaPP(fastgetattr(htup,
1943 : : Anum_pg_trigger_tgargs,
1944 : : tgrel->rd_att, &isnull));
10467 bruce@momjian.us 1945 [ - + ]: 2834 : if (isnull)
8325 tgl@sss.pgh.pa.us 1946 [ # # ]:UBC 0 : elog(ERROR, "tgargs is null in trigger for relation \"%s\"",
1947 : : RelationGetRelationName(relation));
3341 noah@leadboat.com 1948 [ + - ]:CBC 2834 : p = (char *) VARDATA_ANY(val);
8604 tgl@sss.pgh.pa.us 1949 : 2834 : build->tgargs = (char **) palloc(build->tgnargs * sizeof(char *));
10467 bruce@momjian.us 1950 [ + + ]: 6495 : for (i = 0; i < build->tgnargs; i++)
1951 : : {
8604 tgl@sss.pgh.pa.us 1952 : 3661 : build->tgargs[i] = pstrdup(p);
10467 bruce@momjian.us 1953 : 3661 : p += strlen(p) + 1;
1954 : : }
1955 : : }
1956 : : else
1957 : 73687 : build->tgargs = NULL;
1958 : :
3469 kgrittn@postgresql.o 1959 : 76521 : datum = fastgetattr(htup, Anum_pg_trigger_tgoldtable,
1960 : : tgrel->rd_att, &isnull);
1961 [ + + ]: 76521 : if (!isnull)
1962 : 764 : build->tgoldtable =
1963 : 764 : DatumGetCString(DirectFunctionCall1(nameout, datum));
1964 : : else
1965 : 75757 : build->tgoldtable = NULL;
1966 : :
1967 : 76521 : datum = fastgetattr(htup, Anum_pg_trigger_tgnewtable,
1968 : : tgrel->rd_att, &isnull);
1969 [ + + ]: 76521 : if (!isnull)
1970 : 1004 : build->tgnewtable =
1971 : 1004 : DatumGetCString(DirectFunctionCall1(nameout, datum));
1972 : : else
1973 : 75517 : build->tgnewtable = NULL;
1974 : :
6010 tgl@sss.pgh.pa.us 1975 : 76521 : datum = fastgetattr(htup, Anum_pg_trigger_tgqual,
1976 : : tgrel->rd_att, &isnull);
1977 [ + + ]: 76521 : if (!isnull)
1978 : 579 : build->tgqual = TextDatumGetCString(datum);
1979 : : else
1980 : 75942 : build->tgqual = NULL;
1981 : :
6386 1982 : 76521 : numtrigs++;
1983 : : }
1984 : :
8841 1985 : 42333 : systable_endscan(tgscan);
2661 andres@anarazel.de 1986 : 42333 : table_close(tgrel, AccessShareLock);
1987 : :
1988 : : /* There might not be any triggers */
6386 tgl@sss.pgh.pa.us 1989 [ + + ]: 42333 : if (numtrigs == 0)
1990 : : {
1991 : 9843 : pfree(triggers);
1992 : 9843 : return;
1993 : : }
1994 : :
1995 : : /* Build trigdesc */
146 michael@paquier.xyz 1996 :GNC 32490 : trigdesc = palloc0_object(TriggerDesc);
10467 bruce@momjian.us 1997 :CBC 32490 : trigdesc->triggers = triggers;
6386 tgl@sss.pgh.pa.us 1998 : 32490 : trigdesc->numtriggers = numtrigs;
1999 [ + + ]: 109011 : for (i = 0; i < numtrigs; i++)
5686 2000 : 76521 : SetTriggerFlags(trigdesc, &(triggers[i]));
2001 : :
2002 : : /* Copy completed trigdesc into cache storage */
8604 2003 : 32490 : oldContext = MemoryContextSwitchTo(CacheMemoryContext);
2004 : 32490 : relation->trigdesc = CopyTriggerDesc(trigdesc);
2005 : 32490 : MemoryContextSwitchTo(oldContext);
2006 : :
2007 : : /* Release working memory */
2008 : 32490 : FreeTriggerDesc(trigdesc);
2009 : : }
2010 : :
2011 : : /*
2012 : : * Update the TriggerDesc's hint flags to include the specified trigger
2013 : : */
2014 : : static void
5686 2015 : 76521 : SetTriggerFlags(TriggerDesc *trigdesc, Trigger *trigger)
2016 : : {
2017 : 76521 : int16 tgtype = trigger->tgtype;
2018 : :
2019 : 76521 : trigdesc->trig_insert_before_row |=
2020 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2021 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_INSERT);
2022 : 76521 : trigdesc->trig_insert_after_row |=
2023 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2024 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_INSERT);
2025 : 76521 : trigdesc->trig_insert_instead_row |=
2026 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2027 : : TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_INSERT);
2028 : 76521 : trigdesc->trig_insert_before_statement |=
2029 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2030 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_INSERT);
2031 : 76521 : trigdesc->trig_insert_after_statement |=
2032 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2033 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_INSERT);
2034 : 76521 : trigdesc->trig_update_before_row |=
2035 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2036 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_UPDATE);
2037 : 76521 : trigdesc->trig_update_after_row |=
2038 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2039 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_UPDATE);
2040 : 76521 : trigdesc->trig_update_instead_row |=
2041 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2042 : : TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_UPDATE);
2043 : 76521 : trigdesc->trig_update_before_statement |=
2044 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2045 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_UPDATE);
2046 : 76521 : trigdesc->trig_update_after_statement |=
2047 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2048 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_UPDATE);
2049 : 76521 : trigdesc->trig_delete_before_row |=
2050 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2051 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_DELETE);
2052 : 76521 : trigdesc->trig_delete_after_row |=
2053 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2054 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_DELETE);
2055 : 76521 : trigdesc->trig_delete_instead_row |=
2056 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_ROW,
2057 : : TRIGGER_TYPE_INSTEAD, TRIGGER_TYPE_DELETE);
2058 : 76521 : trigdesc->trig_delete_before_statement |=
2059 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2060 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_DELETE);
2061 : 76521 : trigdesc->trig_delete_after_statement |=
2062 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2063 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_DELETE);
2064 : : /* there are no row-level truncate triggers */
2065 : 76521 : trigdesc->trig_truncate_before_statement |=
2066 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2067 : : TRIGGER_TYPE_BEFORE, TRIGGER_TYPE_TRUNCATE);
2068 : 76521 : trigdesc->trig_truncate_after_statement |=
2069 : 76521 : TRIGGER_TYPE_MATCHES(tgtype, TRIGGER_TYPE_STATEMENT,
2070 : : TRIGGER_TYPE_AFTER, TRIGGER_TYPE_TRUNCATE);
2071 : :
3469 kgrittn@postgresql.o 2072 : 153042 : trigdesc->trig_insert_new_table |=
2073 [ + + ]: 102087 : (TRIGGER_FOR_INSERT(tgtype) &&
2074 [ + + ]: 25566 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgnewtable));
2075 : 153042 : trigdesc->trig_update_old_table |=
2076 [ + + ]: 111050 : (TRIGGER_FOR_UPDATE(tgtype) &&
2077 [ + + ]: 34529 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgoldtable));
2078 : 153042 : trigdesc->trig_update_new_table |=
2079 [ + + ]: 111050 : (TRIGGER_FOR_UPDATE(tgtype) &&
2080 [ + + ]: 34529 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgnewtable));
2081 : 153042 : trigdesc->trig_delete_old_table |=
2082 [ + + ]: 97647 : (TRIGGER_FOR_DELETE(tgtype) &&
2083 [ + + ]: 21126 : TRIGGER_USES_TRANSITION_TABLE(trigger->tgoldtable));
10473 vadim4o@yahoo.com 2084 : 76521 : }
2085 : :
2086 : : /*
2087 : : * Copy a TriggerDesc data structure.
2088 : : *
2089 : : * The copy is allocated in the current memory context.
2090 : : */
2091 : : TriggerDesc *
8604 tgl@sss.pgh.pa.us 2092 : 287767 : CopyTriggerDesc(TriggerDesc *trigdesc)
2093 : : {
2094 : : TriggerDesc *newdesc;
2095 : : Trigger *trigger;
2096 : : int i;
2097 : :
2098 [ + + - + ]: 287767 : if (trigdesc == NULL || trigdesc->numtriggers <= 0)
2099 : 242872 : return NULL;
2100 : :
146 michael@paquier.xyz 2101 :GNC 44895 : newdesc = palloc_object(TriggerDesc);
8604 tgl@sss.pgh.pa.us 2102 :CBC 44895 : memcpy(newdesc, trigdesc, sizeof(TriggerDesc));
2103 : :
2104 : 44895 : trigger = (Trigger *) palloc(trigdesc->numtriggers * sizeof(Trigger));
2105 : 44895 : memcpy(trigger, trigdesc->triggers,
2106 : 44895 : trigdesc->numtriggers * sizeof(Trigger));
2107 : 44895 : newdesc->triggers = trigger;
2108 : :
2109 [ + + ]: 155459 : for (i = 0; i < trigdesc->numtriggers; i++)
2110 : : {
2111 : 110564 : trigger->tgname = pstrdup(trigger->tgname);
7707 2112 [ + + ]: 110564 : if (trigger->tgnattr > 0)
2113 : : {
2114 : : int16 *newattr;
2115 : :
5062 peter_e@gmx.net 2116 : 685 : newattr = (int16 *) palloc(trigger->tgnattr * sizeof(int16));
7707 tgl@sss.pgh.pa.us 2117 : 685 : memcpy(newattr, trigger->tgattr,
5062 peter_e@gmx.net 2118 : 685 : trigger->tgnattr * sizeof(int16));
7707 tgl@sss.pgh.pa.us 2119 : 685 : trigger->tgattr = newattr;
2120 : : }
8604 2121 [ + + ]: 110564 : if (trigger->tgnargs > 0)
2122 : : {
2123 : : char **newargs;
2124 : : int16 j;
2125 : :
2126 : 7212 : newargs = (char **) palloc(trigger->tgnargs * sizeof(char *));
2127 [ + + ]: 15847 : for (j = 0; j < trigger->tgnargs; j++)
2128 : 8635 : newargs[j] = pstrdup(trigger->tgargs[j]);
2129 : 7212 : trigger->tgargs = newargs;
2130 : : }
6010 2131 [ + + ]: 110564 : if (trigger->tgqual)
2132 : 983 : trigger->tgqual = pstrdup(trigger->tgqual);
3469 kgrittn@postgresql.o 2133 [ + + ]: 110564 : if (trigger->tgoldtable)
2134 : 1619 : trigger->tgoldtable = pstrdup(trigger->tgoldtable);
2135 [ + + ]: 110564 : if (trigger->tgnewtable)
2136 : 1895 : trigger->tgnewtable = pstrdup(trigger->tgnewtable);
8604 tgl@sss.pgh.pa.us 2137 : 110564 : trigger++;
2138 : : }
2139 : :
2140 : 44895 : return newdesc;
2141 : : }
2142 : :
2143 : : /*
2144 : : * Free a TriggerDesc data structure.
2145 : : */
2146 : : void
9591 2147 : 860819 : FreeTriggerDesc(TriggerDesc *trigdesc)
2148 : : {
2149 : : Trigger *trigger;
2150 : : int i;
2151 : :
2152 [ + + ]: 860819 : if (trigdesc == NULL)
2153 : 798071 : return;
2154 : :
2155 : 62748 : trigger = trigdesc->triggers;
2156 [ + + ]: 208864 : for (i = 0; i < trigdesc->numtriggers; i++)
2157 : : {
2158 : 146116 : pfree(trigger->tgname);
7707 2159 [ + + ]: 146116 : if (trigger->tgnattr > 0)
2160 : 655 : pfree(trigger->tgattr);
9591 2161 [ + + ]: 146116 : if (trigger->tgnargs > 0)
2162 : : {
2163 [ + + ]: 12554 : while (--(trigger->tgnargs) >= 0)
2164 : 7089 : pfree(trigger->tgargs[trigger->tgnargs]);
2165 : 5465 : pfree(trigger->tgargs);
2166 : : }
6010 2167 [ + + ]: 146116 : if (trigger->tgqual)
2168 : 1094 : pfree(trigger->tgqual);
3469 kgrittn@postgresql.o 2169 [ + + ]: 146116 : if (trigger->tgoldtable)
2170 : 1481 : pfree(trigger->tgoldtable);
2171 [ + + ]: 146116 : if (trigger->tgnewtable)
2172 : 1953 : pfree(trigger->tgnewtable);
9591 tgl@sss.pgh.pa.us 2173 : 146116 : trigger++;
2174 : : }
2175 : 62748 : pfree(trigdesc->triggers);
2176 : 62748 : pfree(trigdesc);
2177 : : }
2178 : :
2179 : : /*
2180 : : * Compare two TriggerDesc structures for logical equality.
2181 : : */
2182 : : #ifdef NOT_USED
2183 : : bool
2184 : : equalTriggerDescs(TriggerDesc *trigdesc1, TriggerDesc *trigdesc2)
2185 : : {
2186 : : int i,
2187 : : j;
2188 : :
2189 : : /*
2190 : : * We need not examine the hint flags, just the trigger array itself; if
2191 : : * we have the same triggers with the same types, the flags should match.
2192 : : *
2193 : : * As of 7.3 we assume trigger set ordering is significant in the
2194 : : * comparison; so we just compare corresponding slots of the two sets.
2195 : : *
2196 : : * Note: comparing the stringToNode forms of the WHEN clauses means that
2197 : : * parse column locations will affect the result. This is okay as long as
2198 : : * this function is only used for detecting exact equality, as for example
2199 : : * in checking for staleness of a cache entry.
2200 : : */
2201 : : if (trigdesc1 != NULL)
2202 : : {
2203 : : if (trigdesc2 == NULL)
2204 : : return false;
2205 : : if (trigdesc1->numtriggers != trigdesc2->numtriggers)
2206 : : return false;
2207 : : for (i = 0; i < trigdesc1->numtriggers; i++)
2208 : : {
2209 : : Trigger *trig1 = trigdesc1->triggers + i;
2210 : : Trigger *trig2 = trigdesc2->triggers + i;
2211 : :
2212 : : if (trig1->tgoid != trig2->tgoid)
2213 : : return false;
2214 : : if (strcmp(trig1->tgname, trig2->tgname) != 0)
2215 : : return false;
2216 : : if (trig1->tgfoid != trig2->tgfoid)
2217 : : return false;
2218 : : if (trig1->tgtype != trig2->tgtype)
2219 : : return false;
2220 : : if (trig1->tgenabled != trig2->tgenabled)
2221 : : return false;
2222 : : if (trig1->tgisinternal != trig2->tgisinternal)
2223 : : return false;
2224 : : if (trig1->tgisclone != trig2->tgisclone)
2225 : : return false;
2226 : : if (trig1->tgconstrrelid != trig2->tgconstrrelid)
2227 : : return false;
2228 : : if (trig1->tgconstrindid != trig2->tgconstrindid)
2229 : : return false;
2230 : : if (trig1->tgconstraint != trig2->tgconstraint)
2231 : : return false;
2232 : : if (trig1->tgdeferrable != trig2->tgdeferrable)
2233 : : return false;
2234 : : if (trig1->tginitdeferred != trig2->tginitdeferred)
2235 : : return false;
2236 : : if (trig1->tgnargs != trig2->tgnargs)
2237 : : return false;
2238 : : if (trig1->tgnattr != trig2->tgnattr)
2239 : : return false;
2240 : : if (trig1->tgnattr > 0 &&
2241 : : memcmp(trig1->tgattr, trig2->tgattr,
2242 : : trig1->tgnattr * sizeof(int16)) != 0)
2243 : : return false;
2244 : : for (j = 0; j < trig1->tgnargs; j++)
2245 : : if (strcmp(trig1->tgargs[j], trig2->tgargs[j]) != 0)
2246 : : return false;
2247 : : if (trig1->tgqual == NULL && trig2->tgqual == NULL)
2248 : : /* ok */ ;
2249 : : else if (trig1->tgqual == NULL || trig2->tgqual == NULL)
2250 : : return false;
2251 : : else if (strcmp(trig1->tgqual, trig2->tgqual) != 0)
2252 : : return false;
2253 : : if (trig1->tgoldtable == NULL && trig2->tgoldtable == NULL)
2254 : : /* ok */ ;
2255 : : else if (trig1->tgoldtable == NULL || trig2->tgoldtable == NULL)
2256 : : return false;
2257 : : else if (strcmp(trig1->tgoldtable, trig2->tgoldtable) != 0)
2258 : : return false;
2259 : : if (trig1->tgnewtable == NULL && trig2->tgnewtable == NULL)
2260 : : /* ok */ ;
2261 : : else if (trig1->tgnewtable == NULL || trig2->tgnewtable == NULL)
2262 : : return false;
2263 : : else if (strcmp(trig1->tgnewtable, trig2->tgnewtable) != 0)
2264 : : return false;
2265 : : }
2266 : : }
2267 : : else if (trigdesc2 != NULL)
2268 : : return false;
2269 : : return true;
2270 : : }
2271 : : #endif /* NOT_USED */
2272 : :
2273 : : /*
2274 : : * Check if there is a row-level trigger with transition tables that prevents
2275 : : * a table from becoming an inheritance child or partition. Return the name
2276 : : * of the first such incompatible trigger, or NULL if there is none.
2277 : : */
2278 : : const char *
3233 rhodiumtoad@postgres 2279 : 1797 : FindTriggerIncompatibleWithInheritance(TriggerDesc *trigdesc)
2280 : : {
2281 [ + + ]: 1797 : if (trigdesc != NULL)
2282 : : {
2283 : : int i;
2284 : :
2285 [ + + ]: 408 : for (i = 0; i < trigdesc->numtriggers; ++i)
2286 : : {
3186 tgl@sss.pgh.pa.us 2287 : 292 : Trigger *trigger = &trigdesc->triggers[i];
2288 : :
270 efujita@postgresql.o 2289 [ + + ]: 292 : if (!TRIGGER_FOR_ROW(trigger->tgtype))
2290 : 24 : continue;
3233 rhodiumtoad@postgres 2291 [ + - + + ]: 268 : if (trigger->tgoldtable != NULL || trigger->tgnewtable != NULL)
2292 : 8 : return trigger->tgname;
2293 : : }
2294 : : }
2295 : :
2296 : 1789 : return NULL;
2297 : : }
2298 : :
2299 : : /*
2300 : : * Call a trigger function.
2301 : : *
2302 : : * trigdata: trigger descriptor.
2303 : : * tgindx: trigger's index in finfo and instr arrays.
2304 : : * finfo: array of cached trigger function call information.
2305 : : * instr: optional array of EXPLAIN ANALYZE instrumentation state.
2306 : : * per_tuple_context: memory context to execute the function in.
2307 : : *
2308 : : * Returns the tuple (or NULL) as returned by the function.
2309 : : */
2310 : : static HeapTuple
9104 tgl@sss.pgh.pa.us 2311 : 616401 : ExecCallTriggerFunc(TriggerData *trigdata,
2312 : : int tgindx,
2313 : : FmgrInfo *finfo,
2314 : : TriggerInstrumentation *instr,
2315 : : MemoryContext per_tuple_context)
2316 : : {
2656 andres@anarazel.de 2317 : 616401 : LOCAL_FCINFO(fcinfo, 0);
2318 : : PgStat_FunctionCallUsage fcusage;
2319 : : Datum result;
2320 : : MemoryContext oldContext;
2321 : :
2322 : : /*
2323 : : * Protect against code paths that may fail to initialize transition table
2324 : : * info.
2325 : : */
3469 kgrittn@postgresql.o 2326 [ + + + + : 616401 : Assert(((TRIGGER_FIRED_BY_INSERT(trigdata->tg_event) ||
+ + + + +
- - + + -
- + ]
2327 : : TRIGGER_FIRED_BY_UPDATE(trigdata->tg_event) ||
2328 : : TRIGGER_FIRED_BY_DELETE(trigdata->tg_event)) &&
2329 : : TRIGGER_FIRED_AFTER(trigdata->tg_event) &&
2330 : : !(trigdata->tg_event & AFTER_TRIGGER_DEFERRABLE) &&
2331 : : !(trigdata->tg_event & AFTER_TRIGGER_INITDEFERRED)) ||
2332 : : (trigdata->tg_oldtable == NULL && trigdata->tg_newtable == NULL));
2333 : :
7711 tgl@sss.pgh.pa.us 2334 : 616401 : finfo += tgindx;
2335 : :
2336 : : /*
2337 : : * We cache fmgr lookup info, to avoid making the lookup again on each
2338 : : * call.
2339 : : */
9104 2340 [ + + ]: 616401 : if (finfo->fn_oid == InvalidOid)
2341 : 12579 : fmgr_info(trigdata->tg_trigger->tgfoid, finfo);
2342 : :
2343 [ - + ]: 616401 : Assert(finfo->fn_oid == trigdata->tg_trigger->tgfoid);
2344 : :
2345 : : /*
2346 : : * If doing EXPLAIN ANALYZE, start charging time to this trigger.
2347 : : */
7711 2348 [ - + ]: 616401 : if (instr)
30 andres@anarazel.de 2349 :UNC 0 : InstrStartTrigger(instr + tgindx);
2350 : :
2351 : : /*
2352 : : * Do the function evaluation in the per-tuple memory context, so that
2353 : : * leaked memory will be reclaimed once per tuple. Note in particular that
2354 : : * any new tuple created by the trigger function will live till the end of
2355 : : * the tuple cycle.
2356 : : */
9234 tgl@sss.pgh.pa.us 2357 :CBC 616401 : oldContext = MemoryContextSwitchTo(per_tuple_context);
2358 : :
2359 : : /*
2360 : : * Call the function, passing no arguments but setting a context.
2361 : : */
2656 andres@anarazel.de 2362 : 616401 : InitFunctionCallInfoData(*fcinfo, finfo, 0,
2363 : : InvalidOid, (Node *) trigdata, NULL);
2364 : :
2365 : 616401 : pgstat_init_function_usage(fcinfo, &fcusage);
2366 : :
5214 alvherre@alvh.no-ip. 2367 : 616401 : MyTriggerDepth++;
2368 [ + + ]: 616401 : PG_TRY();
2369 : : {
2656 andres@anarazel.de 2370 : 616401 : result = FunctionCallInvoke(fcinfo);
2371 : : }
2377 peter@eisentraut.org 2372 : 690 : PG_FINALLY();
2373 : : {
5214 alvherre@alvh.no-ip. 2374 : 616401 : MyTriggerDepth--;
2375 : : }
2376 [ + + ]: 616401 : PG_END_TRY();
2377 : :
6564 tgl@sss.pgh.pa.us 2378 : 615711 : pgstat_end_function_usage(&fcusage, true);
2379 : :
9234 2380 : 615711 : MemoryContextSwitchTo(oldContext);
2381 : :
2382 : : /*
2383 : : * Trigger protocol allows function to return a null pointer, but NOT to
2384 : : * set the isnull result flag.
2385 : : */
2656 andres@anarazel.de 2386 [ - + ]: 615711 : if (fcinfo->isnull)
8325 tgl@sss.pgh.pa.us 2387 [ # # ]:UBC 0 : ereport(ERROR,
2388 : : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2389 : : errmsg("trigger function %u returned null value",
2390 : : fcinfo->flinfo->fn_oid)));
2391 : :
2392 : : /*
2393 : : * If doing EXPLAIN ANALYZE, stop charging time to this trigger, and count
2394 : : * the firing of the trigger.
2395 : : */
7711 tgl@sss.pgh.pa.us 2396 [ - + ]:CBC 615711 : if (instr)
30 andres@anarazel.de 2397 :UNC 0 : InstrStopTrigger(instr + tgindx, 1);
2398 : :
9472 tgl@sss.pgh.pa.us 2399 :CBC 615711 : return (HeapTuple) DatumGetPointer(result);
2400 : : }
2401 : :
2402 : : void
8564 bruce@momjian.us 2403 : 56234 : ExecBSInsertTriggers(EState *estate, ResultRelInfo *relinfo)
2404 : : {
2405 : : TriggerDesc *trigdesc;
2406 : : int i;
2262 peter@eisentraut.org 2407 : 56234 : TriggerData LocTriggerData = {0};
2408 : :
8564 bruce@momjian.us 2409 : 56234 : trigdesc = relinfo->ri_TrigDesc;
2410 : :
2411 [ + + ]: 56234 : if (trigdesc == NULL)
2412 : 56063 : return;
5686 tgl@sss.pgh.pa.us 2413 [ + + ]: 5045 : if (!trigdesc->trig_insert_before_statement)
8564 bruce@momjian.us 2414 : 4874 : return;
2415 : :
2416 : : /* no-op if we already fired BS triggers in this context */
3152 tgl@sss.pgh.pa.us 2417 [ - + ]: 171 : if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2418 : : CMD_INSERT))
3152 tgl@sss.pgh.pa.us 2419 :UBC 0 : return;
2420 : :
8564 bruce@momjian.us 2421 :CBC 171 : LocTriggerData.type = T_TriggerData;
2422 : 171 : LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2423 : : TRIGGER_EVENT_BEFORE;
8310 2424 : 171 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
5686 tgl@sss.pgh.pa.us 2425 [ + + ]: 1501 : for (i = 0; i < trigdesc->numtriggers; i++)
2426 : : {
2427 : 1338 : Trigger *trigger = &trigdesc->triggers[i];
2428 : : HeapTuple newtuple;
2429 : :
2430 [ + + ]: 1338 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2431 : : TRIGGER_TYPE_STATEMENT,
2432 : : TRIGGER_TYPE_BEFORE,
2433 : : TRIGGER_TYPE_INSERT))
2434 : 1159 : continue;
6010 2435 [ + + ]: 179 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2436 : : NULL, NULL, NULL))
6047 2437 : 20 : continue;
2438 : :
8564 bruce@momjian.us 2439 : 159 : LocTriggerData.tg_trigger = trigger;
2440 : 159 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2441 : : i,
2442 : : relinfo->ri_TrigFunctions,
2443 : : relinfo->ri_TrigInstrument,
2444 [ + + ]: 159 : GetPerTupleMemoryContext(estate));
2445 : :
2446 [ - + ]: 151 : if (newtuple)
8325 tgl@sss.pgh.pa.us 2447 [ # # ]:UBC 0 : ereport(ERROR,
2448 : : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2449 : : errmsg("BEFORE STATEMENT trigger cannot return a value")));
2450 : : }
2451 : : }
2452 : :
2453 : : void
3233 rhodiumtoad@postgres 2454 :CBC 54440 : ExecASInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2455 : : TransitionCaptureState *transition_capture)
2456 : : {
8564 bruce@momjian.us 2457 : 54440 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2458 : :
5686 tgl@sss.pgh.pa.us 2459 [ + + + + ]: 54440 : if (trigdesc && trigdesc->trig_insert_after_statement)
1507 alvherre@alvh.no-ip. 2460 : 349 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2461 : : TRIGGER_EVENT_INSERT,
2462 : : false, NULL, NULL, NIL, NULL, transition_capture,
2463 : : false);
8564 bruce@momjian.us 2464 : 54440 : }
2465 : :
2466 : : bool
9104 tgl@sss.pgh.pa.us 2467 : 1599 : ExecBRInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2468 : : TupleTableSlot *slot)
2469 : : {
2470 : 1599 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2365 2471 : 1599 : HeapTuple newtuple = NULL;
2472 : : bool should_free;
2262 peter@eisentraut.org 2473 : 1599 : TriggerData LocTriggerData = {0};
2474 : : int i;
2475 : :
9472 tgl@sss.pgh.pa.us 2476 : 1599 : LocTriggerData.type = T_TriggerData;
8564 bruce@momjian.us 2477 : 1599 : LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2478 : : TRIGGER_EVENT_ROW |
2479 : : TRIGGER_EVENT_BEFORE;
9104 tgl@sss.pgh.pa.us 2480 : 1599 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
5686 2481 [ + + ]: 7535 : for (i = 0; i < trigdesc->numtriggers; i++)
2482 : : {
2483 : 6140 : Trigger *trigger = &trigdesc->triggers[i];
2484 : : HeapTuple oldtuple;
2485 : :
2486 [ + + ]: 6140 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2487 : : TRIGGER_TYPE_ROW,
2488 : : TRIGGER_TYPE_BEFORE,
2489 : : TRIGGER_TYPE_INSERT))
2490 : 3015 : continue;
6010 2491 [ + + ]: 3125 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2492 : : NULL, NULL, slot))
6047 2493 : 41 : continue;
2494 : :
2625 andres@anarazel.de 2495 [ + + ]: 3084 : if (!newtuple)
2496 : 1576 : newtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2497 : :
2498 : 3084 : LocTriggerData.tg_trigslot = slot;
9472 tgl@sss.pgh.pa.us 2499 : 3084 : LocTriggerData.tg_trigtuple = oldtuple = newtuple;
9104 2500 : 3084 : LocTriggerData.tg_trigger = trigger;
2501 : 3084 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2502 : : i,
2503 : : relinfo->ri_TrigFunctions,
2504 : : relinfo->ri_TrigInstrument,
9234 2505 [ + + ]: 3084 : GetPerTupleMemoryContext(estate));
10467 bruce@momjian.us 2506 [ + + ]: 3038 : if (newtuple == NULL)
2507 : : {
2728 andres@anarazel.de 2508 [ + + ]: 142 : if (should_free)
2625 2509 : 13 : heap_freetuple(oldtuple);
2510 : 142 : return false; /* "do nothing" */
2511 : : }
2512 [ + + ]: 2896 : else if (newtuple != oldtuple)
2513 : : {
452 peter@eisentraut.org 2514 : 532 : newtuple = check_modified_virtual_generated(RelationGetDescr(relinfo->ri_RelationDesc), newtuple);
2515 : :
2573 andres@anarazel.de 2516 : 532 : ExecForceStoreHeapTuple(newtuple, slot, false);
2517 : :
2518 : : /*
2519 : : * After a tuple in a partition goes through a trigger, the user
2520 : : * could have changed the partition key enough that the tuple no
2521 : : * longer fits the partition. Verify that.
2522 : : */
2239 alvherre@alvh.no-ip. 2523 [ + + ]: 532 : if (trigger->tgisclone &&
2524 [ + + ]: 44 : !ExecPartitionCheck(relinfo, slot, estate, false))
2525 [ + - ]: 16 : ereport(ERROR,
2526 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2527 : : errmsg("moving row to another partition during a BEFORE FOR EACH ROW trigger is not supported"),
2528 : : errdetail("Before executing trigger \"%s\", the row was to be in partition \"%s.%s\".",
2529 : : trigger->tgname,
2530 : : get_namespace_name(RelationGetNamespace(relinfo->ri_RelationDesc)),
2531 : : RelationGetRelationName(relinfo->ri_RelationDesc))));
2532 : :
2625 andres@anarazel.de 2533 [ + + ]: 516 : if (should_free)
2534 : 26 : heap_freetuple(oldtuple);
2535 : :
2536 : : /* signal tuple should be re-fetched if used */
2537 : 516 : newtuple = NULL;
2538 : : }
2539 : : }
2540 : :
2541 : 1395 : return true;
2542 : : }
2543 : :
2544 : : void
9104 tgl@sss.pgh.pa.us 2545 : 8411430 : ExecARInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2546 : : TupleTableSlot *slot, List *recheckIndexes,
2547 : : TransitionCaptureState *transition_capture)
2548 : : {
2549 : 8411430 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2550 : :
270 efujita@postgresql.o 2551 [ + + + + ]: 8411430 : if (relinfo->ri_FdwRoutine && transition_capture &&
2552 [ + - ]: 4 : transition_capture->tcs_insert_new_table)
2553 : : {
2554 [ - + ]: 4 : Assert(relinfo->ri_RootResultRelInfo);
2555 [ + - ]: 4 : ereport(ERROR,
2556 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2557 : : errmsg("cannot collect transition tuples from child foreign tables")));
2558 : : }
2559 : :
3233 rhodiumtoad@postgres 2560 [ + + + + : 8411426 : if ((trigdesc && trigdesc->trig_insert_after_row) ||
+ + ]
2561 [ + + ]: 40224 : (transition_capture && transition_capture->tcs_insert_new_table))
1507 alvherre@alvh.no-ip. 2562 : 445537 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2563 : : TRIGGER_EVENT_INSERT,
2564 : : true, NULL, slot,
2565 : : recheckIndexes, NULL,
2566 : : transition_capture,
2567 : : false);
8564 bruce@momjian.us 2568 : 8411426 : }
2569 : :
2570 : : bool
5686 tgl@sss.pgh.pa.us 2571 : 119 : ExecIRInsertTriggers(EState *estate, ResultRelInfo *relinfo,
2572 : : TupleTableSlot *slot)
2573 : : {
2574 : 119 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2625 andres@anarazel.de 2575 : 119 : HeapTuple newtuple = NULL;
2576 : : bool should_free;
2262 peter@eisentraut.org 2577 : 119 : TriggerData LocTriggerData = {0};
2578 : : int i;
2579 : :
5686 tgl@sss.pgh.pa.us 2580 : 119 : LocTriggerData.type = T_TriggerData;
2581 : 119 : LocTriggerData.tg_event = TRIGGER_EVENT_INSERT |
2582 : : TRIGGER_EVENT_ROW |
2583 : : TRIGGER_EVENT_INSTEAD;
2584 : 119 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2585 [ + + ]: 362 : for (i = 0; i < trigdesc->numtriggers; i++)
2586 : : {
2587 : 255 : Trigger *trigger = &trigdesc->triggers[i];
2588 : : HeapTuple oldtuple;
2589 : :
2590 [ + + ]: 255 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2591 : : TRIGGER_TYPE_ROW,
2592 : : TRIGGER_TYPE_INSTEAD,
2593 : : TRIGGER_TYPE_INSERT))
2594 : 136 : continue;
2595 [ - + ]: 119 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2596 : : NULL, NULL, slot))
5686 tgl@sss.pgh.pa.us 2597 :UBC 0 : continue;
2598 : :
2625 andres@anarazel.de 2599 [ + - ]:CBC 119 : if (!newtuple)
2600 : 119 : newtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2601 : :
2602 : 119 : LocTriggerData.tg_trigslot = slot;
5686 tgl@sss.pgh.pa.us 2603 : 119 : LocTriggerData.tg_trigtuple = oldtuple = newtuple;
2604 : 119 : LocTriggerData.tg_trigger = trigger;
2605 : 119 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2606 : : i,
2607 : : relinfo->ri_TrigFunctions,
2608 : : relinfo->ri_TrigInstrument,
2609 [ + + ]: 119 : GetPerTupleMemoryContext(estate));
2610 [ + + ]: 119 : if (newtuple == NULL)
2611 : : {
2728 andres@anarazel.de 2612 [ + - ]: 12 : if (should_free)
2625 2613 : 12 : heap_freetuple(oldtuple);
2614 : 12 : return false; /* "do nothing" */
2615 : : }
2616 [ + + ]: 107 : else if (newtuple != oldtuple)
2617 : : {
2573 2618 : 36 : ExecForceStoreHeapTuple(newtuple, slot, false);
2619 : :
2625 2620 [ + - ]: 36 : if (should_free)
2621 : 36 : heap_freetuple(oldtuple);
2622 : :
2623 : : /* signal tuple should be re-fetched if used */
2624 : 36 : newtuple = NULL;
2625 : : }
2626 : : }
2627 : :
2628 : 107 : return true;
2629 : : }
2630 : :
2631 : : void
8564 bruce@momjian.us 2632 : 8585 : ExecBSDeleteTriggers(EState *estate, ResultRelInfo *relinfo)
2633 : : {
2634 : : TriggerDesc *trigdesc;
2635 : : int i;
2262 peter@eisentraut.org 2636 : 8585 : TriggerData LocTriggerData = {0};
2637 : :
8564 bruce@momjian.us 2638 : 8585 : trigdesc = relinfo->ri_TrigDesc;
2639 : :
2640 [ + + ]: 8585 : if (trigdesc == NULL)
2641 : 8526 : return;
5686 tgl@sss.pgh.pa.us 2642 [ + + ]: 1067 : if (!trigdesc->trig_delete_before_statement)
8564 bruce@momjian.us 2643 : 980 : return;
2644 : :
2645 : : /* no-op if we already fired BS triggers in this context */
3152 tgl@sss.pgh.pa.us 2646 [ + + ]: 87 : if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2647 : : CMD_DELETE))
2648 : 28 : return;
2649 : :
8564 bruce@momjian.us 2650 : 59 : LocTriggerData.type = T_TriggerData;
2651 : 59 : LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2652 : : TRIGGER_EVENT_BEFORE;
8310 2653 : 59 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
5686 tgl@sss.pgh.pa.us 2654 [ + + ]: 540 : for (i = 0; i < trigdesc->numtriggers; i++)
2655 : : {
2656 : 481 : Trigger *trigger = &trigdesc->triggers[i];
2657 : : HeapTuple newtuple;
2658 : :
2659 [ + + ]: 481 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2660 : : TRIGGER_TYPE_STATEMENT,
2661 : : TRIGGER_TYPE_BEFORE,
2662 : : TRIGGER_TYPE_DELETE))
2663 : 422 : continue;
6010 2664 [ + + ]: 59 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2665 : : NULL, NULL, NULL))
6047 2666 : 8 : continue;
2667 : :
8564 bruce@momjian.us 2668 : 51 : LocTriggerData.tg_trigger = trigger;
2669 : 51 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2670 : : i,
2671 : : relinfo->ri_TrigFunctions,
2672 : : relinfo->ri_TrigInstrument,
2673 [ + + ]: 51 : GetPerTupleMemoryContext(estate));
2674 : :
2675 [ - + ]: 51 : if (newtuple)
8325 tgl@sss.pgh.pa.us 2676 [ # # ]:UBC 0 : ereport(ERROR,
2677 : : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2678 : : errmsg("BEFORE STATEMENT trigger cannot return a value")));
2679 : : }
2680 : : }
2681 : :
2682 : : void
3233 rhodiumtoad@postgres 2683 :CBC 8462 : ExecASDeleteTriggers(EState *estate, ResultRelInfo *relinfo,
2684 : : TransitionCaptureState *transition_capture)
2685 : : {
8564 bruce@momjian.us 2686 : 8462 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2687 : :
5686 tgl@sss.pgh.pa.us 2688 [ + + + + ]: 8462 : if (trigdesc && trigdesc->trig_delete_after_statement)
1507 alvherre@alvh.no-ip. 2689 : 165 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2690 : : TRIGGER_EVENT_DELETE,
2691 : : false, NULL, NULL, NIL, NULL, transition_capture,
2692 : : false);
10473 vadim4o@yahoo.com 2693 : 8462 : }
2694 : :
2695 : : /*
2696 : : * Execute BEFORE ROW DELETE triggers.
2697 : : *
2698 : : * True indicates caller can proceed with the delete. False indicates caller
2699 : : * need to suppress the delete and additionally if requested, we need to pass
2700 : : * back the concurrently updated tuple if any.
2701 : : */
2702 : : bool
6035 tgl@sss.pgh.pa.us 2703 : 217 : ExecBRDeleteTriggers(EState *estate, EPQState *epqstate,
2704 : : ResultRelInfo *relinfo,
2705 : : ItemPointer tupleid,
2706 : : HeapTuple fdw_trigtuple,
2707 : : TupleTableSlot **epqslot,
2708 : : TM_Result *tmresult,
2709 : : TM_FailureData *tmfd,
2710 : : bool is_merge_delete)
2711 : : {
2625 andres@anarazel.de 2712 : 217 : TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
9104 tgl@sss.pgh.pa.us 2713 : 217 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
7559 2714 : 217 : bool result = true;
2262 peter@eisentraut.org 2715 : 217 : TriggerData LocTriggerData = {0};
2716 : : HeapTuple trigtuple;
2625 andres@anarazel.de 2717 : 217 : bool should_free = false;
2718 : : int i;
2719 : :
4426 noah@leadboat.com 2720 [ - + ]: 217 : Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2721 [ + + ]: 217 : if (fdw_trigtuple == NULL)
2722 : : {
2405 andres@anarazel.de 2723 : 209 : TupleTableSlot *epqslot_candidate = NULL;
2724 : :
2725 : : /*
2726 : : * Get a copy of the on-disk tuple we are planning to delete. In
2727 : : * general, if the tuple has been concurrently updated, we should
2728 : : * recheck it using EPQ. However, if this is a MERGE DELETE action,
2729 : : * we skip this EPQ recheck and leave it to the caller (it must do
2730 : : * additional rechecking, and might end up executing a different
2731 : : * action entirely).
2732 : : */
2625 2733 [ + + ]: 205 : if (!GetTupleForTrigger(estate, epqstate, relinfo, tupleid,
291 dean.a.rasheed@gmail 2734 : 209 : LockTupleExclusive, slot, !is_merge_delete,
2735 : 209 : &epqslot_candidate, tmresult, tmfd))
4426 noah@leadboat.com 2736 : 6 : return false;
2737 : :
2738 : : /*
2739 : : * If the tuple was concurrently updated and the caller of this
2740 : : * function requested for the updated tuple, skip the trigger
2741 : : * execution.
2742 : : */
2405 andres@anarazel.de 2743 [ + + + - ]: 200 : if (epqslot_candidate != NULL && epqslot != NULL)
2744 : : {
2745 : 1 : *epqslot = epqslot_candidate;
2854 akapila@postgresql.o 2746 : 1 : return false;
2747 : : }
2748 : :
2625 andres@anarazel.de 2749 : 199 : trigtuple = ExecFetchSlotHeapTuple(slot, true, &should_free);
2750 : : }
2751 : : else
2752 : : {
4426 noah@leadboat.com 2753 : 8 : trigtuple = fdw_trigtuple;
2573 andres@anarazel.de 2754 : 8 : ExecForceStoreHeapTuple(trigtuple, slot, false);
2755 : : }
2756 : :
9472 tgl@sss.pgh.pa.us 2757 : 207 : LocTriggerData.type = T_TriggerData;
8564 bruce@momjian.us 2758 : 207 : LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2759 : : TRIGGER_EVENT_ROW |
2760 : : TRIGGER_EVENT_BEFORE;
9104 tgl@sss.pgh.pa.us 2761 : 207 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
5686 2762 [ + + ]: 832 : for (i = 0; i < trigdesc->numtriggers; i++)
2763 : : {
2764 : : HeapTuple newtuple;
2765 : 665 : Trigger *trigger = &trigdesc->triggers[i];
2766 : :
2767 [ + + ]: 665 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2768 : : TRIGGER_TYPE_ROW,
2769 : : TRIGGER_TYPE_BEFORE,
2770 : : TRIGGER_TYPE_DELETE))
2771 : 454 : continue;
6010 2772 [ + + ]: 211 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2773 : : NULL, slot, NULL))
6047 2774 : 9 : continue;
2775 : :
2625 andres@anarazel.de 2776 : 202 : LocTriggerData.tg_trigslot = slot;
9472 tgl@sss.pgh.pa.us 2777 : 202 : LocTriggerData.tg_trigtuple = trigtuple;
9104 2778 : 202 : LocTriggerData.tg_trigger = trigger;
2779 : 202 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2780 : : i,
2781 : : relinfo->ri_TrigFunctions,
2782 : : relinfo->ri_TrigInstrument,
9234 2783 [ + + ]: 202 : GetPerTupleMemoryContext(estate));
10463 vadim4o@yahoo.com 2784 [ + + ]: 196 : if (newtuple == NULL)
2785 : : {
7559 tgl@sss.pgh.pa.us 2786 : 34 : result = false; /* tell caller to suppress delete */
10463 vadim4o@yahoo.com 2787 : 34 : break;
2788 : : }
9955 JanWieck@Yahoo.com 2789 [ + + ]: 162 : if (newtuple != trigtuple)
9637 2790 : 33 : heap_freetuple(newtuple);
2791 : : }
2625 andres@anarazel.de 2792 [ - + ]: 201 : if (should_free)
4426 noah@leadboat.com 2793 :UBC 0 : heap_freetuple(trigtuple);
2794 : :
7559 tgl@sss.pgh.pa.us 2795 :CBC 201 : return result;
2796 : : }
2797 : :
2798 : : /*
2799 : : * Note: is_crosspart_update must be true if the DELETE is being performed
2800 : : * as part of a cross-partition update.
2801 : : */
2802 : : void
1507 alvherre@alvh.no-ip. 2803 : 1025292 : ExecARDeleteTriggers(EState *estate,
2804 : : ResultRelInfo *relinfo,
2805 : : ItemPointer tupleid,
2806 : : HeapTuple fdw_trigtuple,
2807 : : TransitionCaptureState *transition_capture,
2808 : : bool is_crosspart_update)
2809 : : {
9104 tgl@sss.pgh.pa.us 2810 : 1025292 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2811 : :
270 efujita@postgresql.o 2812 [ + + + + ]: 1025292 : if (relinfo->ri_FdwRoutine && transition_capture &&
2813 [ + - ]: 2 : transition_capture->tcs_delete_old_table)
2814 : : {
2815 [ - + ]: 2 : Assert(relinfo->ri_RootResultRelInfo);
2816 [ + - ]: 2 : ereport(ERROR,
2817 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
2818 : : errmsg("cannot collect transition tuples from child foreign tables")));
2819 : : }
2820 : :
3233 rhodiumtoad@postgres 2821 [ + + + + : 1025290 : if ((trigdesc && trigdesc->trig_delete_after_row) ||
+ + ]
2822 [ + + ]: 3344 : (transition_capture && transition_capture->tcs_delete_old_table))
2823 : : {
754 akorotkov@postgresql 2824 : 4151 : TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2825 : :
2826 [ - + ]: 4151 : Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2827 [ + + ]: 4151 : if (fdw_trigtuple == NULL)
2828 : 4143 : GetTupleForTrigger(estate,
2829 : : NULL,
2830 : : relinfo,
2831 : : tupleid,
2832 : : LockTupleExclusive,
2833 : : slot,
2834 : : false,
2835 : : NULL,
2836 : : NULL,
2837 : : NULL);
2838 : : else
2573 andres@anarazel.de 2839 : 8 : ExecForceStoreHeapTuple(fdw_trigtuple, slot, false);
2840 : :
1507 alvherre@alvh.no-ip. 2841 : 4151 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2842 : : TRIGGER_EVENT_DELETE,
2843 : : true, slot, NULL, NIL, NULL,
2844 : : transition_capture,
2845 : : is_crosspart_update);
2846 : : }
10473 vadim4o@yahoo.com 2847 : 1025290 : }
2848 : :
2849 : : bool
5686 tgl@sss.pgh.pa.us 2850 : 39 : ExecIRDeleteTriggers(EState *estate, ResultRelInfo *relinfo,
2851 : : HeapTuple trigtuple)
2852 : : {
2853 : 39 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2625 andres@anarazel.de 2854 : 39 : TupleTableSlot *slot = ExecGetTriggerOldSlot(estate, relinfo);
2262 peter@eisentraut.org 2855 : 39 : TriggerData LocTriggerData = {0};
2856 : : int i;
2857 : :
5686 tgl@sss.pgh.pa.us 2858 : 39 : LocTriggerData.type = T_TriggerData;
2859 : 39 : LocTriggerData.tg_event = TRIGGER_EVENT_DELETE |
2860 : : TRIGGER_EVENT_ROW |
2861 : : TRIGGER_EVENT_INSTEAD;
2862 : 39 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2863 : :
2573 andres@anarazel.de 2864 : 39 : ExecForceStoreHeapTuple(trigtuple, slot, false);
2865 : :
5686 tgl@sss.pgh.pa.us 2866 [ + + ]: 234 : for (i = 0; i < trigdesc->numtriggers; i++)
2867 : : {
2868 : : HeapTuple rettuple;
2869 : 199 : Trigger *trigger = &trigdesc->triggers[i];
2870 : :
2871 [ + + ]: 199 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2872 : : TRIGGER_TYPE_ROW,
2873 : : TRIGGER_TYPE_INSTEAD,
2874 : : TRIGGER_TYPE_DELETE))
2875 : 160 : continue;
2876 [ - + ]: 39 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2877 : : NULL, slot, NULL))
5686 tgl@sss.pgh.pa.us 2878 :UBC 0 : continue;
2879 : :
2625 andres@anarazel.de 2880 :CBC 39 : LocTriggerData.tg_trigslot = slot;
5686 tgl@sss.pgh.pa.us 2881 : 39 : LocTriggerData.tg_trigtuple = trigtuple;
2882 : 39 : LocTriggerData.tg_trigger = trigger;
2883 : 39 : rettuple = ExecCallTriggerFunc(&LocTriggerData,
2884 : : i,
2885 : : relinfo->ri_TrigFunctions,
2886 : : relinfo->ri_TrigInstrument,
2887 [ + + ]: 39 : GetPerTupleMemoryContext(estate));
2888 [ + + ]: 39 : if (rettuple == NULL)
2889 : 4 : return false; /* Delete was suppressed */
2890 [ - + ]: 35 : if (rettuple != trigtuple)
5686 tgl@sss.pgh.pa.us 2891 :UBC 0 : heap_freetuple(rettuple);
2892 : : }
5686 tgl@sss.pgh.pa.us 2893 :CBC 35 : return true;
2894 : : }
2895 : :
2896 : : void
8564 bruce@momjian.us 2897 : 10211 : ExecBSUpdateTriggers(EState *estate, ResultRelInfo *relinfo)
2898 : : {
2899 : : TriggerDesc *trigdesc;
2900 : : int i;
2262 peter@eisentraut.org 2901 : 10211 : TriggerData LocTriggerData = {0};
2902 : : Bitmapset *updatedCols;
2903 : :
8564 bruce@momjian.us 2904 : 10211 : trigdesc = relinfo->ri_TrigDesc;
2905 : :
2906 [ + + ]: 10211 : if (trigdesc == NULL)
2907 : 10082 : return;
5686 tgl@sss.pgh.pa.us 2908 [ + + ]: 2743 : if (!trigdesc->trig_update_before_statement)
8564 bruce@momjian.us 2909 : 2614 : return;
2910 : :
2911 : : /* no-op if we already fired BS triggers in this context */
3152 tgl@sss.pgh.pa.us 2912 [ - + ]: 129 : if (before_stmt_triggers_fired(RelationGetRelid(relinfo->ri_RelationDesc),
2913 : : CMD_UPDATE))
3152 tgl@sss.pgh.pa.us 2914 :UBC 0 : return;
2915 : :
2916 : : /* statement-level triggers operate on the parent table */
1912 heikki.linnakangas@i 2917 [ - + ]:CBC 129 : Assert(relinfo->ri_RootResultRelInfo == NULL);
2918 : :
2919 : 129 : updatedCols = ExecGetAllUpdatedCols(relinfo, estate);
2920 : :
8564 bruce@momjian.us 2921 : 129 : LocTriggerData.type = T_TriggerData;
2922 : 129 : LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
2923 : : TRIGGER_EVENT_BEFORE;
8310 2924 : 129 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
2248 peter@eisentraut.org 2925 : 129 : LocTriggerData.tg_updatedcols = updatedCols;
5686 tgl@sss.pgh.pa.us 2926 [ + + ]: 1168 : for (i = 0; i < trigdesc->numtriggers; i++)
2927 : : {
2928 : 1039 : Trigger *trigger = &trigdesc->triggers[i];
2929 : : HeapTuple newtuple;
2930 : :
2931 [ + + ]: 1039 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
2932 : : TRIGGER_TYPE_STATEMENT,
2933 : : TRIGGER_TYPE_BEFORE,
2934 : : TRIGGER_TYPE_UPDATE))
2935 : 910 : continue;
6010 2936 [ + + ]: 129 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
2937 : : updatedCols, NULL, NULL))
6047 2938 : 4 : continue;
2939 : :
8564 bruce@momjian.us 2940 : 125 : LocTriggerData.tg_trigger = trigger;
2941 : 125 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
2942 : : i,
2943 : : relinfo->ri_TrigFunctions,
2944 : : relinfo->ri_TrigInstrument,
2945 [ + - ]: 125 : GetPerTupleMemoryContext(estate));
2946 : :
2947 [ - + ]: 125 : if (newtuple)
8325 tgl@sss.pgh.pa.us 2948 [ # # ]:UBC 0 : ereport(ERROR,
2949 : : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
2950 : : errmsg("BEFORE STATEMENT trigger cannot return a value")));
2951 : : }
2952 : : }
2953 : :
2954 : : void
3233 rhodiumtoad@postgres 2955 :CBC 9569 : ExecASUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
2956 : : TransitionCaptureState *transition_capture)
2957 : : {
8564 bruce@momjian.us 2958 : 9569 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2959 : :
2960 : : /* statement-level triggers operate on the parent table */
1912 heikki.linnakangas@i 2961 [ - + ]: 9569 : Assert(relinfo->ri_RootResultRelInfo == NULL);
2962 : :
5686 tgl@sss.pgh.pa.us 2963 [ + + + + ]: 9569 : if (trigdesc && trigdesc->trig_update_after_statement)
1507 alvherre@alvh.no-ip. 2964 : 277 : AfterTriggerSaveEvent(estate, relinfo, NULL, NULL,
2965 : : TRIGGER_EVENT_UPDATE,
2966 : : false, NULL, NULL, NIL,
2967 : : ExecGetAllUpdatedCols(relinfo, estate),
2968 : : transition_capture,
2969 : : false);
8564 bruce@momjian.us 2970 : 9569 : }
2971 : :
2972 : : bool
6035 tgl@sss.pgh.pa.us 2973 : 1572 : ExecBRUpdateTriggers(EState *estate, EPQState *epqstate,
2974 : : ResultRelInfo *relinfo,
2975 : : ItemPointer tupleid,
2976 : : HeapTuple fdw_trigtuple,
2977 : : TupleTableSlot *newslot,
2978 : : TM_Result *tmresult,
2979 : : TM_FailureData *tmfd,
2980 : : bool is_merge_update)
2981 : : {
9104 2982 : 1572 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2625 andres@anarazel.de 2983 : 1572 : TupleTableSlot *oldslot = ExecGetTriggerOldSlot(estate, relinfo);
2984 : 1572 : HeapTuple newtuple = NULL;
2985 : : HeapTuple trigtuple;
2986 : 1572 : bool should_free_trig = false;
2987 : 1572 : bool should_free_new = false;
2262 peter@eisentraut.org 2988 : 1572 : TriggerData LocTriggerData = {0};
2989 : : int i;
2990 : : Bitmapset *updatedCols;
2991 : : LockTupleMode lockmode;
2992 : :
2993 : : /* Determine lock mode to use */
4015 andres@anarazel.de 2994 : 1572 : lockmode = ExecUpdateLockMode(estate, relinfo);
2995 : :
4426 noah@leadboat.com 2996 [ - + ]: 1572 : Assert(HeapTupleIsValid(fdw_trigtuple) ^ ItemPointerIsValid(tupleid));
2997 [ + + ]: 1572 : if (fdw_trigtuple == NULL)
2998 : : {
2405 andres@anarazel.de 2999 : 1553 : TupleTableSlot *epqslot_candidate = NULL;
3000 : :
3001 : : /*
3002 : : * Get a copy of the on-disk tuple we are planning to update. In
3003 : : * general, if the tuple has been concurrently updated, we should
3004 : : * recheck it using EPQ. However, if this is a MERGE UPDATE action,
3005 : : * we skip this EPQ recheck and leave it to the caller (it must do
3006 : : * additional rechecking, and might end up executing a different
3007 : : * action entirely).
3008 : : */
2625 3009 [ + + ]: 1549 : if (!GetTupleForTrigger(estate, epqstate, relinfo, tupleid,
291 dean.a.rasheed@gmail 3010 : 1553 : lockmode, oldslot, !is_merge_update,
3011 : 1553 : &epqslot_candidate, tmresult, tmfd))
2625 andres@anarazel.de 3012 : 12 : return false; /* cancel the update action */
3013 : :
3014 : : /*
3015 : : * In READ COMMITTED isolation level it's possible that target tuple
3016 : : * was changed due to concurrent update. In that case we have a raw
3017 : : * subplan output tuple in epqslot_candidate, and need to form a new
3018 : : * insertable tuple using ExecGetUpdateNewTuple to replace the one we
3019 : : * received in newslot. Neither we nor our callers have any further
3020 : : * interest in the passed-in tuple, so it's okay to overwrite newslot
3021 : : * with the newer data.
3022 : : */
2405 3023 [ + + ]: 1537 : if (epqslot_candidate != NULL)
3024 : : {
3025 : : TupleTableSlot *epqslot_clean;
3026 : :
1861 tgl@sss.pgh.pa.us 3027 : 3 : epqslot_clean = ExecGetUpdateNewTuple(relinfo, epqslot_candidate,
3028 : : oldslot);
3029 : :
3030 : : /*
3031 : : * Typically, the caller's newslot was also generated by
3032 : : * ExecGetUpdateNewTuple, so that epqslot_clean will be the same
3033 : : * slot and copying is not needed. But do the right thing if it
3034 : : * isn't.
3035 : : */
842 3036 [ - + ]: 3 : if (unlikely(newslot != epqslot_clean))
2405 andres@anarazel.de 3037 :UBC 0 : ExecCopySlot(newslot, epqslot_clean);
3038 : :
3039 : : /*
3040 : : * At this point newslot contains a virtual tuple that may
3041 : : * reference some fields of oldslot's tuple in some disk buffer.
3042 : : * If that tuple is in a different page than the original target
3043 : : * tuple, then our only pin on that buffer is oldslot's, and we're
3044 : : * about to release it. Hence we'd better materialize newslot to
3045 : : * ensure it doesn't contain references into an unpinned buffer.
3046 : : * (We'd materialize it below anyway, but too late for safety.)
3047 : : */
842 tgl@sss.pgh.pa.us 3048 :CBC 3 : ExecMaterializeSlot(newslot);
3049 : : }
3050 : :
3051 : : /*
3052 : : * Here we convert oldslot to a materialized slot holding trigtuple.
3053 : : * Neither slot passed to the triggers will hold any buffer pin.
3054 : : */
2017 3055 : 1537 : trigtuple = ExecFetchSlotHeapTuple(oldslot, true, &should_free_trig);
3056 : : }
3057 : : else
3058 : : {
3059 : : /* Put the FDW-supplied tuple into oldslot to unify the cases */
2573 andres@anarazel.de 3060 : 19 : ExecForceStoreHeapTuple(fdw_trigtuple, oldslot, false);
4426 noah@leadboat.com 3061 : 19 : trigtuple = fdw_trigtuple;
3062 : : }
3063 : :
9472 tgl@sss.pgh.pa.us 3064 : 1556 : LocTriggerData.type = T_TriggerData;
8518 bruce@momjian.us 3065 : 1556 : LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
3066 : : TRIGGER_EVENT_ROW |
3067 : : TRIGGER_EVENT_BEFORE;
9104 tgl@sss.pgh.pa.us 3068 : 1556 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
1912 heikki.linnakangas@i 3069 : 1556 : updatedCols = ExecGetAllUpdatedCols(relinfo, estate);
2248 peter@eisentraut.org 3070 : 1556 : LocTriggerData.tg_updatedcols = updatedCols;
5686 tgl@sss.pgh.pa.us 3071 [ + + ]: 7408 : for (i = 0; i < trigdesc->numtriggers; i++)
3072 : : {
3073 : 5946 : Trigger *trigger = &trigdesc->triggers[i];
3074 : : HeapTuple oldtuple;
3075 : :
3076 [ + + ]: 5946 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3077 : : TRIGGER_TYPE_ROW,
3078 : : TRIGGER_TYPE_BEFORE,
3079 : : TRIGGER_TYPE_UPDATE))
3080 : 3018 : continue;
6010 3081 [ + + ]: 2928 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3082 : : updatedCols, oldslot, newslot))
6047 3083 : 65 : continue;
3084 : :
2625 andres@anarazel.de 3085 [ + + ]: 2863 : if (!newtuple)
3086 : 1547 : newtuple = ExecFetchSlotHeapTuple(newslot, true, &should_free_new);
3087 : :
3088 : 2863 : LocTriggerData.tg_trigslot = oldslot;
9472 tgl@sss.pgh.pa.us 3089 : 2863 : LocTriggerData.tg_trigtuple = trigtuple;
3090 : 2863 : LocTriggerData.tg_newtuple = oldtuple = newtuple;
2625 andres@anarazel.de 3091 : 2863 : LocTriggerData.tg_newslot = newslot;
9104 tgl@sss.pgh.pa.us 3092 : 2863 : LocTriggerData.tg_trigger = trigger;
3093 : 2863 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
3094 : : i,
3095 : : relinfo->ri_TrigFunctions,
3096 : : relinfo->ri_TrigInstrument,
9234 3097 [ + - ]: 2863 : GetPerTupleMemoryContext(estate));
3098 : :
10463 vadim4o@yahoo.com 3099 [ + + ]: 2855 : if (newtuple == NULL)
3100 : : {
2625 andres@anarazel.de 3101 [ - + ]: 86 : if (should_free_trig)
4426 noah@leadboat.com 3102 :UBC 0 : heap_freetuple(trigtuple);
2625 andres@anarazel.de 3103 [ + + ]:CBC 86 : if (should_free_new)
3104 : 2 : heap_freetuple(oldtuple);
3105 : 86 : return false; /* "do nothing" */
3106 : : }
3107 [ + + ]: 2769 : else if (newtuple != oldtuple)
3108 : : {
452 peter@eisentraut.org 3109 : 736 : newtuple = check_modified_virtual_generated(RelationGetDescr(relinfo->ri_RelationDesc), newtuple);
3110 : :
2573 andres@anarazel.de 3111 : 736 : ExecForceStoreHeapTuple(newtuple, newslot, false);
3112 : :
3113 : : /*
3114 : : * If the tuple returned by the trigger / being stored, is the old
3115 : : * row version, and the heap tuple passed to the trigger was
3116 : : * allocated locally, materialize the slot. Otherwise we might
3117 : : * free it while still referenced by the slot.
3118 : : */
2574 3119 [ - + - - ]: 736 : if (should_free_trig && newtuple == trigtuple)
2574 andres@anarazel.de 3120 :UBC 0 : ExecMaterializeSlot(newslot);
3121 : :
2625 andres@anarazel.de 3122 [ + + ]:CBC 736 : if (should_free_new)
3123 : 1 : heap_freetuple(oldtuple);
3124 : :
3125 : : /* signal tuple should be re-fetched if used */
3126 : 736 : newtuple = NULL;
3127 : : }
3128 : : }
3129 [ - + ]: 1462 : if (should_free_trig)
2625 andres@anarazel.de 3130 :UBC 0 : heap_freetuple(trigtuple);
3131 : :
2625 andres@anarazel.de 3132 :CBC 1462 : return true;
3133 : : }
3134 : :
3135 : : /*
3136 : : * Note: 'src_partinfo' and 'dst_partinfo', when non-NULL, refer to the source
3137 : : * and destination partitions, respectively, of a cross-partition update of
3138 : : * the root partitioned table mentioned in the query, given by 'relinfo'.
3139 : : * 'tupleid' in that case refers to the ctid of the "old" tuple in the source
3140 : : * partition, and 'newslot' contains the "new" tuple in the destination
3141 : : * partition. This interface allows to support the requirements of
3142 : : * ExecCrossPartitionUpdateForeignKey(); is_crosspart_update must be true in
3143 : : * that case.
3144 : : */
3145 : : void
9104 tgl@sss.pgh.pa.us 3146 : 2244545 : ExecARUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
3147 : : ResultRelInfo *src_partinfo,
3148 : : ResultRelInfo *dst_partinfo,
3149 : : ItemPointer tupleid,
3150 : : HeapTuple fdw_trigtuple,
3151 : : TupleTableSlot *newslot,
3152 : : List *recheckIndexes,
3153 : : TransitionCaptureState *transition_capture,
3154 : : bool is_crosspart_update)
3155 : : {
3156 : 2244545 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3157 : :
270 efujita@postgresql.o 3158 [ + + + + ]: 2244545 : if (relinfo->ri_FdwRoutine && transition_capture &&
3159 [ - + ]: 2 : (transition_capture->tcs_update_old_table ||
270 efujita@postgresql.o 3160 [ # # ]:UBC 0 : transition_capture->tcs_update_new_table))
3161 : : {
270 efujita@postgresql.o 3162 [ - + ]:CBC 2 : Assert(relinfo->ri_RootResultRelInfo);
3163 [ + - ]: 2 : ereport(ERROR,
3164 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
3165 : : errmsg("cannot collect transition tuples from child foreign tables")));
3166 : : }
3167 : :
3233 rhodiumtoad@postgres 3168 [ + + + + : 2244543 : if ((trigdesc && trigdesc->trig_update_after_row) ||
+ + ]
3169 : 247 : (transition_capture &&
3170 [ + + ]: 247 : (transition_capture->tcs_update_old_table ||
3171 [ + - ]: 12 : transition_capture->tcs_update_new_table)))
3172 : : {
3173 : : /*
3174 : : * Note: if the UPDATE is converted into a DELETE+INSERT as part of
3175 : : * update-partition-key operation, then this function is also called
3176 : : * separately for DELETE and INSERT to capture transition table rows.
3177 : : * In such case, either old tuple or new tuple can be NULL.
3178 : : */
3179 : : TupleTableSlot *oldslot;
3180 : : ResultRelInfo *tupsrc;
3181 : :
1507 alvherre@alvh.no-ip. 3182 [ + + - + : 2444 : Assert((src_partinfo != NULL && dst_partinfo != NULL) ||
- + ]
3183 : : !is_crosspart_update);
3184 : :
754 akorotkov@postgresql 3185 [ + + ]: 2444 : tupsrc = src_partinfo ? src_partinfo : relinfo;
3186 : 2444 : oldslot = ExecGetTriggerOldSlot(estate, tupsrc);
3187 : :
3188 [ + + + + ]: 2444 : if (fdw_trigtuple == NULL && ItemPointerIsValid(tupleid))
3189 : 2402 : GetTupleForTrigger(estate,
3190 : : NULL,
3191 : : tupsrc,
3192 : : tupleid,
3193 : : LockTupleExclusive,
3194 : : oldslot,
3195 : : false,
3196 : : NULL,
3197 : : NULL,
3198 : : NULL);
3199 [ + + ]: 42 : else if (fdw_trigtuple != NULL)
2573 andres@anarazel.de 3200 : 10 : ExecForceStoreHeapTuple(fdw_trigtuple, oldslot, false);
3201 : : else
754 akorotkov@postgresql 3202 : 32 : ExecClearTuple(oldslot);
3203 : :
1507 alvherre@alvh.no-ip. 3204 : 2444 : AfterTriggerSaveEvent(estate, relinfo,
3205 : : src_partinfo, dst_partinfo,
3206 : : TRIGGER_EVENT_UPDATE,
3207 : : true,
3208 : : oldslot, newslot, recheckIndexes,
3209 : : ExecGetAllUpdatedCols(relinfo, estate),
3210 : : transition_capture,
3211 : : is_crosspart_update);
3212 : : }
10473 vadim4o@yahoo.com 3213 : 2244543 : }
3214 : :
3215 : : bool
5686 tgl@sss.pgh.pa.us 3216 : 135 : ExecIRUpdateTriggers(EState *estate, ResultRelInfo *relinfo,
3217 : : HeapTuple trigtuple, TupleTableSlot *newslot)
3218 : : {
3219 : 135 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
2625 andres@anarazel.de 3220 : 135 : TupleTableSlot *oldslot = ExecGetTriggerOldSlot(estate, relinfo);
2365 tgl@sss.pgh.pa.us 3221 : 135 : HeapTuple newtuple = NULL;
3222 : : bool should_free;
2262 peter@eisentraut.org 3223 : 135 : TriggerData LocTriggerData = {0};
3224 : : int i;
3225 : :
5686 tgl@sss.pgh.pa.us 3226 : 135 : LocTriggerData.type = T_TriggerData;
3227 : 135 : LocTriggerData.tg_event = TRIGGER_EVENT_UPDATE |
3228 : : TRIGGER_EVENT_ROW |
3229 : : TRIGGER_EVENT_INSTEAD;
3230 : 135 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3231 : :
2573 andres@anarazel.de 3232 : 135 : ExecForceStoreHeapTuple(trigtuple, oldslot, false);
3233 : :
5686 tgl@sss.pgh.pa.us 3234 [ + + ]: 502 : for (i = 0; i < trigdesc->numtriggers; i++)
3235 : : {
3236 : 387 : Trigger *trigger = &trigdesc->triggers[i];
3237 : : HeapTuple oldtuple;
3238 : :
3239 [ + + ]: 387 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3240 : : TRIGGER_TYPE_ROW,
3241 : : TRIGGER_TYPE_INSTEAD,
3242 : : TRIGGER_TYPE_UPDATE))
3243 : 252 : continue;
3244 [ - + ]: 135 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3245 : : NULL, oldslot, newslot))
5686 tgl@sss.pgh.pa.us 3246 :UBC 0 : continue;
3247 : :
2625 andres@anarazel.de 3248 [ + - ]:CBC 135 : if (!newtuple)
3249 : 135 : newtuple = ExecFetchSlotHeapTuple(newslot, true, &should_free);
3250 : :
3251 : 135 : LocTriggerData.tg_trigslot = oldslot;
5552 tgl@sss.pgh.pa.us 3252 : 135 : LocTriggerData.tg_trigtuple = trigtuple;
2625 andres@anarazel.de 3253 : 135 : LocTriggerData.tg_newslot = newslot;
5552 tgl@sss.pgh.pa.us 3254 : 135 : LocTriggerData.tg_newtuple = oldtuple = newtuple;
3255 : :
5686 3256 : 135 : LocTriggerData.tg_trigger = trigger;
5552 3257 : 135 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
3258 : : i,
3259 : : relinfo->ri_TrigFunctions,
3260 : : relinfo->ri_TrigInstrument,
5686 3261 [ + + ]: 135 : GetPerTupleMemoryContext(estate));
3262 [ + + ]: 127 : if (newtuple == NULL)
3263 : : {
2625 andres@anarazel.de 3264 : 12 : return false; /* "do nothing" */
3265 : : }
3266 [ + + ]: 115 : else if (newtuple != oldtuple)
3267 : : {
2573 3268 : 92 : ExecForceStoreHeapTuple(newtuple, newslot, false);
3269 : :
2625 3270 [ + - ]: 92 : if (should_free)
3271 : 92 : heap_freetuple(oldtuple);
3272 : :
3273 : : /* signal tuple should be re-fetched if used */
3274 : 92 : newtuple = NULL;
3275 : : }
3276 : : }
3277 : :
3278 : 115 : return true;
3279 : : }
3280 : :
3281 : : void
6612 tgl@sss.pgh.pa.us 3282 : 2481 : ExecBSTruncateTriggers(EState *estate, ResultRelInfo *relinfo)
3283 : : {
3284 : : TriggerDesc *trigdesc;
3285 : : int i;
2262 peter@eisentraut.org 3286 : 2481 : TriggerData LocTriggerData = {0};
3287 : :
6612 tgl@sss.pgh.pa.us 3288 : 2481 : trigdesc = relinfo->ri_TrigDesc;
3289 : :
3290 [ + + ]: 2481 : if (trigdesc == NULL)
3291 : 2474 : return;
5686 3292 [ + + ]: 489 : if (!trigdesc->trig_truncate_before_statement)
6612 3293 : 482 : return;
3294 : :
3295 : 7 : LocTriggerData.type = T_TriggerData;
3296 : 7 : LocTriggerData.tg_event = TRIGGER_EVENT_TRUNCATE |
3297 : : TRIGGER_EVENT_BEFORE;
3298 : 7 : LocTriggerData.tg_relation = relinfo->ri_RelationDesc;
3299 : :
5686 3300 [ + + ]: 20 : for (i = 0; i < trigdesc->numtriggers; i++)
3301 : : {
3302 : 13 : Trigger *trigger = &trigdesc->triggers[i];
3303 : : HeapTuple newtuple;
3304 : :
3305 [ + + ]: 13 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
3306 : : TRIGGER_TYPE_STATEMENT,
3307 : : TRIGGER_TYPE_BEFORE,
3308 : : TRIGGER_TYPE_TRUNCATE))
3309 : 6 : continue;
6010 3310 [ - + ]: 7 : if (!TriggerEnabled(estate, relinfo, trigger, LocTriggerData.tg_event,
3311 : : NULL, NULL, NULL))
6047 tgl@sss.pgh.pa.us 3312 :UBC 0 : continue;
3313 : :
6612 tgl@sss.pgh.pa.us 3314 :CBC 7 : LocTriggerData.tg_trigger = trigger;
3315 : 7 : newtuple = ExecCallTriggerFunc(&LocTriggerData,
3316 : : i,
3317 : : relinfo->ri_TrigFunctions,
3318 : : relinfo->ri_TrigInstrument,
3319 [ - + ]: 7 : GetPerTupleMemoryContext(estate));
3320 : :
3321 [ - + ]: 7 : if (newtuple)
6612 tgl@sss.pgh.pa.us 3322 [ # # ]:UBC 0 : ereport(ERROR,
3323 : : (errcode(ERRCODE_E_R_I_E_TRIGGER_PROTOCOL_VIOLATED),
3324 : : errmsg("BEFORE STATEMENT trigger cannot return a value")));
3325 : : }
3326 : : }
3327 : :
3328 : : void
6612 tgl@sss.pgh.pa.us 3329 :CBC 2477 : ExecASTruncateTriggers(EState *estate, ResultRelInfo *relinfo)
3330 : : {
3331 : 2477 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
3332 : :
5686 3333 [ + + + + ]: 2477 : if (trigdesc && trigdesc->trig_truncate_after_statement)
1507 alvherre@alvh.no-ip. 3334 : 5 : AfterTriggerSaveEvent(estate, relinfo,
3335 : : NULL, NULL,
3336 : : TRIGGER_EVENT_TRUNCATE,
3337 : : false, NULL, NULL, NIL, NULL, NULL,
3338 : : false);
6612 tgl@sss.pgh.pa.us 3339 : 2477 : }
3340 : :
3341 : :
3342 : : /*
3343 : : * Fetch tuple into "oldslot", dealing with locking and EPQ if necessary
3344 : : */
3345 : : static bool
6051 3346 : 8307 : GetTupleForTrigger(EState *estate,
3347 : : EPQState *epqstate,
3348 : : ResultRelInfo *relinfo,
3349 : : ItemPointer tid,
3350 : : LockTupleMode lockmode,
3351 : : TupleTableSlot *oldslot,
3352 : : bool do_epq_recheck,
3353 : : TupleTableSlot **epqslot,
3354 : : TM_Result *tmresultp,
3355 : : TM_FailureData *tmfdp)
3356 : : {
9104 3357 : 8307 : Relation relation = relinfo->ri_RelationDesc;
3358 : :
2405 andres@anarazel.de 3359 [ + + ]: 8307 : if (epqslot != NULL)
3360 : : {
3361 : : TM_Result test;
3362 : : TM_FailureData tmfd;
2600 3363 : 1762 : int lockflags = 0;
3364 : :
2405 3365 : 1762 : *epqslot = NULL;
3366 : :
3367 : : /* caller must pass an epqstate if EvalPlanQual is possible */
6035 tgl@sss.pgh.pa.us 3368 [ - + ]: 1762 : Assert(epqstate != NULL);
3369 : :
3370 : : /*
3371 : : * lock tuple for update
3372 : : */
2600 andres@anarazel.de 3373 [ + + ]: 1762 : if (!IsolationUsesXactSnapshot())
3374 : 1329 : lockflags |= TUPLE_LOCK_FLAG_FIND_LAST_VERSION;
2539 3375 : 1762 : test = table_tuple_lock(relation, tid, estate->es_snapshot, oldslot,
3376 : : estate->es_output_cid,
3377 : : lockmode, LockWaitBlock,
3378 : : lockflags,
3379 : : &tmfd);
3380 : :
3381 : : /* Let the caller know about the status of this operation */
1149 dean.a.rasheed@gmail 3382 [ + + ]: 1760 : if (tmresultp)
3383 : 148 : *tmresultp = test;
1499 alvherre@alvh.no-ip. 3384 [ + + ]: 1760 : if (tmfdp)
3385 : 1757 : *tmfdp = tmfd;
3386 : :
10003 vadim4o@yahoo.com 3387 [ + + + + : 1760 : switch (test)
- - ]
3388 : : {
2600 andres@anarazel.de 3389 : 4 : case TM_SelfModified:
3390 : :
3391 : : /*
3392 : : * The target tuple was already updated or deleted by the
3393 : : * current command, or by a later command in the current
3394 : : * transaction. We ignore the tuple in the former case, and
3395 : : * throw error in the latter case, for the same reasons
3396 : : * enumerated in ExecUpdate and ExecDelete in
3397 : : * nodeModifyTable.c.
3398 : : */
3399 [ + - ]: 4 : if (tmfd.cmax != estate->es_output_cid)
4939 kgrittn@postgresql.o 3400 [ + - ]: 4 : ereport(ERROR,
3401 : : (errcode(ERRCODE_TRIGGERED_DATA_CHANGE_VIOLATION),
3402 : : errmsg("tuple to be updated was already modified by an operation triggered by the current command"),
3403 : : errhint("Consider using an AFTER trigger instead of a BEFORE trigger to propagate changes to other rows.")));
3404 : :
3405 : : /* treat it as deleted; do not process */
2625 andres@anarazel.de 3406 : 17 : return false;
3407 : :
2600 3408 : 1747 : case TM_Ok:
3409 [ + + ]: 1747 : if (tmfd.traversed)
3410 : : {
3411 : : /*
3412 : : * Recheck the tuple using EPQ, if requested. Otherwise,
3413 : : * just return that it was concurrently updated.
3414 : : */
291 dean.a.rasheed@gmail 3415 [ + + ]: 14 : if (do_epq_recheck)
3416 : : {
3417 : 6 : *epqslot = EvalPlanQual(epqstate,
3418 : : relation,
3419 : : relinfo->ri_RangeTableIndex,
3420 : : oldslot);
3421 : :
3422 : : /*
3423 : : * If PlanQual failed for updated tuple - we must not
3424 : : * process this tuple!
3425 : : */
3426 [ + - + + ]: 6 : if (TupIsNull(*epqslot))
3427 : : {
3428 : 2 : *epqslot = NULL;
3429 : 2 : return false;
3430 : : }
3431 : : }
3432 : : else
3433 : : {
3434 [ + - ]: 8 : if (tmresultp)
3435 : 8 : *tmresultp = TM_Updated;
2600 andres@anarazel.de 3436 : 8 : return false;
3437 : : }
3438 : : }
3439 : 1737 : break;
3440 : :
3441 : 1 : case TM_Updated:
3442 [ + - ]: 1 : if (IsolationUsesXactSnapshot())
3443 [ + - ]: 1 : ereport(ERROR,
3444 : : (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE),
3445 : : errmsg("could not serialize access due to concurrent update")));
2539 andres@anarazel.de 3446 [ # # ]:UBC 0 : elog(ERROR, "unexpected table_tuple_lock status: %u", test);
3447 : : break;
3448 : :
2600 andres@anarazel.de 3449 :CBC 8 : case TM_Deleted:
3450 [ + + ]: 8 : if (IsolationUsesXactSnapshot())
3451 [ + - ]: 1 : ereport(ERROR,
3452 : : (errcode(ERRCODE_T_R_SERIALIZATION_FAILURE),
3453 : : errmsg("could not serialize access due to concurrent delete")));
3454 : : /* tuple was deleted */
2625 3455 : 7 : return false;
3456 : :
2600 andres@anarazel.de 3457 :UBC 0 : case TM_Invisible:
4015 3458 [ # # ]: 0 : elog(ERROR, "attempted to lock invisible tuple");
3459 : : break;
3460 : :
10003 vadim4o@yahoo.com 3461 : 0 : default:
2539 andres@anarazel.de 3462 [ # # ]: 0 : elog(ERROR, "unrecognized table_tuple_lock status: %u", test);
3463 : : return false; /* keep compiler quiet */
3464 : : }
3465 : : }
3466 : : else
3467 : : {
3468 : : /*
3469 : : * We expect the tuple to be present, thus very simple error handling
3470 : : * suffices.
3471 : : */
2539 andres@anarazel.de 3472 [ - + ]:CBC 6545 : if (!table_tuple_fetch_row_version(relation, tid, SnapshotAny,
3473 : : oldslot))
2598 andres@anarazel.de 3474 [ # # ]:UBC 0 : elog(ERROR, "failed to fetch tuple for trigger");
3475 : : }
3476 : :
2625 andres@anarazel.de 3477 :CBC 8282 : return true;
3478 : : }
3479 : :
3480 : : /*
3481 : : * Is trigger enabled to fire?
3482 : : */
3483 : : static bool
6010 tgl@sss.pgh.pa.us 3484 : 618021 : TriggerEnabled(EState *estate, ResultRelInfo *relinfo,
3485 : : Trigger *trigger, TriggerEvent event,
3486 : : Bitmapset *modifiedCols,
3487 : : TupleTableSlot *oldslot, TupleTableSlot *newslot)
3488 : : {
3489 : : /* Check replication-role-dependent enable state */
6047 3490 [ + + ]: 618021 : if (SessionReplicationRole == SESSION_REPLICATION_ROLE_REPLICA)
3491 : : {
3492 [ + + ]: 72 : if (trigger->tgenabled == TRIGGER_FIRES_ON_ORIGIN ||
3493 [ + + ]: 47 : trigger->tgenabled == TRIGGER_DISABLED)
3494 : 49 : return false;
3495 : : }
3496 : : else /* ORIGIN or LOCAL role */
3497 : : {
3498 [ + + ]: 617949 : if (trigger->tgenabled == TRIGGER_FIRES_ON_REPLICA ||
3499 [ + + ]: 617948 : trigger->tgenabled == TRIGGER_DISABLED)
3500 : 105 : return false;
3501 : : }
3502 : :
3503 : : /*
3504 : : * Check for column-specific trigger (only possible for UPDATE, and in
3505 : : * fact we *must* ignore tgattr for other event types)
3506 : : */
3507 [ + + + + ]: 617867 : if (trigger->tgnattr > 0 && TRIGGER_FIRED_BY_UPDATE(event))
3508 : : {
3509 : : int i;
3510 : : bool modified;
3511 : :
3512 : 286 : modified = false;
3513 [ + + ]: 374 : for (i = 0; i < trigger->tgnattr; i++)
3514 : : {
3515 [ + + ]: 318 : if (bms_is_member(trigger->tgattr[i] - FirstLowInvalidHeapAttributeNumber,
3516 : : modifiedCols))
3517 : : {
3518 : 230 : modified = true;
3519 : 230 : break;
3520 : : }
3521 : : }
3522 [ + + ]: 286 : if (!modified)
3523 : 56 : return false;
3524 : : }
3525 : :
3526 : : /* Check for WHEN clause */
6010 3527 [ + + ]: 617811 : if (trigger->tgqual)
3528 : : {
3529 : : ExprState **predicate;
3530 : : ExprContext *econtext;
3531 : : MemoryContext oldContext;
3532 : : int i;
3533 : :
3534 [ - + ]: 376 : Assert(estate != NULL);
3535 : :
3536 : : /*
3537 : : * trigger is an element of relinfo->ri_TrigDesc->triggers[]; find the
3538 : : * matching element of relinfo->ri_TrigWhenExprs[]
3539 : : */
3540 : 376 : i = trigger - relinfo->ri_TrigDesc->triggers;
3541 : 376 : predicate = &relinfo->ri_TrigWhenExprs[i];
3542 : :
3543 : : /*
3544 : : * If first time through for this WHEN expression, build expression
3545 : : * nodetrees for it. Keep them in the per-query memory context so
3546 : : * they'll survive throughout the query.
3547 : : */
3339 andres@anarazel.de 3548 [ + + ]: 376 : if (*predicate == NULL)
3549 : : {
3550 : : Node *tgqual;
3551 : :
6010 tgl@sss.pgh.pa.us 3552 : 198 : oldContext = MemoryContextSwitchTo(estate->es_query_cxt);
3553 : 198 : tgqual = stringToNode(trigger->tgqual);
452 peter@eisentraut.org 3554 : 198 : tgqual = expand_generated_columns_in_expr(tgqual, relinfo->ri_RelationDesc, PRS2_OLD_VARNO);
3555 : 198 : tgqual = expand_generated_columns_in_expr(tgqual, relinfo->ri_RelationDesc, PRS2_NEW_VARNO);
3556 : : /* Change references to OLD and NEW to INNER_VAR and OUTER_VAR */
5320 tgl@sss.pgh.pa.us 3557 : 198 : ChangeVarNodes(tgqual, PRS2_OLD_VARNO, INNER_VAR, 0);
3558 : 198 : ChangeVarNodes(tgqual, PRS2_NEW_VARNO, OUTER_VAR, 0);
3559 : : /* ExecPrepareQual wants implicit-AND form */
6010 3560 : 198 : tgqual = (Node *) make_ands_implicit((Expr *) tgqual);
3339 andres@anarazel.de 3561 : 198 : *predicate = ExecPrepareQual((List *) tgqual, estate);
6010 tgl@sss.pgh.pa.us 3562 : 198 : MemoryContextSwitchTo(oldContext);
3563 : : }
3564 : :
3565 : : /*
3566 : : * We will use the EState's per-tuple context for evaluating WHEN
3567 : : * expressions (creating it if it's not already there).
3568 : : */
3569 [ + + ]: 376 : econtext = GetPerTupleExprContext(estate);
3570 : :
3571 : : /*
3572 : : * Finally evaluate the expression, making the old and/or new tuples
3573 : : * available as INNER_VAR/OUTER_VAR respectively.
3574 : : */
3575 : 376 : econtext->ecxt_innertuple = oldslot;
3576 : 376 : econtext->ecxt_outertuple = newslot;
3339 andres@anarazel.de 3577 [ + + ]: 376 : if (!ExecQual(*predicate, econtext))
6010 tgl@sss.pgh.pa.us 3578 : 210 : return false;
3579 : : }
3580 : :
6047 3581 : 617601 : return true;
3582 : : }
3583 : :
3584 : :
3585 : : /* ----------
3586 : : * After-trigger stuff
3587 : : *
3588 : : * The AfterTriggersData struct holds data about pending AFTER trigger events
3589 : : * during the current transaction tree. (BEFORE triggers are fired
3590 : : * immediately so we don't need any persistent state about them.) The struct
3591 : : * and most of its subsidiary data are kept in TopTransactionContext; however
3592 : : * some data that can be discarded sooner appears in the CurTransactionContext
3593 : : * of the relevant subtransaction. Also, the individual event records are
3594 : : * kept in a separate sub-context of TopTransactionContext. This is done
3595 : : * mainly so that it's easy to tell from a memory context dump how much space
3596 : : * is being eaten by trigger events.
3597 : : *
3598 : : * Because the list of pending events can grow large, we go to some
3599 : : * considerable effort to minimize per-event memory consumption. The event
3600 : : * records are grouped into chunks and common data for similar events in the
3601 : : * same chunk is only stored once.
3602 : : *
3603 : : * XXX We need to be able to save the per-event data in a file if it grows too
3604 : : * large.
3605 : : * ----------
3606 : : */
3607 : :
3608 : : /* Per-trigger SET CONSTRAINT status */
3609 : : typedef struct SetConstraintTriggerData
3610 : : {
3611 : : Oid sct_tgoid;
3612 : : bool sct_tgisdeferred;
3613 : : } SetConstraintTriggerData;
3614 : :
3615 : : typedef struct SetConstraintTriggerData *SetConstraintTrigger;
3616 : :
3617 : : /*
3618 : : * SET CONSTRAINT intra-transaction status.
3619 : : *
3620 : : * We make this a single palloc'd object so it can be copied and freed easily.
3621 : : *
3622 : : * all_isset and all_isdeferred are used to keep track
3623 : : * of SET CONSTRAINTS ALL {DEFERRED, IMMEDIATE}.
3624 : : *
3625 : : * trigstates[] stores per-trigger tgisdeferred settings.
3626 : : */
3627 : : typedef struct SetConstraintStateData
3628 : : {
3629 : : bool all_isset;
3630 : : bool all_isdeferred;
3631 : : int numstates; /* number of trigstates[] entries in use */
3632 : : int numalloc; /* allocated size of trigstates[] */
3633 : : SetConstraintTriggerData trigstates[FLEXIBLE_ARRAY_MEMBER];
3634 : : } SetConstraintStateData;
3635 : :
3636 : : typedef SetConstraintStateData *SetConstraintState;
3637 : :
3638 : :
3639 : : /*
3640 : : * Per-trigger-event data
3641 : : *
3642 : : * The actual per-event data, AfterTriggerEventData, includes DONE/IN_PROGRESS
3643 : : * status bits, up to two tuple CTIDs, and optionally two OIDs of partitions.
3644 : : * Each event record also has an associated AfterTriggerSharedData that is
3645 : : * shared across all instances of similar events within a "chunk".
3646 : : *
3647 : : * For row-level triggers, we arrange not to waste storage on unneeded ctid
3648 : : * fields. Updates of regular tables use two; inserts and deletes of regular
3649 : : * tables use one; foreign tables always use zero and save the tuple(s) to a
3650 : : * tuplestore. AFTER_TRIGGER_FDW_FETCH directs AfterTriggerExecute() to
3651 : : * retrieve a fresh tuple or pair of tuples from that tuplestore, while
3652 : : * AFTER_TRIGGER_FDW_REUSE directs it to use the most-recently-retrieved
3653 : : * tuple(s). This permits storing tuples once regardless of the number of
3654 : : * row-level triggers on a foreign table.
3655 : : *
3656 : : * When updates on partitioned tables cause rows to move between partitions,
3657 : : * the OIDs of both partitions are stored too, so that the tuples can be
3658 : : * fetched; such entries are marked AFTER_TRIGGER_CP_UPDATE (for "cross-
3659 : : * partition update").
3660 : : *
3661 : : * Note that we need triggers on foreign tables to be fired in exactly the
3662 : : * order they were queued, so that the tuples come out of the tuplestore in
3663 : : * the right order. To ensure that, we forbid deferrable (constraint)
3664 : : * triggers on foreign tables. This also ensures that such triggers do not
3665 : : * get deferred into outer trigger query levels, meaning that it's okay to
3666 : : * destroy the tuplestore at the end of the query level.
3667 : : *
3668 : : * Statement-level triggers always bear AFTER_TRIGGER_1CTID, though they
3669 : : * require no ctid field. We lack the flag bit space to neatly represent that
3670 : : * distinct case, and it seems unlikely to be worth much trouble.
3671 : : *
3672 : : * Note: ats_firing_id is initially zero and is set to something else when
3673 : : * AFTER_TRIGGER_IN_PROGRESS is set. It indicates which trigger firing
3674 : : * cycle the trigger will be fired in (or was fired in, if DONE is set).
3675 : : * Although this is mutable state, we can keep it in AfterTriggerSharedData
3676 : : * because all instances of the same type of event in a given event list will
3677 : : * be fired at the same time, if they were queued between the same firing
3678 : : * cycles. So we need only ensure that ats_firing_id is zero when attaching
3679 : : * a new event to an existing AfterTriggerSharedData record.
3680 : : */
3681 : : typedef uint32 TriggerFlags;
3682 : :
3683 : : #define AFTER_TRIGGER_OFFSET 0x07FFFFFF /* must be low-order bits */
3684 : : #define AFTER_TRIGGER_DONE 0x80000000
3685 : : #define AFTER_TRIGGER_IN_PROGRESS 0x40000000
3686 : : /* bits describing the size and tuple sources of this event */
3687 : : #define AFTER_TRIGGER_FDW_REUSE 0x00000000
3688 : : #define AFTER_TRIGGER_FDW_FETCH 0x20000000
3689 : : #define AFTER_TRIGGER_1CTID 0x10000000
3690 : : #define AFTER_TRIGGER_2CTID 0x30000000
3691 : : #define AFTER_TRIGGER_CP_UPDATE 0x08000000
3692 : : #define AFTER_TRIGGER_TUP_BITS 0x38000000
3693 : : typedef struct AfterTriggerSharedData *AfterTriggerShared;
3694 : :
3695 : : typedef struct AfterTriggerSharedData
3696 : : {
3697 : : TriggerEvent ats_event; /* event type indicator, see trigger.h */
3698 : : Oid ats_tgoid; /* the trigger's ID */
3699 : : Oid ats_relid; /* the relation it's on */
3700 : : Oid ats_rolid; /* role to execute the trigger */
3701 : : CommandId ats_firing_id; /* ID for firing cycle */
3702 : : struct AfterTriggersTableData *ats_table; /* transition table access */
3703 : : Bitmapset *ats_modifiedcols; /* modified columns */
3704 : : } AfterTriggerSharedData;
3705 : :
3706 : : typedef struct AfterTriggerEventData *AfterTriggerEvent;
3707 : :
3708 : : typedef struct AfterTriggerEventData
3709 : : {
3710 : : TriggerFlags ate_flags; /* status bits and offset to shared data */
3711 : : ItemPointerData ate_ctid1; /* inserted, deleted, or old updated tuple */
3712 : : ItemPointerData ate_ctid2; /* new updated tuple */
3713 : :
3714 : : /*
3715 : : * During a cross-partition update of a partitioned table, we also store
3716 : : * the OIDs of source and destination partitions that are needed to fetch
3717 : : * the old (ctid1) and the new tuple (ctid2) from, respectively.
3718 : : */
3719 : : Oid ate_src_part;
3720 : : Oid ate_dst_part;
3721 : : } AfterTriggerEventData;
3722 : :
3723 : : /* AfterTriggerEventData, minus ate_src_part, ate_dst_part */
3724 : : typedef struct AfterTriggerEventDataNoOids
3725 : : {
3726 : : TriggerFlags ate_flags;
3727 : : ItemPointerData ate_ctid1;
3728 : : ItemPointerData ate_ctid2;
3729 : : } AfterTriggerEventDataNoOids;
3730 : :
3731 : : /* AfterTriggerEventData, minus ate_*_part and ate_ctid2 */
3732 : : typedef struct AfterTriggerEventDataOneCtid
3733 : : {
3734 : : TriggerFlags ate_flags; /* status bits and offset to shared data */
3735 : : ItemPointerData ate_ctid1; /* inserted, deleted, or old updated tuple */
3736 : : } AfterTriggerEventDataOneCtid;
3737 : :
3738 : : /* AfterTriggerEventData, minus ate_*_part, ate_ctid1 and ate_ctid2 */
3739 : : typedef struct AfterTriggerEventDataZeroCtids
3740 : : {
3741 : : TriggerFlags ate_flags; /* status bits and offset to shared data */
3742 : : } AfterTriggerEventDataZeroCtids;
3743 : :
3744 : : #define SizeofTriggerEvent(evt) \
3745 : : (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_CP_UPDATE ? \
3746 : : sizeof(AfterTriggerEventData) : \
3747 : : (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_2CTID ? \
3748 : : sizeof(AfterTriggerEventDataNoOids) : \
3749 : : (((evt)->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_1CTID ? \
3750 : : sizeof(AfterTriggerEventDataOneCtid) : \
3751 : : sizeof(AfterTriggerEventDataZeroCtids))))
3752 : :
3753 : : #define GetTriggerSharedData(evt) \
3754 : : ((AfterTriggerShared) ((char *) (evt) + ((evt)->ate_flags & AFTER_TRIGGER_OFFSET)))
3755 : :
3756 : : /*
3757 : : * To avoid palloc overhead, we keep trigger events in arrays in successively-
3758 : : * larger chunks (a slightly more sophisticated version of an expansible
3759 : : * array). The space between CHUNK_DATA_START and freeptr is occupied by
3760 : : * AfterTriggerEventData records; the space between endfree and endptr is
3761 : : * occupied by AfterTriggerSharedData records.
3762 : : */
3763 : : typedef struct AfterTriggerEventChunk
3764 : : {
3765 : : struct AfterTriggerEventChunk *next; /* list link */
3766 : : char *freeptr; /* start of free space in chunk */
3767 : : char *endfree; /* end of free space in chunk */
3768 : : char *endptr; /* end of chunk */
3769 : : /* event data follows here */
3770 : : } AfterTriggerEventChunk;
3771 : :
3772 : : #define CHUNK_DATA_START(cptr) ((char *) (cptr) + MAXALIGN(sizeof(AfterTriggerEventChunk)))
3773 : :
3774 : : /* A list of events */
3775 : : typedef struct AfterTriggerEventList
3776 : : {
3777 : : AfterTriggerEventChunk *head;
3778 : : AfterTriggerEventChunk *tail;
3779 : : char *tailfree; /* freeptr of tail chunk */
3780 : : } AfterTriggerEventList;
3781 : :
3782 : : /* Macros to help in iterating over a list of events */
3783 : : #define for_each_chunk(cptr, evtlist) \
3784 : : for (cptr = (evtlist).head; cptr != NULL; cptr = cptr->next)
3785 : : #define for_each_event(eptr, cptr) \
3786 : : for (eptr = (AfterTriggerEvent) CHUNK_DATA_START(cptr); \
3787 : : (char *) eptr < (cptr)->freeptr; \
3788 : : eptr = (AfterTriggerEvent) (((char *) eptr) + SizeofTriggerEvent(eptr)))
3789 : : /* Use this if no special per-chunk processing is needed */
3790 : : #define for_each_event_chunk(eptr, cptr, evtlist) \
3791 : : for_each_chunk(cptr, evtlist) for_each_event(eptr, cptr)
3792 : :
3793 : : /* Macros for iterating from a start point that might not be list start */
3794 : : #define for_each_chunk_from(cptr) \
3795 : : for (; cptr != NULL; cptr = cptr->next)
3796 : : #define for_each_event_from(eptr, cptr) \
3797 : : for (; \
3798 : : (char *) eptr < (cptr)->freeptr; \
3799 : : eptr = (AfterTriggerEvent) (((char *) eptr) + SizeofTriggerEvent(eptr)))
3800 : :
3801 : :
3802 : : /*
3803 : : * All per-transaction data for the AFTER TRIGGERS module.
3804 : : *
3805 : : * AfterTriggersData has the following fields:
3806 : : *
3807 : : * firing_counter is incremented for each call of afterTriggerInvokeEvents.
3808 : : * We mark firable events with the current firing cycle's ID so that we can
3809 : : * tell which ones to work on. This ensures sane behavior if a trigger
3810 : : * function chooses to do SET CONSTRAINTS: the inner SET CONSTRAINTS will
3811 : : * only fire those events that weren't already scheduled for firing.
3812 : : *
3813 : : * state keeps track of the transaction-local effects of SET CONSTRAINTS.
3814 : : * This is saved and restored across failed subtransactions.
3815 : : *
3816 : : * events is the current list of deferred events. This is global across
3817 : : * all subtransactions of the current transaction. In a subtransaction
3818 : : * abort, we know that the events added by the subtransaction are at the
3819 : : * end of the list, so it is relatively easy to discard them. The event
3820 : : * list chunks themselves are stored in event_cxt.
3821 : : *
3822 : : * query_depth is the current depth of nested AfterTriggerBeginQuery calls
3823 : : * (-1 when the stack is empty).
3824 : : *
3825 : : * query_stack[query_depth] is the per-query-level data, including these fields:
3826 : : *
3827 : : * events is a list of AFTER trigger events queued by the current query.
3828 : : * None of these are valid until the matching AfterTriggerEndQuery call
3829 : : * occurs. At that point we fire immediate-mode triggers, and append any
3830 : : * deferred events to the main events list.
3831 : : *
3832 : : * fdw_tuplestore is a tuplestore containing the foreign-table tuples
3833 : : * needed by events queued by the current query. (Note: we use just one
3834 : : * tuplestore even though more than one foreign table might be involved.
3835 : : * This is okay because tuplestores don't really care what's in the tuples
3836 : : * they store; but it's possible that someday it'd break.)
3837 : : *
3838 : : * tables is a List of AfterTriggersTableData structs for target tables
3839 : : * of the current query (see below).
3840 : : *
3841 : : * maxquerydepth is just the allocated length of query_stack.
3842 : : *
3843 : : * trans_stack holds per-subtransaction data, including these fields:
3844 : : *
3845 : : * state is NULL or a pointer to a saved copy of the SET CONSTRAINTS
3846 : : * state data. Each subtransaction level that modifies that state first
3847 : : * saves a copy, which we use to restore the state if we abort.
3848 : : *
3849 : : * events is a copy of the events head/tail pointers,
3850 : : * which we use to restore those values during subtransaction abort.
3851 : : *
3852 : : * query_depth is the subtransaction-start-time value of query_depth,
3853 : : * which we similarly use to clean up at subtransaction abort.
3854 : : *
3855 : : * firing_counter is the subtransaction-start-time value of firing_counter.
3856 : : * We use this to recognize which deferred triggers were fired (or marked
3857 : : * for firing) within an aborted subtransaction.
3858 : : *
3859 : : * We use GetCurrentTransactionNestLevel() to determine the correct array
3860 : : * index in trans_stack. maxtransdepth is the number of allocated entries in
3861 : : * trans_stack. (By not keeping our own stack pointer, we can avoid trouble
3862 : : * in cases where errors during subxact abort cause multiple invocations
3863 : : * of AfterTriggerEndSubXact() at the same nesting depth.)
3864 : : *
3865 : : * We create an AfterTriggersTableData struct for each target table of the
3866 : : * current query, and each operation mode (INSERT/UPDATE/DELETE), that has
3867 : : * either transition tables or statement-level triggers. This is used to
3868 : : * hold the relevant transition tables, as well as info tracking whether
3869 : : * we already queued the statement triggers. (We use that info to prevent
3870 : : * firing the same statement triggers more than once per statement, or really
3871 : : * once per transition table set.) These structs, along with the transition
3872 : : * table tuplestores, live in the (sub)transaction's CurTransactionContext.
3873 : : * That's sufficient lifespan because we don't allow transition tables to be
3874 : : * used by deferrable triggers, so they only need to survive until
3875 : : * AfterTriggerEndQuery.
3876 : : */
3877 : : typedef struct AfterTriggersQueryData AfterTriggersQueryData;
3878 : : typedef struct AfterTriggersTransData AfterTriggersTransData;
3879 : : typedef struct AfterTriggersTableData AfterTriggersTableData;
3880 : :
3881 : : typedef struct AfterTriggersData
3882 : : {
3883 : : CommandId firing_counter; /* next firing ID to assign */
3884 : : SetConstraintState state; /* the active S C state */
3885 : : AfterTriggerEventList events; /* deferred-event list */
3886 : : MemoryContext event_cxt; /* memory context for events, if any */
3887 : :
3888 : : /* per-query-level data: */
3889 : : AfterTriggersQueryData *query_stack; /* array of structs shown below */
3890 : : int query_depth; /* current index in above array */
3891 : : int maxquerydepth; /* allocated len of above array */
3892 : :
3893 : : /* per-subtransaction-level data: */
3894 : : AfterTriggersTransData *trans_stack; /* array of structs shown below */
3895 : : int maxtransdepth; /* allocated len of above array */
3896 : :
3897 : : List *batch_callbacks; /* List of AfterTriggerCallbackItem; for
3898 : : * deferred constraints */
3899 : : bool firing_batch_callbacks; /* true when in
3900 : : * FireAfterTriggerBatchCallbacks() */
3901 : :
3902 : : /*
3903 : : * Incremented around the trigger-firing loops in AfterTriggerEndQuery,
3904 : : * AfterTriggerFireDeferred, and AfterTriggerSetState. Used by
3905 : : * AfterTriggerIsActive() to signal that after-trigger firing is active.
3906 : : */
3907 : : int firing_depth;
3908 : : } AfterTriggersData;
3909 : :
3910 : : struct AfterTriggersQueryData
3911 : : {
3912 : : AfterTriggerEventList events; /* events pending from this query */
3913 : : Tuplestorestate *fdw_tuplestore; /* foreign tuples for said events */
3914 : : List *tables; /* list of AfterTriggersTableData, see below */
3915 : : List *batch_callbacks; /* List of AfterTriggerCallbackItem */
3916 : : };
3917 : :
3918 : : struct AfterTriggersTransData
3919 : : {
3920 : : /* these fields are just for resetting at subtrans abort: */
3921 : : SetConstraintState state; /* saved S C state, or NULL if not yet saved */
3922 : : AfterTriggerEventList events; /* saved list pointer */
3923 : : int query_depth; /* saved query_depth */
3924 : : CommandId firing_counter; /* saved firing_counter */
3925 : : };
3926 : :
3927 : : struct AfterTriggersTableData
3928 : : {
3929 : : /* relid + cmdType form the lookup key for these structs: */
3930 : : Oid relid; /* target table's OID */
3931 : : CmdType cmdType; /* event type, CMD_INSERT/UPDATE/DELETE */
3932 : : bool closed; /* true when no longer OK to add tuples */
3933 : : bool before_trig_done; /* did we already queue BS triggers? */
3934 : : bool after_trig_done; /* did we already queue AS triggers? */
3935 : : AfterTriggerEventList after_trig_events; /* if so, saved list pointer */
3936 : :
3937 : : /* "old" transition table for UPDATE/DELETE, if any */
3938 : : Tuplestorestate *old_tuplestore;
3939 : : /* "new" transition table for INSERT/UPDATE, if any */
3940 : : Tuplestorestate *new_tuplestore;
3941 : :
3942 : : TupleTableSlot *storeslot; /* for converting to tuplestore's format */
3943 : : };
3944 : :
3945 : : /* Entry in afterTriggers.batch_callbacks */
3946 : : typedef struct AfterTriggerCallbackItem
3947 : : {
3948 : : AfterTriggerBatchCallback callback;
3949 : : void *arg;
3950 : : } AfterTriggerCallbackItem;
3951 : :
3952 : : static AfterTriggersData afterTriggers;
3953 : :
3954 : : static void AfterTriggerExecute(EState *estate,
3955 : : AfterTriggerEvent event,
3956 : : ResultRelInfo *relInfo,
3957 : : ResultRelInfo *src_relInfo,
3958 : : ResultRelInfo *dst_relInfo,
3959 : : TriggerDesc *trigdesc,
3960 : : FmgrInfo *finfo,
3961 : : TriggerInstrumentation *instr,
3962 : : MemoryContext per_tuple_context,
3963 : : TupleTableSlot *trig_tuple_slot1,
3964 : : TupleTableSlot *trig_tuple_slot2);
3965 : : static AfterTriggersTableData *GetAfterTriggersTableData(Oid relid,
3966 : : CmdType cmdType);
3967 : : static TupleTableSlot *GetAfterTriggersStoreSlot(AfterTriggersTableData *table,
3968 : : TupleDesc tupdesc);
3969 : : static Tuplestorestate *GetAfterTriggersTransitionTable(int event,
3970 : : TupleTableSlot *oldslot,
3971 : : TupleTableSlot *newslot,
3972 : : TransitionCaptureState *transition_capture);
3973 : : static void TransitionTableAddTuple(EState *estate,
3974 : : int event,
3975 : : TransitionCaptureState *transition_capture,
3976 : : ResultRelInfo *relinfo,
3977 : : TupleTableSlot *slot,
3978 : : TupleTableSlot *original_insert_tuple,
3979 : : Tuplestorestate *tuplestore);
3980 : : static void AfterTriggerFreeQuery(AfterTriggersQueryData *qs);
3981 : : static SetConstraintState SetConstraintStateCreate(int numalloc);
3982 : : static SetConstraintState SetConstraintStateCopy(SetConstraintState origstate);
3983 : : static SetConstraintState SetConstraintStateAddItem(SetConstraintState state,
3984 : : Oid tgoid, bool tgisdeferred);
3985 : : static void cancel_prior_stmt_triggers(Oid relid, CmdType cmdType, int tgevent);
3986 : :
3987 : : static void FireAfterTriggerBatchCallbacks(List *callbacks);
3988 : :
3989 : : /*
3990 : : * Get the FDW tuplestore for the current trigger query level, creating it
3991 : : * if necessary.
3992 : : */
3993 : : static Tuplestorestate *
3153 3994 : 50 : GetCurrentFDWTuplestore(void)
3995 : : {
3996 : : Tuplestorestate *ret;
3997 : :
3998 : 50 : ret = afterTriggers.query_stack[afterTriggers.query_depth].fdw_tuplestore;
4426 noah@leadboat.com 3999 [ + + ]: 50 : if (ret == NULL)
4000 : : {
4001 : : MemoryContext oldcxt;
4002 : : ResourceOwner saveResourceOwner;
4003 : :
4004 : : /*
4005 : : * Make the tuplestore valid until end of subtransaction. We really
4006 : : * only need it until AfterTriggerEndQuery().
4007 : : */
3153 tgl@sss.pgh.pa.us 4008 : 18 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4426 noah@leadboat.com 4009 : 18 : saveResourceOwner = CurrentResourceOwner;
3128 tgl@sss.pgh.pa.us 4010 : 18 : CurrentResourceOwner = CurTransactionResourceOwner;
4011 : :
4012 : 18 : ret = tuplestore_begin_heap(false, false, work_mem);
4013 : :
4426 noah@leadboat.com 4014 : 18 : CurrentResourceOwner = saveResourceOwner;
4015 : 18 : MemoryContextSwitchTo(oldcxt);
4016 : :
3153 tgl@sss.pgh.pa.us 4017 : 18 : afterTriggers.query_stack[afterTriggers.query_depth].fdw_tuplestore = ret;
4018 : : }
4019 : :
4426 noah@leadboat.com 4020 : 50 : return ret;
4021 : : }
4022 : :
4023 : : /* ----------
4024 : : * afterTriggerCheckState()
4025 : : *
4026 : : * Returns true if the trigger event is actually in state DEFERRED.
4027 : : * ----------
4028 : : */
4029 : : static bool
6402 tgl@sss.pgh.pa.us 4030 : 609803 : afterTriggerCheckState(AfterTriggerShared evtshared)
4031 : : {
4032 : 609803 : Oid tgoid = evtshared->ats_tgoid;
4212 rhaas@postgresql.org 4033 : 609803 : SetConstraintState state = afterTriggers.state;
4034 : : int i;
4035 : :
4036 : : /*
4037 : : * For not-deferrable triggers (i.e. normal AFTER ROW triggers and
4038 : : * constraints declared NOT DEFERRABLE), the state is always false.
4039 : : */
6402 tgl@sss.pgh.pa.us 4040 [ + + ]: 609803 : if ((evtshared->ats_event & AFTER_TRIGGER_DEFERRABLE) == 0)
9715 JanWieck@Yahoo.com 4041 : 609262 : return false;
4042 : :
4043 : : /*
4044 : : * If constraint state exists, SET CONSTRAINTS might have been executed
4045 : : * either for this trigger or for all triggers.
4046 : : */
4212 rhaas@postgresql.org 4047 [ + + ]: 541 : if (state != NULL)
4048 : : {
4049 : : /* Check for SET CONSTRAINTS for this specific trigger. */
4050 [ + + ]: 220 : for (i = 0; i < state->numstates; i++)
4051 : : {
4052 [ + + ]: 165 : if (state->trigstates[i].sct_tgoid == tgoid)
4053 : 40 : return state->trigstates[i].sct_tgisdeferred;
4054 : : }
4055 : :
4056 : : /* Check for SET CONSTRAINTS ALL. */
4057 [ + + ]: 55 : if (state->all_isset)
4058 : 47 : return state->all_isdeferred;
4059 : : }
4060 : :
4061 : : /*
4062 : : * Otherwise return the default state for the trigger.
4063 : : */
6402 tgl@sss.pgh.pa.us 4064 : 454 : return ((evtshared->ats_event & AFTER_TRIGGER_INITDEFERRED) != 0);
4065 : : }
4066 : :
4067 : : /* ----------
4068 : : * afterTriggerCopyBitmap()
4069 : : *
4070 : : * Copy bitmap into AfterTriggerEvents memory context, which is where the after
4071 : : * trigger events are kept.
4072 : : * ----------
4073 : : */
4074 : : static Bitmapset *
1038 tomas.vondra@postgre 4075 : 7622 : afterTriggerCopyBitmap(Bitmapset *src)
4076 : : {
4077 : : Bitmapset *dst;
4078 : : MemoryContext oldcxt;
4079 : :
4080 [ + + ]: 7622 : if (src == NULL)
4081 : 5399 : return NULL;
4082 : :
4083 : 2223 : oldcxt = MemoryContextSwitchTo(afterTriggers.event_cxt);
4084 : :
4085 : 2223 : dst = bms_copy(src);
4086 : :
4087 : 2223 : MemoryContextSwitchTo(oldcxt);
4088 : :
4089 : 2223 : return dst;
4090 : : }
4091 : :
4092 : : /* ----------
4093 : : * afterTriggerAddEvent()
4094 : : *
4095 : : * Add a new trigger event to the specified queue.
4096 : : * The passed-in event data is copied.
4097 : : * ----------
4098 : : */
4099 : : static void
6402 tgl@sss.pgh.pa.us 4100 : 610288 : afterTriggerAddEvent(AfterTriggerEventList *events,
4101 : : AfterTriggerEvent event, AfterTriggerShared evtshared)
4102 : : {
4103 [ + + + + : 610288 : Size eventsize = SizeofTriggerEvent(event);
+ + ]
4104 : 610288 : Size needed = eventsize + sizeof(AfterTriggerSharedData);
4105 : : AfterTriggerEventChunk *chunk;
4106 : : AfterTriggerShared newshared;
4107 : : AfterTriggerEvent newevent;
4108 : :
4109 : : /*
4110 : : * If empty list or not enough room in the tail chunk, make a new chunk.
4111 : : * We assume here that a new shared record will always be needed.
4112 : : */
4113 : 610288 : chunk = events->tail;
4114 [ + + ]: 610288 : if (chunk == NULL ||
4115 [ + + ]: 604911 : chunk->endfree - chunk->freeptr < needed)
4116 : : {
4117 : : Size chunksize;
4118 : :
4119 : : /* Create event context if we didn't already */
4212 rhaas@postgresql.org 4120 [ + + ]: 5431 : if (afterTriggers.event_cxt == NULL)
4121 : 4475 : afterTriggers.event_cxt =
6402 tgl@sss.pgh.pa.us 4122 : 4475 : AllocSetContextCreate(TopTransactionContext,
4123 : : "AfterTriggerEvents",
4124 : : ALLOCSET_DEFAULT_SIZES);
4125 : :
4126 : : /*
4127 : : * Chunk size starts at 1KB and is allowed to increase up to 1MB.
4128 : : * These numbers are fairly arbitrary, though there is a hard limit at
4129 : : * AFTER_TRIGGER_OFFSET; else we couldn't link event records to their
4130 : : * shared records using the available space in ate_flags. Another
4131 : : * constraint is that if the chunk size gets too huge, the search loop
4132 : : * below would get slow given a (not too common) usage pattern with
4133 : : * many distinct event types in a chunk. Therefore, we double the
4134 : : * preceding chunk size only if there weren't too many shared records
4135 : : * in the preceding chunk; otherwise we halve it. This gives us some
4136 : : * ability to adapt to the actual usage pattern of the current query
4137 : : * while still having large chunk sizes in typical usage. All chunk
4138 : : * sizes used should be MAXALIGN multiples, to ensure that the shared
4139 : : * records will be aligned safely.
4140 : : */
4141 : : #define MIN_CHUNK_SIZE 1024
4142 : : #define MAX_CHUNK_SIZE (1024*1024)
4143 : :
4144 : : #if MAX_CHUNK_SIZE > (AFTER_TRIGGER_OFFSET+1)
4145 : : #error MAX_CHUNK_SIZE must not exceed AFTER_TRIGGER_OFFSET
4146 : : #endif
4147 : :
4148 [ + + ]: 5431 : if (chunk == NULL)
4149 : 5377 : chunksize = MIN_CHUNK_SIZE;
4150 : : else
4151 : : {
4152 : : /* preceding chunk size... */
6402 tgl@sss.pgh.pa.us 4153 :GBC 54 : chunksize = chunk->endptr - (char *) chunk;
4154 : : /* check number of shared records in preceding chunk */
4155 [ + - ]: 54 : if ((chunk->endptr - chunk->endfree) <=
4156 : : (100 * sizeof(AfterTriggerSharedData)))
6172 bruce@momjian.us 4157 : 54 : chunksize *= 2; /* okay, double it */
4158 : : else
6172 bruce@momjian.us 4159 :UBC 0 : chunksize /= 2; /* too many shared records */
6402 tgl@sss.pgh.pa.us 4160 :GBC 54 : chunksize = Min(chunksize, MAX_CHUNK_SIZE);
4161 : : }
4212 rhaas@postgresql.org 4162 :CBC 5431 : chunk = MemoryContextAlloc(afterTriggers.event_cxt, chunksize);
6402 tgl@sss.pgh.pa.us 4163 : 5431 : chunk->next = NULL;
4164 : 5431 : chunk->freeptr = CHUNK_DATA_START(chunk);
4165 : 5431 : chunk->endptr = chunk->endfree = (char *) chunk + chunksize;
4166 [ - + ]: 5431 : Assert(chunk->endfree - chunk->freeptr >= needed);
4167 : :
587 alvherre@alvh.no-ip. 4168 [ + + ]: 5431 : if (events->tail == NULL)
4169 : : {
4170 [ - + ]: 5377 : Assert(events->head == NULL);
6402 tgl@sss.pgh.pa.us 4171 : 5377 : events->head = chunk;
4172 : : }
4173 : : else
6402 tgl@sss.pgh.pa.us 4174 :GBC 54 : events->tail->next = chunk;
6402 tgl@sss.pgh.pa.us 4175 :CBC 5431 : events->tail = chunk;
4176 : : /* events->tailfree is now out of sync, but we'll fix it below */
4177 : : }
4178 : :
4179 : : /*
4180 : : * Try to locate a matching shared-data record already in the chunk. If
4181 : : * none, make a new one. The search begins with the most recently added
4182 : : * record, since newer ones are most likely to match.
4183 : : */
467 4184 : 610288 : for (newshared = (AfterTriggerShared) chunk->endfree;
4185 [ + + ]: 813852 : (char *) newshared < chunk->endptr;
4186 : 203564 : newshared++)
4187 : : {
4188 : : /* compare fields roughly by probability of them being different */
6402 4189 [ + + ]: 806230 : if (newshared->ats_tgoid == evtshared->ats_tgoid &&
4190 [ + + ]: 602811 : newshared->ats_event == evtshared->ats_event &&
468 4191 [ + + ]: 602807 : newshared->ats_firing_id == 0 &&
3153 4192 [ + - ]: 602691 : newshared->ats_table == evtshared->ats_table &&
468 4193 [ + - ]: 602691 : newshared->ats_relid == evtshared->ats_relid &&
467 4194 [ + + + + ]: 1205378 : newshared->ats_rolid == evtshared->ats_rolid &&
468 4195 : 602687 : bms_equal(newshared->ats_modifiedcols,
4196 : 602687 : evtshared->ats_modifiedcols))
6402 4197 : 602666 : break;
4198 : : }
467 4199 [ + + ]: 610288 : if ((char *) newshared >= chunk->endptr)
4200 : : {
4201 : 7622 : newshared = ((AfterTriggerShared) chunk->endfree) - 1;
6402 4202 : 7622 : *newshared = *evtshared;
4203 : : /* now we must make a suitably-long-lived copy of the bitmap */
468 4204 : 7622 : newshared->ats_modifiedcols = afterTriggerCopyBitmap(evtshared->ats_modifiedcols);
6172 bruce@momjian.us 4205 : 7622 : newshared->ats_firing_id = 0; /* just to be sure */
6402 tgl@sss.pgh.pa.us 4206 : 7622 : chunk->endfree = (char *) newshared;
4207 : : }
4208 : :
4209 : : /* Insert the data */
4210 : 610288 : newevent = (AfterTriggerEvent) chunk->freeptr;
4211 : 610288 : memcpy(newevent, event, eventsize);
4212 : : /* ... and link the new event to its shared record */
4213 : 610288 : newevent->ate_flags &= ~AFTER_TRIGGER_OFFSET;
4214 : 610288 : newevent->ate_flags |= (char *) newshared - (char *) newevent;
4215 : :
4216 : 610288 : chunk->freeptr += eventsize;
4217 : 610288 : events->tailfree = chunk->freeptr;
4218 : 610288 : }
4219 : :
4220 : : /* ----------
4221 : : * afterTriggerFreeEventList()
4222 : : *
4223 : : * Free all the event storage in the given list.
4224 : : * ----------
4225 : : */
4226 : : static void
4227 : 10571 : afterTriggerFreeEventList(AfterTriggerEventList *events)
4228 : : {
4229 : : AfterTriggerEventChunk *chunk;
4230 : :
3152 4231 [ + + ]: 14920 : while ((chunk = events->head) != NULL)
4232 : : {
4233 : 4349 : events->head = chunk->next;
6402 4234 : 4349 : pfree(chunk);
4235 : : }
4236 : 10571 : events->tail = NULL;
4237 : 10571 : events->tailfree = NULL;
4238 : 10571 : }
4239 : :
4240 : : /* ----------
4241 : : * afterTriggerRestoreEventList()
4242 : : *
4243 : : * Restore an event list to its prior length, removing all the events
4244 : : * added since it had the value old_events.
4245 : : * ----------
4246 : : */
4247 : : static void
4248 : 5380 : afterTriggerRestoreEventList(AfterTriggerEventList *events,
4249 : : const AfterTriggerEventList *old_events)
4250 : : {
4251 : : AfterTriggerEventChunk *chunk;
4252 : : AfterTriggerEventChunk *next_chunk;
4253 : :
4254 [ + + ]: 5380 : if (old_events->tail == NULL)
4255 : : {
4256 : : /* restoring to a completely empty state, so free everything */
4257 : 5366 : afterTriggerFreeEventList(events);
4258 : : }
4259 : : else
4260 : : {
4261 : 14 : *events = *old_events;
4262 : : /* free any chunks after the last one we want to keep */
4263 [ - + ]: 14 : for (chunk = events->tail->next; chunk != NULL; chunk = next_chunk)
4264 : : {
6402 tgl@sss.pgh.pa.us 4265 :UBC 0 : next_chunk = chunk->next;
4266 : 0 : pfree(chunk);
4267 : : }
4268 : : /* and clean up the tail chunk to be the right length */
6402 tgl@sss.pgh.pa.us 4269 :CBC 14 : events->tail->next = NULL;
4270 : 14 : events->tail->freeptr = events->tailfree;
4271 : :
4272 : : /*
4273 : : * We don't make any effort to remove now-unused shared data records.
4274 : : * They might still be useful, anyway.
4275 : : */
4276 : : }
9715 JanWieck@Yahoo.com 4277 : 5380 : }
4278 : :
4279 : : /* ----------
4280 : : * afterTriggerDeleteHeadEventChunk()
4281 : : *
4282 : : * Remove the first chunk of events from the query level's event list.
4283 : : * Keep any event list pointers elsewhere in the query level's data
4284 : : * structures in sync.
4285 : : * ----------
4286 : : */
4287 : : static void
3152 tgl@sss.pgh.pa.us 4288 :UBC 0 : afterTriggerDeleteHeadEventChunk(AfterTriggersQueryData *qs)
4289 : : {
4290 : 0 : AfterTriggerEventChunk *target = qs->events.head;
4291 : : ListCell *lc;
4292 : :
4293 [ # # # # ]: 0 : Assert(target && target->next);
4294 : :
4295 : : /*
4296 : : * First, update any pointers in the per-table data, so that they won't be
4297 : : * dangling. Resetting obsoleted pointers to NULL will make
4298 : : * cancel_prior_stmt_triggers start from the list head, which is fine.
4299 : : */
4300 [ # # # # : 0 : foreach(lc, qs->tables)
# # ]
4301 : : {
4302 : 0 : AfterTriggersTableData *table = (AfterTriggersTableData *) lfirst(lc);
4303 : :
4304 [ # # ]: 0 : if (table->after_trig_done &&
4305 [ # # ]: 0 : table->after_trig_events.tail == target)
4306 : : {
4307 : 0 : table->after_trig_events.head = NULL;
4308 : 0 : table->after_trig_events.tail = NULL;
4309 : 0 : table->after_trig_events.tailfree = NULL;
4310 : : }
4311 : : }
4312 : :
4313 : : /* Now we can flush the head chunk */
4314 : 0 : qs->events.head = target->next;
4315 : 0 : pfree(target);
4316 : 0 : }
4317 : :
4318 : :
4319 : : /* ----------
4320 : : * AfterTriggerExecute()
4321 : : *
4322 : : * Fetch the required tuples back from the heap and fire one
4323 : : * single trigger function.
4324 : : *
4325 : : * Frequently, this will be fired many times in a row for triggers of
4326 : : * a single relation. Therefore, we cache the open relation and provide
4327 : : * fmgr lookup cache space at the caller level. (For triggers fired at
4328 : : * the end of a query, we can even piggyback on the executor's state.)
4329 : : *
4330 : : * When fired for a cross-partition update of a partitioned table, the old
4331 : : * tuple is fetched using 'src_relInfo' (the source leaf partition) and
4332 : : * the new tuple using 'dst_relInfo' (the destination leaf partition), though
4333 : : * both are converted into the root partitioned table's format before passing
4334 : : * to the trigger function.
4335 : : *
4336 : : * event: event currently being fired.
4337 : : * relInfo: result relation for event.
4338 : : * src_relInfo: source partition of a cross-partition update
4339 : : * dst_relInfo: its destination partition
4340 : : * trigdesc: working copy of rel's trigger info.
4341 : : * finfo: array of fmgr lookup cache entries (one per trigger in trigdesc).
4342 : : * instr: array of EXPLAIN ANALYZE instrumentation nodes (one per trigger),
4343 : : * or NULL if no instrumentation is wanted.
4344 : : * per_tuple_context: memory context to call trigger function in.
4345 : : * trig_tuple_slot1: scratch slot for tg_trigtuple (foreign tables only)
4346 : : * trig_tuple_slot2: scratch slot for tg_newtuple (foreign tables only)
4347 : : * ----------
4348 : : */
4349 : : static void
2625 andres@anarazel.de 4350 :CBC 609621 : AfterTriggerExecute(EState *estate,
4351 : : AfterTriggerEvent event,
4352 : : ResultRelInfo *relInfo,
4353 : : ResultRelInfo *src_relInfo,
4354 : : ResultRelInfo *dst_relInfo,
4355 : : TriggerDesc *trigdesc,
4356 : : FmgrInfo *finfo, TriggerInstrumentation *instr,
4357 : : MemoryContext per_tuple_context,
4358 : : TupleTableSlot *trig_tuple_slot1,
4359 : : TupleTableSlot *trig_tuple_slot2)
4360 : : {
4361 : 609621 : Relation rel = relInfo->ri_RelationDesc;
1507 alvherre@alvh.no-ip. 4362 : 609621 : Relation src_rel = src_relInfo->ri_RelationDesc;
4363 : 609621 : Relation dst_rel = dst_relInfo->ri_RelationDesc;
6402 tgl@sss.pgh.pa.us 4364 : 609621 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
4365 : 609621 : Oid tgoid = evtshared->ats_tgoid;
2262 peter@eisentraut.org 4366 : 609621 : TriggerData LocTriggerData = {0};
4367 : : Oid save_rolid;
4368 : : int save_sec_context;
4369 : : HeapTuple rettuple;
4370 : : int tgindx;
2625 andres@anarazel.de 4371 : 609621 : bool should_free_trig = false;
4372 : 609621 : bool should_free_new = false;
4373 : :
4374 : : /*
4375 : : * Locate trigger in trigdesc. It might not be present, and in fact the
4376 : : * trigdesc could be NULL, if the trigger was dropped since the event was
4377 : : * queued. In that case, silently do nothing.
4378 : : */
684 tgl@sss.pgh.pa.us 4379 [ + + ]: 609621 : if (trigdesc == NULL)
4380 : 4 : return;
7711 4381 [ + - ]: 1019451 : for (tgindx = 0; tgindx < trigdesc->numtriggers; tgindx++)
4382 : : {
4383 [ + + ]: 1019451 : if (trigdesc->triggers[tgindx].tgoid == tgoid)
4384 : : {
4385 : 609617 : LocTriggerData.tg_trigger = &(trigdesc->triggers[tgindx]);
4386 : 609617 : break;
4387 : : }
4388 : : }
4389 [ - + ]: 609617 : if (LocTriggerData.tg_trigger == NULL)
684 tgl@sss.pgh.pa.us 4390 :UBC 0 : return;
4391 : :
4392 : : /*
4393 : : * If doing EXPLAIN ANALYZE, start charging time to this trigger. We want
4394 : : * to include time spent re-fetching tuples in the trigger cost.
4395 : : */
7711 tgl@sss.pgh.pa.us 4396 [ - + ]:CBC 609617 : if (instr)
30 andres@anarazel.de 4397 :UNC 0 : InstrStartTrigger(instr + tgindx);
4398 : :
4399 : : /*
4400 : : * Fetch the required tuple(s).
4401 : : */
4426 noah@leadboat.com 4402 [ + + + ]:CBC 609617 : switch (event->ate_flags & AFTER_TRIGGER_TUP_BITS)
4403 : : {
4404 : 25 : case AFTER_TRIGGER_FDW_FETCH:
4405 : : {
3153 tgl@sss.pgh.pa.us 4406 : 25 : Tuplestorestate *fdw_tuplestore = GetCurrentFDWTuplestore();
4407 : :
4426 noah@leadboat.com 4408 [ - + ]: 25 : if (!tuplestore_gettupleslot(fdw_tuplestore, true, false,
4409 : : trig_tuple_slot1))
4426 noah@leadboat.com 4410 [ # # ]:UBC 0 : elog(ERROR, "failed to fetch tuple1 for AFTER trigger");
4411 : :
4426 noah@leadboat.com 4412 [ + + ]:CBC 25 : if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) ==
4413 : 9 : TRIGGER_EVENT_UPDATE &&
4414 [ - + ]: 9 : !tuplestore_gettupleslot(fdw_tuplestore, true, false,
4415 : : trig_tuple_slot2))
4426 noah@leadboat.com 4416 [ # # ]:UBC 0 : elog(ERROR, "failed to fetch tuple2 for AFTER trigger");
4417 : : }
4418 : : pg_fallthrough;
4419 : : case AFTER_TRIGGER_FDW_REUSE:
4420 : :
4421 : : /*
4422 : : * Store tuple in the slot so that tg_trigtuple does not reference
4423 : : * tuplestore memory. (It is formally possible for the trigger
4424 : : * function to queue trigger events that add to the same
4425 : : * tuplestore, which can push other tuples out of memory.) The
4426 : : * distinction is academic, because we start with a minimal tuple
4427 : : * that is stored as a heap tuple, constructed in different memory
4428 : : * context, in the slot anyway.
4429 : : */
2625 andres@anarazel.de 4430 :CBC 29 : LocTriggerData.tg_trigslot = trig_tuple_slot1;
4431 : 29 : LocTriggerData.tg_trigtuple =
4432 : 29 : ExecFetchSlotHeapTuple(trig_tuple_slot1, true, &should_free_trig);
4433 : :
2338 efujita@postgresql.o 4434 [ + + ]: 29 : if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) ==
4435 : : TRIGGER_EVENT_UPDATE)
4436 : : {
4437 : 11 : LocTriggerData.tg_newslot = trig_tuple_slot2;
4438 : 11 : LocTriggerData.tg_newtuple =
4439 : 11 : ExecFetchSlotHeapTuple(trig_tuple_slot2, true, &should_free_new);
4440 : : }
4441 : : else
4442 : : {
4443 : 18 : LocTriggerData.tg_newtuple = NULL;
4444 : : }
4426 noah@leadboat.com 4445 : 29 : break;
4446 : :
4447 : 609588 : default:
4448 [ + + ]: 609588 : if (ItemPointerIsValid(&(event->ate_ctid1)))
4449 : : {
1507 alvherre@alvh.no-ip. 4450 : 608832 : TupleTableSlot *src_slot = ExecGetTriggerOldSlot(estate,
4451 : : src_relInfo);
4452 : :
4453 [ - + ]: 608832 : if (!table_tuple_fetch_row_version(src_rel,
4454 : : &(event->ate_ctid1),
4455 : : SnapshotAny,
4456 : : src_slot))
4426 noah@leadboat.com 4457 [ # # ]:UBC 0 : elog(ERROR, "failed to fetch tuple1 for AFTER trigger");
4458 : :
4459 : : /*
4460 : : * Store the tuple fetched from the source partition into the
4461 : : * target (root partitioned) table slot, converting if needed.
4462 : : */
1507 alvherre@alvh.no-ip. 4463 [ + + ]:CBC 608832 : if (src_relInfo != relInfo)
4464 : : {
4465 : 100 : TupleConversionMap *map = ExecGetChildToRootMap(src_relInfo);
4466 : :
4467 : 100 : LocTriggerData.tg_trigslot = ExecGetTriggerOldSlot(estate, relInfo);
4468 [ + + ]: 100 : if (map)
4469 : : {
4470 : 24 : execute_attr_map_slot(map->attrMap,
4471 : : src_slot,
4472 : : LocTriggerData.tg_trigslot);
4473 : : }
4474 : : else
4475 : 76 : ExecCopySlot(LocTriggerData.tg_trigslot, src_slot);
4476 : : }
4477 : : else
4478 : 608732 : LocTriggerData.tg_trigslot = src_slot;
2625 andres@anarazel.de 4479 : 608832 : LocTriggerData.tg_trigtuple =
2598 4480 : 608832 : ExecFetchSlotHeapTuple(LocTriggerData.tg_trigslot, false, &should_free_trig);
4481 : : }
4482 : : else
4483 : : {
4426 noah@leadboat.com 4484 : 756 : LocTriggerData.tg_trigtuple = NULL;
4485 : : }
4486 : :
4487 : : /* don't touch ctid2 if not there */
1507 alvherre@alvh.no-ip. 4488 [ + + ]: 609588 : if (((event->ate_flags & AFTER_TRIGGER_TUP_BITS) == AFTER_TRIGGER_2CTID ||
4489 [ + + + - ]: 609688 : (event->ate_flags & AFTER_TRIGGER_CP_UPDATE)) &&
4426 noah@leadboat.com 4490 : 2094 : ItemPointerIsValid(&(event->ate_ctid2)))
4491 : 2094 : {
1507 alvherre@alvh.no-ip. 4492 : 2094 : TupleTableSlot *dst_slot = ExecGetTriggerNewSlot(estate,
4493 : : dst_relInfo);
4494 : :
4495 [ - + ]: 2094 : if (!table_tuple_fetch_row_version(dst_rel,
4496 : : &(event->ate_ctid2),
4497 : : SnapshotAny,
4498 : : dst_slot))
4426 noah@leadboat.com 4499 [ # # ]:UBC 0 : elog(ERROR, "failed to fetch tuple2 for AFTER trigger");
4500 : :
4501 : : /*
4502 : : * Store the tuple fetched from the destination partition into
4503 : : * the target (root partitioned) table slot, converting if
4504 : : * needed.
4505 : : */
1507 alvherre@alvh.no-ip. 4506 [ + + ]:CBC 2094 : if (dst_relInfo != relInfo)
4507 : : {
4508 : 100 : TupleConversionMap *map = ExecGetChildToRootMap(dst_relInfo);
4509 : :
4510 : 100 : LocTriggerData.tg_newslot = ExecGetTriggerNewSlot(estate, relInfo);
4511 [ + + ]: 100 : if (map)
4512 : : {
4513 : 28 : execute_attr_map_slot(map->attrMap,
4514 : : dst_slot,
4515 : : LocTriggerData.tg_newslot);
4516 : : }
4517 : : else
4518 : 72 : ExecCopySlot(LocTriggerData.tg_newslot, dst_slot);
4519 : : }
4520 : : else
4521 : 1994 : LocTriggerData.tg_newslot = dst_slot;
2625 andres@anarazel.de 4522 : 2094 : LocTriggerData.tg_newtuple =
2598 4523 : 2094 : ExecFetchSlotHeapTuple(LocTriggerData.tg_newslot, false, &should_free_new);
4524 : : }
4525 : : else
4526 : : {
4426 noah@leadboat.com 4527 : 607494 : LocTriggerData.tg_newtuple = NULL;
4528 : : }
4529 : : }
4530 : :
4531 : : /*
4532 : : * Set up the tuplestore information to let the trigger have access to
4533 : : * transition tables. When we first make a transition table available to
4534 : : * a trigger, mark it "closed" so that it cannot change anymore. If any
4535 : : * additional events of the same type get queued in the current trigger
4536 : : * query level, they'll go into new transition tables.
4537 : : */
3233 rhodiumtoad@postgres 4538 : 609617 : LocTriggerData.tg_oldtable = LocTriggerData.tg_newtable = NULL;
3153 tgl@sss.pgh.pa.us 4539 [ + + ]: 609617 : if (evtshared->ats_table)
4540 : : {
3233 rhodiumtoad@postgres 4541 [ + + ]: 435 : if (LocTriggerData.tg_trigger->tgoldtable)
4542 : : {
101 dean.a.rasheed@gmail 4543 : 231 : LocTriggerData.tg_oldtable = evtshared->ats_table->old_tuplestore;
3153 tgl@sss.pgh.pa.us 4544 : 231 : evtshared->ats_table->closed = true;
4545 : : }
4546 : :
4547 [ + + ]: 435 : if (LocTriggerData.tg_trigger->tgnewtable)
4548 : : {
101 dean.a.rasheed@gmail 4549 : 319 : LocTriggerData.tg_newtable = evtshared->ats_table->new_tuplestore;
3153 tgl@sss.pgh.pa.us 4550 : 319 : evtshared->ats_table->closed = true;
4551 : : }
4552 : : }
4553 : :
4554 : : /*
4555 : : * Setup the remaining trigger information
4556 : : */
9472 4557 : 609617 : LocTriggerData.type = T_TriggerData;
7907 4558 : 609617 : LocTriggerData.tg_event =
6402 4559 : 609617 : evtshared->ats_event & (TRIGGER_EVENT_OPMASK | TRIGGER_EVENT_ROW);
9472 4560 : 609617 : LocTriggerData.tg_relation = rel;
2248 peter@eisentraut.org 4561 [ + + ]: 609617 : if (TRIGGER_FOR_UPDATE(LocTriggerData.tg_trigger->tgtype))
4562 : 3640 : LocTriggerData.tg_updatedcols = evtshared->ats_modifiedcols;
4563 : :
7907 tgl@sss.pgh.pa.us 4564 : 609617 : MemoryContextReset(per_tuple_context);
4565 : :
4566 : : /*
4567 : : * If necessary, become the role that was active when the trigger got
4568 : : * queued. Note that the role might have been dropped since the trigger
4569 : : * was queued, but if that is a problem, we will get an error later.
4570 : : * Checking here would still leave a race condition.
4571 : : */
467 4572 : 609617 : GetUserIdAndSecContext(&save_rolid, &save_sec_context);
4573 [ + + ]: 609617 : if (save_rolid != evtshared->ats_rolid)
4574 : 16 : SetUserIdAndSecContext(evtshared->ats_rolid,
4575 : : save_sec_context | SECURITY_LOCAL_USERID_CHANGE);
4576 : :
4577 : : /*
4578 : : * Call the trigger and throw away any possibly returned updated tuple.
4579 : : * (Don't let ExecCallTriggerFunc measure EXPLAIN time.)
4580 : : */
9104 4581 : 609617 : rettuple = ExecCallTriggerFunc(&LocTriggerData,
4582 : : tgindx,
4583 : : finfo,
4584 : : NULL,
4585 : : per_tuple_context);
4426 noah@leadboat.com 4586 [ + + ]: 609003 : if (rettuple != NULL &&
4587 [ + + ]: 2352 : rettuple != LocTriggerData.tg_trigtuple &&
4588 [ - + ]: 979 : rettuple != LocTriggerData.tg_newtuple)
9637 JanWieck@Yahoo.com 4589 :UBC 0 : heap_freetuple(rettuple);
4590 : :
4591 : : /* Restore the current role if necessary */
467 tgl@sss.pgh.pa.us 4592 [ + + ]:CBC 609003 : if (save_rolid != evtshared->ats_rolid)
4593 : 12 : SetUserIdAndSecContext(save_rolid, save_sec_context);
4594 : :
4595 : : /*
4596 : : * Release resources
4597 : : */
2625 andres@anarazel.de 4598 [ + + ]: 609003 : if (should_free_trig)
4599 : 109 : heap_freetuple(LocTriggerData.tg_trigtuple);
4600 [ + + ]: 609003 : if (should_free_new)
4601 : 91 : heap_freetuple(LocTriggerData.tg_newtuple);
4602 : :
4603 : : /* don't clear slots' contents if foreign table */
2338 efujita@postgresql.o 4604 [ + + ]: 609003 : if (trig_tuple_slot1 == NULL)
4605 : : {
4606 [ + + ]: 608968 : if (LocTriggerData.tg_trigslot)
4607 : 608250 : ExecClearTuple(LocTriggerData.tg_trigslot);
4608 [ + + ]: 608968 : if (LocTriggerData.tg_newslot)
4609 : 1897 : ExecClearTuple(LocTriggerData.tg_newslot);
4610 : : }
4611 : :
4612 : : /*
4613 : : * If doing EXPLAIN ANALYZE, stop charging time to this trigger, and count
4614 : : * the firing of the trigger.
4615 : : */
7711 tgl@sss.pgh.pa.us 4616 [ - + ]: 609003 : if (instr)
30 andres@anarazel.de 4617 :UNC 0 : InstrStopTrigger(instr + tgindx, 1);
4618 : : }
4619 : :
4620 : :
4621 : : /*
4622 : : * afterTriggerMarkEvents()
4623 : : *
4624 : : * Scan the given event list for not yet invoked events. Mark the ones
4625 : : * that can be invoked now with the current firing ID.
4626 : : *
4627 : : * If move_list isn't NULL, events that are not to be invoked now are
4628 : : * transferred to move_list.
4629 : : *
4630 : : * When immediate_only is true, do not invoke currently-deferred triggers.
4631 : : * (This will be false only at main transaction exit.)
4632 : : *
4633 : : * Returns true if any invokable events were found.
4634 : : */
4635 : : static bool
7907 tgl@sss.pgh.pa.us 4636 :CBC 398605 : afterTriggerMarkEvents(AfterTriggerEventList *events,
4637 : : AfterTriggerEventList *move_list,
4638 : : bool immediate_only)
4639 : : {
4640 : 398605 : bool found = false;
2003 noah@leadboat.com 4641 : 398605 : bool deferred_found = false;
4642 : : AfterTriggerEvent event;
4643 : : AfterTriggerEventChunk *chunk;
4644 : :
6402 tgl@sss.pgh.pa.us 4645 [ + + + + : 1014952 : for_each_event_chunk(event, chunk, *events)
+ + + + +
+ ]
4646 : : {
4647 : 610758 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
7907 4648 : 610758 : bool defer_it = false;
4649 : :
6402 4650 [ + + ]: 610758 : if (!(event->ate_flags &
4651 : : (AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS)))
4652 : : {
4653 : : /*
4654 : : * This trigger hasn't been called or scheduled yet. Check if we
4655 : : * should call it now.
4656 : : */
4657 [ + + + + ]: 610182 : if (immediate_only && afterTriggerCheckState(evtshared))
4658 : : {
7907 4659 : 451 : defer_it = true;
4660 : : }
4661 : : else
4662 : : {
4663 : : /*
4664 : : * Mark it as to be fired in this firing cycle.
4665 : : */
4212 rhaas@postgresql.org 4666 : 609731 : evtshared->ats_firing_id = afterTriggers.firing_counter;
6402 tgl@sss.pgh.pa.us 4667 : 609731 : event->ate_flags |= AFTER_TRIGGER_IN_PROGRESS;
7907 4668 : 609731 : found = true;
4669 : : }
4670 : : }
4671 : :
4672 : : /*
4673 : : * If it's deferred, move it to move_list, if requested.
4674 : : */
4675 [ + + + - ]: 610758 : if (defer_it && move_list != NULL)
4676 : : {
2003 noah@leadboat.com 4677 : 451 : deferred_found = true;
4678 : : /* add it to move_list */
6402 tgl@sss.pgh.pa.us 4679 : 451 : afterTriggerAddEvent(move_list, event, evtshared);
4680 : : /* mark original copy "done" so we don't do it again */
4681 : 451 : event->ate_flags |= AFTER_TRIGGER_DONE;
4682 : : }
4683 : : }
4684 : :
4685 : : /*
4686 : : * We could allow deferred triggers if, before the end of the
4687 : : * security-restricted operation, we were to verify that a SET CONSTRAINTS
4688 : : * ... IMMEDIATE has fired all such triggers. For now, don't bother.
4689 : : */
2003 noah@leadboat.com 4690 [ + + + + ]: 398605 : if (deferred_found && InSecurityRestrictedOperation())
4691 [ + - ]: 8 : ereport(ERROR,
4692 : : (errcode(ERRCODE_INSUFFICIENT_PRIVILEGE),
4693 : : errmsg("cannot fire deferred trigger within security-restricted operation")));
4694 : :
7907 tgl@sss.pgh.pa.us 4695 : 398597 : return found;
4696 : : }
4697 : :
4698 : : /*
4699 : : * afterTriggerInvokeEvents()
4700 : : *
4701 : : * Scan the given event list for events that are marked as to be fired
4702 : : * in the current firing cycle, and fire them.
4703 : : *
4704 : : * If estate isn't NULL, we use its result relation info to avoid repeated
4705 : : * openings and closing of trigger target relations. If it is NULL, we
4706 : : * make one locally to cache the info in case there are multiple trigger
4707 : : * events per rel.
4708 : : *
4709 : : * When delete_ok is true, it's safe to delete fully-processed events.
4710 : : * (We are not very tense about that: we simply reset a chunk to be empty
4711 : : * if all its events got fired. The objective here is just to avoid useless
4712 : : * rescanning of events when a trigger queues new events during transaction
4713 : : * end, so it's not necessary to worry much about the case where only
4714 : : * some events are fired.)
4715 : : *
4716 : : * Returns true if no unfired events remain in the list (this allows us
4717 : : * to avoid repeating afterTriggerMarkEvents).
4718 : : */
4719 : : static bool
4720 : 5188 : afterTriggerInvokeEvents(AfterTriggerEventList *events,
4721 : : CommandId firing_id,
4722 : : EState *estate,
4723 : : bool delete_ok)
4724 : : {
6402 4725 : 5188 : bool all_fired = true;
4726 : : AfterTriggerEventChunk *chunk;
4727 : : MemoryContext per_tuple_context;
6838 4728 : 5188 : bool local_estate = false;
2624 andres@anarazel.de 4729 : 5188 : ResultRelInfo *rInfo = NULL;
9104 tgl@sss.pgh.pa.us 4730 : 5188 : Relation rel = NULL;
8604 4731 : 5188 : TriggerDesc *trigdesc = NULL;
9104 4732 : 5188 : FmgrInfo *finfo = NULL;
30 andres@anarazel.de 4733 :GNC 5188 : TriggerInstrumentation *instr = NULL;
4426 noah@leadboat.com 4734 :CBC 5188 : TupleTableSlot *slot1 = NULL,
4735 : 5188 : *slot2 = NULL;
4736 : :
4737 : : /* Make a local EState if need be */
6838 tgl@sss.pgh.pa.us 4738 [ + + ]: 5188 : if (estate == NULL)
4739 : : {
4740 : 252 : estate = CreateExecutorState();
4741 : 252 : local_estate = true;
4742 : : }
4743 : :
4744 : : /* Make a per-tuple memory context for trigger function calls */
4745 : : per_tuple_context =
9234 4746 : 5188 : AllocSetContextCreate(CurrentMemoryContext,
4747 : : "AfterTriggerTupleContext",
4748 : : ALLOCSET_DEFAULT_SIZES);
4749 : :
6402 4750 [ + + ]: 9816 : for_each_chunk(chunk, *events)
4751 : : {
4752 : : AfterTriggerEvent event;
4753 : 5242 : bool all_fired_in_chunk = true;
4754 : :
4755 [ + + + + : 615261 : for_each_event(event, chunk)
+ + + + ]
4756 : : {
4757 : 610633 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
4758 : :
4759 : : /*
4760 : : * Is it one for me to fire?
4761 : : */
4762 [ + + ]: 610633 : if ((event->ate_flags & AFTER_TRIGGER_IN_PROGRESS) &&
4763 [ + - ]: 609621 : evtshared->ats_firing_id == firing_id)
9104 4764 : 609007 : {
4765 : : ResultRelInfo *src_rInfo,
4766 : : *dst_rInfo;
4767 : :
4768 : : /*
4769 : : * So let's fire it... but first, find the correct relation if
4770 : : * this is not the same relation as before.
4771 : : */
6402 4772 [ + + + + ]: 609621 : if (rel == NULL || RelationGetRelid(rel) != evtshared->ats_relid)
4773 : : {
1507 alvherre@alvh.no-ip. 4774 : 5413 : rInfo = ExecGetTriggerResultRel(estate, evtshared->ats_relid,
4775 : : NULL);
6402 tgl@sss.pgh.pa.us 4776 : 5413 : rel = rInfo->ri_RelationDesc;
4777 : : /* Catch calls with insufficient relcache refcounting */
1809 4778 [ - + ]: 5413 : Assert(!RelationHasReferenceCountZero(rel));
6402 4779 : 5413 : trigdesc = rInfo->ri_TrigDesc;
4780 : : /* caution: trigdesc could be NULL here */
4781 : 5413 : finfo = rInfo->ri_TrigFunctions;
4782 : 5413 : instr = rInfo->ri_TrigInstrument;
2338 efujita@postgresql.o 4783 [ - + ]: 5413 : if (slot1 != NULL)
4784 : : {
2338 efujita@postgresql.o 4785 :UBC 0 : ExecDropSingleTupleTableSlot(slot1);
4786 : 0 : ExecDropSingleTupleTableSlot(slot2);
4787 : 0 : slot1 = slot2 = NULL;
4788 : : }
4426 noah@leadboat.com 4789 [ + + ]:CBC 5413 : if (rel->rd_rel->relkind == RELKIND_FOREIGN_TABLE)
4790 : : {
2728 andres@anarazel.de 4791 : 19 : slot1 = MakeSingleTupleTableSlot(rel->rd_att,
4792 : : &TTSOpsMinimalTuple);
4793 : 19 : slot2 = MakeSingleTupleTableSlot(rel->rd_att,
4794 : : &TTSOpsMinimalTuple);
4795 : : }
4796 : : }
4797 : :
4798 : : /*
4799 : : * Look up source and destination partition result rels of a
4800 : : * cross-partition update event.
4801 : : */
1507 alvherre@alvh.no-ip. 4802 [ + + ]: 609621 : if ((event->ate_flags & AFTER_TRIGGER_TUP_BITS) ==
4803 : : AFTER_TRIGGER_CP_UPDATE)
4804 : : {
4805 [ + - - + ]: 100 : Assert(OidIsValid(event->ate_src_part) &&
4806 : : OidIsValid(event->ate_dst_part));
4807 : 100 : src_rInfo = ExecGetTriggerResultRel(estate,
4808 : : event->ate_src_part,
4809 : : rInfo);
4810 : 100 : dst_rInfo = ExecGetTriggerResultRel(estate,
4811 : : event->ate_dst_part,
4812 : : rInfo);
4813 : : }
4814 : : else
4815 : 609521 : src_rInfo = dst_rInfo = rInfo;
4816 : :
4817 : : /*
4818 : : * Fire it. Note that the AFTER_TRIGGER_IN_PROGRESS flag is
4819 : : * still set, so recursive examinations of the event list
4820 : : * won't try to re-fire it.
4821 : : */
4822 : 609621 : AfterTriggerExecute(estate, event, rInfo,
4823 : : src_rInfo, dst_rInfo,
4824 : : trigdesc, finfo, instr,
4825 : : per_tuple_context, slot1, slot2);
4826 : :
4827 : : /*
4828 : : * Mark the event as done.
4829 : : */
6402 tgl@sss.pgh.pa.us 4830 : 609007 : event->ate_flags &= ~AFTER_TRIGGER_IN_PROGRESS;
4831 : 609007 : event->ate_flags |= AFTER_TRIGGER_DONE;
4832 : : }
4833 [ + + ]: 1012 : else if (!(event->ate_flags & AFTER_TRIGGER_DONE))
4834 : : {
4835 : : /* something remains to be done */
4836 : 352 : all_fired = all_fired_in_chunk = false;
4837 : : }
4838 : : }
4839 : :
4840 : : /* Clear the chunk if delete_ok and nothing left of interest */
4841 [ + + + - ]: 4628 : if (delete_ok && all_fired_in_chunk)
4842 : : {
4843 : 177 : chunk->freeptr = CHUNK_DATA_START(chunk);
4844 : 177 : chunk->endfree = chunk->endptr;
4845 : :
4846 : : /*
4847 : : * If it's last chunk, must sync event list's tailfree too. Note
4848 : : * that delete_ok must NOT be passed as true if there could be
4849 : : * additional AfterTriggerEventList values pointing at this event
4850 : : * list, since we'd fail to fix their copies of tailfree.
4851 : : */
5738 4852 [ + - ]: 177 : if (chunk == events->tail)
4853 : 177 : events->tailfree = chunk->freeptr;
4854 : : }
4855 : : }
4426 noah@leadboat.com 4856 [ + + ]: 4574 : if (slot1 != NULL)
4857 : : {
4858 : 19 : ExecDropSingleTupleTableSlot(slot1);
4859 : 19 : ExecDropSingleTupleTableSlot(slot2);
4860 : : }
4861 : :
4862 : : /* Release working resources */
6838 tgl@sss.pgh.pa.us 4863 : 4574 : MemoryContextDelete(per_tuple_context);
4864 : :
4865 [ + + ]: 4574 : if (local_estate)
4866 : : {
2030 heikki.linnakangas@i 4867 : 179 : ExecCloseResultRelations(estate);
2625 andres@anarazel.de 4868 : 179 : ExecResetTupleTable(estate->es_tupleTable, false);
6838 tgl@sss.pgh.pa.us 4869 : 179 : FreeExecutorState(estate);
4870 : : }
4871 : :
6402 4872 : 4574 : return all_fired;
4873 : : }
4874 : :
4875 : :
4876 : : /*
4877 : : * GetAfterTriggersTableData
4878 : : *
4879 : : * Find or create an AfterTriggersTableData struct for the specified
4880 : : * trigger event (relation + operation type). Ignore existing structs
4881 : : * marked "closed"; we don't want to put any additional tuples into them,
4882 : : * nor change their stmt-triggers-fired state.
4883 : : *
4884 : : * Note: the AfterTriggersTableData list is allocated in the current
4885 : : * (sub)transaction's CurTransactionContext. This is OK because
4886 : : * we don't need it to live past AfterTriggerEndQuery.
4887 : : */
4888 : : static AfterTriggersTableData *
3153 4889 : 1633 : GetAfterTriggersTableData(Oid relid, CmdType cmdType)
4890 : : {
4891 : : AfterTriggersTableData *table;
4892 : : AfterTriggersQueryData *qs;
4893 : : MemoryContext oldcxt;
4894 : : ListCell *lc;
4895 : :
4896 : : /* At this level, cmdType should not be, eg, CMD_MERGE */
101 dean.a.rasheed@gmail 4897 [ + + + + : 1633 : Assert(cmdType == CMD_INSERT ||
- + ]
4898 : : cmdType == CMD_UPDATE ||
4899 : : cmdType == CMD_DELETE);
4900 : :
4901 : : /* Caller should have ensured query_depth is OK. */
3153 tgl@sss.pgh.pa.us 4902 [ + - - + ]: 1633 : Assert(afterTriggers.query_depth >= 0 &&
4903 : : afterTriggers.query_depth < afterTriggers.maxquerydepth);
4904 : 1633 : qs = &afterTriggers.query_stack[afterTriggers.query_depth];
4905 : :
4906 [ + + + + : 1873 : foreach(lc, qs->tables)
+ + ]
4907 : : {
4908 : 1084 : table = (AfterTriggersTableData *) lfirst(lc);
4909 [ + + + + ]: 1084 : if (table->relid == relid && table->cmdType == cmdType &&
4910 [ + + ]: 868 : !table->closed)
4911 : 844 : return table;
4912 : : }
4913 : :
4914 : 789 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
4915 : :
146 michael@paquier.xyz 4916 :GNC 789 : table = palloc0_object(AfterTriggersTableData);
3153 tgl@sss.pgh.pa.us 4917 :CBC 789 : table->relid = relid;
4918 : 789 : table->cmdType = cmdType;
4919 : 789 : qs->tables = lappend(qs->tables, table);
4920 : :
4921 : 789 : MemoryContextSwitchTo(oldcxt);
4922 : :
4923 : 789 : return table;
4924 : : }
4925 : :
4926 : : /*
4927 : : * Returns a TupleTableSlot suitable for holding the tuples to be put
4928 : : * into AfterTriggersTableData's transition table tuplestores.
4929 : : */
4930 : : static TupleTableSlot *
1893 alvherre@alvh.no-ip. 4931 : 196 : GetAfterTriggersStoreSlot(AfterTriggersTableData *table,
4932 : : TupleDesc tupdesc)
4933 : : {
4934 : : /* Create it if not already done. */
4935 [ + + ]: 196 : if (!table->storeslot)
4936 : : {
4937 : : MemoryContext oldcxt;
4938 : :
4939 : : /*
4940 : : * We need this slot only until AfterTriggerEndQuery, but making it
4941 : : * last till end-of-subxact is good enough. It'll be freed by
4942 : : * AfterTriggerFreeQuery(). However, the passed-in tupdesc might have
4943 : : * a different lifespan, so we'd better make a copy of that.
4944 : : */
4945 : 56 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
1318 tgl@sss.pgh.pa.us 4946 : 56 : tupdesc = CreateTupleDescCopy(tupdesc);
1893 alvherre@alvh.no-ip. 4947 : 56 : table->storeslot = MakeSingleTupleTableSlot(tupdesc, &TTSOpsVirtual);
4948 : 56 : MemoryContextSwitchTo(oldcxt);
4949 : : }
4950 : :
4951 : 196 : return table->storeslot;
4952 : : }
4953 : :
4954 : : /*
4955 : : * MakeTransitionCaptureState
4956 : : *
4957 : : * Make a TransitionCaptureState object for the given TriggerDesc, target
4958 : : * relation, and operation type. The TCS object holds all the state needed
4959 : : * to decide whether to capture tuples in transition tables.
4960 : : *
4961 : : * If there are no triggers in 'trigdesc' that request relevant transition
4962 : : * tables, then return NULL.
4963 : : *
4964 : : * The resulting object can be passed to the ExecAR* functions. When
4965 : : * dealing with child tables, the caller can set tcs_original_insert_tuple
4966 : : * to avoid having to reconstruct the original tuple in the root table's
4967 : : * format.
4968 : : *
4969 : : * Note that we copy the flags from a parent table into this struct (rather
4970 : : * than subsequently using the relation's TriggerDesc directly) so that we can
4971 : : * use it to control collection of transition tuples from child tables.
4972 : : *
4973 : : * Per SQL spec, all operations of the same kind (INSERT/UPDATE/DELETE)
4974 : : * on the same table during one query should share one transition table.
4975 : : * Therefore, the Tuplestores are owned by an AfterTriggersTableData struct
4976 : : * looked up using the table OID + CmdType, and are merely referenced by
4977 : : * the TransitionCaptureState objects we hand out to callers.
4978 : : */
4979 : : TransitionCaptureState *
3153 tgl@sss.pgh.pa.us 4980 : 74834 : MakeTransitionCaptureState(TriggerDesc *trigdesc, Oid relid, CmdType cmdType)
4981 : : {
4982 : : TransitionCaptureState *state;
4983 : : bool need_old_upd,
4984 : : need_new_upd,
4985 : : need_old_del,
4986 : : need_new_ins;
4987 : : AfterTriggersTableData *ins_table;
4988 : : AfterTriggersTableData *upd_table;
4989 : : AfterTriggersTableData *del_table;
4990 : : MemoryContext oldcxt;
4991 : : ResourceOwner saveResourceOwner;
4992 : :
4993 [ + + ]: 74834 : if (trigdesc == NULL)
4994 : 66099 : return NULL;
4995 : :
4996 : : /* Detect which table(s) we need. */
4997 [ + + + + : 8735 : switch (cmdType)
- ]
4998 : : {
4999 : 4913 : case CMD_INSERT:
1499 alvherre@alvh.no-ip. 5000 : 4913 : need_old_upd = need_old_del = need_new_upd = false;
5001 : 4913 : need_new_ins = trigdesc->trig_insert_new_table;
3153 tgl@sss.pgh.pa.us 5002 : 4913 : break;
5003 : 2581 : case CMD_UPDATE:
1499 alvherre@alvh.no-ip. 5004 : 2581 : need_old_upd = trigdesc->trig_update_old_table;
5005 : 2581 : need_new_upd = trigdesc->trig_update_new_table;
5006 : 2581 : need_old_del = need_new_ins = false;
3153 tgl@sss.pgh.pa.us 5007 : 2581 : break;
5008 : 995 : case CMD_DELETE:
1499 alvherre@alvh.no-ip. 5009 : 995 : need_old_del = trigdesc->trig_delete_old_table;
5010 : 995 : need_old_upd = need_new_upd = need_new_ins = false;
5011 : 995 : break;
5012 : 246 : case CMD_MERGE:
5013 : 246 : need_old_upd = trigdesc->trig_update_old_table;
5014 : 246 : need_new_upd = trigdesc->trig_update_new_table;
5015 : 246 : need_old_del = trigdesc->trig_delete_old_table;
5016 : 246 : need_new_ins = trigdesc->trig_insert_new_table;
3153 tgl@sss.pgh.pa.us 5017 : 246 : break;
3153 tgl@sss.pgh.pa.us 5018 :UBC 0 : default:
5019 [ # # ]: 0 : elog(ERROR, "unexpected CmdType: %d", (int) cmdType);
5020 : : /* keep compiler quiet */
5021 : : need_old_upd = need_new_upd = need_old_del = need_new_ins = false;
5022 : : break;
5023 : : }
1499 alvherre@alvh.no-ip. 5024 [ + + + + :CBC 8735 : if (!need_old_upd && !need_new_upd && !need_new_ins && !need_old_del)
+ + + + ]
3153 tgl@sss.pgh.pa.us 5025 : 8304 : return NULL;
5026 : :
5027 : : /* Check state, like AfterTriggerSaveEvent. */
5028 [ - + ]: 431 : if (afterTriggers.query_depth < 0)
3153 tgl@sss.pgh.pa.us 5029 [ # # ]:UBC 0 : elog(ERROR, "MakeTransitionCaptureState() called outside of query");
5030 : :
5031 : : /* Be sure we have enough space to record events at this query depth. */
3153 tgl@sss.pgh.pa.us 5032 [ + + ]:CBC 431 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
5033 : 315 : AfterTriggerEnlargeQueryState();
5034 : :
5035 : : /*
5036 : : * Find or create AfterTriggersTableData struct(s) to hold the
5037 : : * tuplestore(s). If there's a matching struct but it's marked closed,
5038 : : * ignore it; we need a newer one.
5039 : : *
5040 : : * Note: MERGE must use the same AfterTriggersTableData structs as INSERT,
5041 : : * UPDATE, and DELETE, so that any MERGE'd tuples are added to the same
5042 : : * tuplestores as tuples from any INSERT, UPDATE, or DELETE commands
5043 : : * running in the same top-level command (e.g., in a writable CTE).
5044 : : *
5045 : : * Note: the AfterTriggersTableData list, as well as the tuplestores, are
5046 : : * allocated in the current (sub)transaction's CurTransactionContext, and
5047 : : * the tuplestores are managed by the (sub)transaction's resource owner.
5048 : : * This is sufficient lifespan because we do not allow triggers using
5049 : : * transition tables to be deferrable; they will be fired during
5050 : : * AfterTriggerEndQuery, after which it's okay to delete the data.
5051 : : */
101 dean.a.rasheed@gmail 5052 [ + + ]: 431 : if (need_new_ins)
5053 : 180 : ins_table = GetAfterTriggersTableData(relid, CMD_INSERT);
5054 : : else
5055 : 251 : ins_table = NULL;
5056 : :
5057 [ + + + + ]: 431 : if (need_old_upd || need_new_upd)
5058 : 145 : upd_table = GetAfterTriggersTableData(relid, CMD_UPDATE);
5059 : : else
5060 : 286 : upd_table = NULL;
5061 : :
5062 [ + + ]: 431 : if (need_old_del)
5063 : 130 : del_table = GetAfterTriggersTableData(relid, CMD_DELETE);
5064 : : else
5065 : 301 : del_table = NULL;
5066 : :
5067 : : /* Now create required tuplestore(s), if we don't have them already. */
3153 tgl@sss.pgh.pa.us 5068 : 431 : oldcxt = MemoryContextSwitchTo(CurTransactionContext);
5069 : 431 : saveResourceOwner = CurrentResourceOwner;
3128 5070 : 431 : CurrentResourceOwner = CurTransactionResourceOwner;
5071 : :
101 dean.a.rasheed@gmail 5072 [ + + + + ]: 431 : if (need_old_upd && upd_table->old_tuplestore == NULL)
5073 : 121 : upd_table->old_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5074 [ + + + + ]: 431 : if (need_new_upd && upd_table->new_tuplestore == NULL)
5075 : 129 : upd_table->new_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5076 [ + + + + ]: 431 : if (need_old_del && del_table->old_tuplestore == NULL)
5077 : 98 : del_table->old_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5078 [ + + + + ]: 431 : if (need_new_ins && ins_table->new_tuplestore == NULL)
5079 : 172 : ins_table->new_tuplestore = tuplestore_begin_heap(false, false, work_mem);
5080 : :
3153 tgl@sss.pgh.pa.us 5081 : 431 : CurrentResourceOwner = saveResourceOwner;
5082 : 431 : MemoryContextSwitchTo(oldcxt);
5083 : :
5084 : : /* Now build the TransitionCaptureState struct, in caller's context */
146 michael@paquier.xyz 5085 :GNC 431 : state = palloc0_object(TransitionCaptureState);
446 michael@paquier.xyz 5086 :CBC 431 : state->tcs_delete_old_table = need_old_del;
5087 : 431 : state->tcs_update_old_table = need_old_upd;
5088 : 431 : state->tcs_update_new_table = need_new_upd;
5089 : 431 : state->tcs_insert_new_table = need_new_ins;
101 dean.a.rasheed@gmail 5090 : 431 : state->tcs_insert_private = ins_table;
5091 : 431 : state->tcs_update_private = upd_table;
5092 : 431 : state->tcs_delete_private = del_table;
5093 : :
3153 tgl@sss.pgh.pa.us 5094 : 431 : return state;
5095 : : }
5096 : :
5097 : :
5098 : : /* ----------
5099 : : * AfterTriggerBeginXact()
5100 : : *
5101 : : * Called at transaction start (either BEGIN or implicit for single
5102 : : * statement outside of transaction block).
5103 : : * ----------
5104 : : */
5105 : : void
7907 5106 : 421968 : AfterTriggerBeginXact(void)
5107 : : {
5108 : : /*
5109 : : * Initialize after-trigger state structure to empty
5110 : : */
3240 5111 : 421968 : afterTriggers.firing_counter = (CommandId) 1; /* mustn't be 0 */
4212 rhaas@postgresql.org 5112 : 421968 : afterTriggers.query_depth = -1;
25 amitlan@postgresql.o 5113 :GNC 421968 : afterTriggers.firing_depth = 0;
32 5114 : 421968 : afterTriggers.batch_callbacks = NIL;
25 5115 : 421968 : afterTriggers.firing_batch_callbacks = false;
5116 : :
5117 : : /*
5118 : : * Verify that there is no leftover state remaining. If these assertions
5119 : : * trip, it means that AfterTriggerEndXact wasn't called or didn't clean
5120 : : * up properly.
5121 : : */
4212 rhaas@postgresql.org 5122 [ - + ]:CBC 421968 : Assert(afterTriggers.state == NULL);
5123 [ - + ]: 421968 : Assert(afterTriggers.query_stack == NULL);
5124 [ - + ]: 421968 : Assert(afterTriggers.maxquerydepth == 0);
5125 [ - + ]: 421968 : Assert(afterTriggers.event_cxt == NULL);
5126 [ - + ]: 421968 : Assert(afterTriggers.events.head == NULL);
3153 tgl@sss.pgh.pa.us 5127 [ - + ]: 421968 : Assert(afterTriggers.trans_stack == NULL);
4212 rhaas@postgresql.org 5128 [ - + ]: 421968 : Assert(afterTriggers.maxtransdepth == 0);
7907 tgl@sss.pgh.pa.us 5129 : 421968 : }
5130 : :
5131 : :
5132 : : /* ----------
5133 : : * AfterTriggerBeginQuery()
5134 : : *
5135 : : * Called just before we start processing a single query within a
5136 : : * transaction (or subtransaction). Most of the real work gets deferred
5137 : : * until somebody actually tries to queue a trigger event.
5138 : : * ----------
5139 : : */
5140 : : void
5141 : 243749 : AfterTriggerBeginQuery(void)
5142 : : {
5143 : : /* Increase the query stack depth */
4212 rhaas@postgresql.org 5144 : 243749 : afterTriggers.query_depth++;
9715 JanWieck@Yahoo.com 5145 : 243749 : }
5146 : :
5147 : :
5148 : : /* ----------
5149 : : * AfterTriggerEndQuery()
5150 : : *
5151 : : * Called after one query has been completely processed. At this time
5152 : : * we invoke all AFTER IMMEDIATE trigger events queued by the query, and
5153 : : * transfer deferred trigger events to the global deferred-trigger list.
5154 : : *
5155 : : * Note that this must be called BEFORE closing down the executor
5156 : : * with ExecutorEnd, because we make use of the EState's info about
5157 : : * target relations. Normally it is called from ExecutorFinish.
5158 : : * ----------
5159 : : */
5160 : : void
7711 tgl@sss.pgh.pa.us 5161 : 240552 : AfterTriggerEndQuery(EState *estate)
5162 : : {
5163 : : AfterTriggersQueryData *qs;
5164 : :
5165 : : /* Must be inside a query, too */
4212 rhaas@postgresql.org 5166 [ - + ]: 240552 : Assert(afterTriggers.query_depth >= 0);
5167 : :
5168 : : /*
5169 : : * If we never even got as far as initializing the event stack, there
5170 : : * certainly won't be any events, so exit quickly.
5171 : : */
5172 [ + + ]: 240552 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
5173 : : {
5174 : 234567 : afterTriggers.query_depth--;
5175 : 234567 : return;
5176 : : }
5177 : :
5178 : : /*
5179 : : * Process all immediate-mode triggers queued by the query, and move the
5180 : : * deferred ones to the main list of deferred events.
5181 : : *
5182 : : * Notice that we decide which ones will be fired, and put the deferred
5183 : : * ones on the main list, before anything is actually fired. This ensures
5184 : : * reasonably sane behavior if a trigger function does SET CONSTRAINTS ...
5185 : : * IMMEDIATE: all events we have decided to defer will be available for it
5186 : : * to fire.
5187 : : *
5188 : : * We loop in case a trigger queues more events at the same query level.
5189 : : * Ordinary trigger functions, including all PL/pgSQL trigger functions,
5190 : : * will instead fire any triggers in a dedicated query level. Foreign key
5191 : : * enforcement triggers do add to the current query level, thanks to their
5192 : : * passing fire_triggers = false to SPI_execute_snapshot(). Other
5193 : : * C-language triggers might do likewise.
5194 : : *
5195 : : * If we find no firable events, we don't have to increment
5196 : : * firing_counter.
5197 : : */
3152 tgl@sss.pgh.pa.us 5198 : 5985 : qs = &afterTriggers.query_stack[afterTriggers.query_depth];
5199 : :
25 amitlan@postgresql.o 5200 :GNC 5985 : afterTriggers.firing_depth++;
5201 : : for (;;)
5202 : : {
3153 tgl@sss.pgh.pa.us 5203 [ + + ]:CBC 6193 : if (afterTriggerMarkEvents(&qs->events, &afterTriggers.events, true))
5204 : : {
4212 rhaas@postgresql.org 5205 : 4936 : CommandId firing_id = afterTriggers.firing_counter++;
3152 tgl@sss.pgh.pa.us 5206 : 4936 : AfterTriggerEventChunk *oldtail = qs->events.tail;
5207 : :
5208 [ + + ]: 4936 : if (afterTriggerInvokeEvents(&qs->events, firing_id, estate, false))
6402 5209 : 4187 : break; /* all fired */
5210 : :
5211 : : /*
5212 : : * Firing a trigger could result in query_stack being repalloc'd,
5213 : : * so we must recalculate qs after each afterTriggerInvokeEvents
5214 : : * call. Furthermore, it's unsafe to pass delete_ok = true here,
5215 : : * because that could cause afterTriggerInvokeEvents to try to
5216 : : * access qs->events after the stack has been repalloc'd.
5217 : : */
3152 5218 : 208 : qs = &afterTriggers.query_stack[afterTriggers.query_depth];
5219 : :
5220 : : /*
5221 : : * We'll need to scan the events list again. To reduce the cost
5222 : : * of doing so, get rid of completely-fired chunks. We know that
5223 : : * all events were marked IN_PROGRESS or DONE at the conclusion of
5224 : : * afterTriggerMarkEvents, so any still-interesting events must
5225 : : * have been added after that, and so must be in the chunk that
5226 : : * was then the tail chunk, or in later chunks. So, zap all
5227 : : * chunks before oldtail. This is approximately the same set of
5228 : : * events we would have gotten rid of by passing delete_ok = true.
5229 : : */
5230 [ - + ]: 208 : Assert(oldtail != NULL);
5231 [ - + ]: 208 : while (qs->events.head != oldtail)
3152 tgl@sss.pgh.pa.us 5232 :UBC 0 : afterTriggerDeleteHeadEventChunk(qs);
5233 : : }
5234 : : else
6402 tgl@sss.pgh.pa.us 5235 :CBC 1249 : break;
5236 : : }
5237 : :
5238 : : /*
5239 : : * Fire batch callbacks before releasing query-level storage and before
5240 : : * decrementing query_depth. Callbacks may do real work (index probes,
5241 : : * error reporting).
5242 : : */
25 amitlan@postgresql.o 5243 :GNC 5436 : FireAfterTriggerBatchCallbacks(qs->batch_callbacks);
5244 : :
5245 : : /* Release query-level-local storage, including tuplestores if any */
3153 tgl@sss.pgh.pa.us 5246 :CBC 5188 : AfterTriggerFreeQuery(&afterTriggers.query_stack[afterTriggers.query_depth]);
5247 : :
5248 : 5188 : afterTriggers.query_depth--;
25 amitlan@postgresql.o 5249 :GNC 5188 : afterTriggers.firing_depth--;
5250 : : }
5251 : :
5252 : :
5253 : : /*
5254 : : * AfterTriggerFreeQuery
5255 : : * Release subsidiary storage for a trigger query level.
5256 : : * This includes closing down tuplestores.
5257 : : * Note: it's important for this to be safe if interrupted by an error
5258 : : * and then called again for the same query level.
5259 : : */
5260 : : static void
3153 tgl@sss.pgh.pa.us 5261 :CBC 5205 : AfterTriggerFreeQuery(AfterTriggersQueryData *qs)
5262 : : {
5263 : : Tuplestorestate *ts;
5264 : : List *tables;
5265 : : ListCell *lc;
5266 : :
5267 : : /* Drop the trigger events */
5268 : 5205 : afterTriggerFreeEventList(&qs->events);
5269 : :
5270 : : /* Drop FDW tuplestore if any */
5271 : 5205 : ts = qs->fdw_tuplestore;
5272 : 5205 : qs->fdw_tuplestore = NULL;
5273 [ + + ]: 5205 : if (ts)
5274 : 18 : tuplestore_end(ts);
5275 : :
5276 : : /* Release per-table subsidiary storage */
5277 : 5205 : tables = qs->tables;
5278 [ + + + + : 5948 : foreach(lc, tables)
+ + ]
5279 : : {
5280 : 743 : AfterTriggersTableData *table = (AfterTriggersTableData *) lfirst(lc);
5281 : :
101 dean.a.rasheed@gmail 5282 : 743 : ts = table->old_tuplestore;
5283 : 743 : table->old_tuplestore = NULL;
3153 tgl@sss.pgh.pa.us 5284 [ + + ]: 743 : if (ts)
5285 : 199 : tuplestore_end(ts);
101 dean.a.rasheed@gmail 5286 : 743 : ts = table->new_tuplestore;
5287 : 743 : table->new_tuplestore = NULL;
3153 tgl@sss.pgh.pa.us 5288 [ + + ]: 743 : if (ts)
5289 : 275 : tuplestore_end(ts);
1893 alvherre@alvh.no-ip. 5290 [ + + ]: 743 : if (table->storeslot)
5291 : : {
1318 tgl@sss.pgh.pa.us 5292 : 56 : TupleTableSlot *slot = table->storeslot;
5293 : :
5294 : 56 : table->storeslot = NULL;
5295 : 56 : ExecDropSingleTupleTableSlot(slot);
5296 : : }
5297 : : }
5298 : :
5299 : : /*
5300 : : * Now free the AfterTriggersTableData structs and list cells. Reset list
5301 : : * pointer first; if list_free_deep somehow gets an error, better to leak
5302 : : * that storage than have an infinite loop.
5303 : : */
3153 5304 : 5205 : qs->tables = NIL;
5305 : 5205 : list_free_deep(tables);
5306 : :
25 amitlan@postgresql.o 5307 :GNC 5205 : list_free_deep(qs->batch_callbacks);
5308 : 5205 : qs->batch_callbacks = NIL;
9715 JanWieck@Yahoo.com 5309 :CBC 5205 : }
5310 : :
5311 : :
5312 : : /* ----------
5313 : : * AfterTriggerFireDeferred()
5314 : : *
5315 : : * Called just before the current transaction is committed. At this
5316 : : * time we invoke all pending DEFERRED triggers.
5317 : : *
5318 : : * It is possible for other modules to queue additional deferred triggers
5319 : : * during pre-commit processing; therefore xact.c may have to call this
5320 : : * multiple times.
5321 : : * ----------
5322 : : */
5323 : : void
7694 tgl@sss.pgh.pa.us 5324 : 392386 : AfterTriggerFireDeferred(void)
5325 : : {
5326 : : AfterTriggerEventList *events;
6567 alvherre@alvh.no-ip. 5327 : 392386 : bool snap_pushed = false;
5328 : :
5329 : : /* Must not be inside a query */
4212 rhaas@postgresql.org 5330 [ - + ]: 392386 : Assert(afterTriggers.query_depth == -1);
5331 : :
5332 : : /*
5333 : : * If there are any triggers to fire, make sure we have set a snapshot for
5334 : : * them to use. (Since PortalRunUtility doesn't set a snap for COMMIT, we
5335 : : * can't assume ActiveSnapshot is valid on entry.)
5336 : : */
5337 : 392386 : events = &afterTriggers.events;
7694 tgl@sss.pgh.pa.us 5338 [ + + ]: 392386 : if (events->head != NULL)
5339 : : {
6567 alvherre@alvh.no-ip. 5340 : 238 : PushActiveSnapshot(GetTransactionSnapshot());
5341 : 238 : snap_pushed = true;
5342 : : }
5343 : :
5344 : : /*
5345 : : * Run all the remaining triggers. Loop until they are all gone, in case
5346 : : * some trigger queues more for us to do.
5347 : : */
25 amitlan@postgresql.o 5348 :GNC 392386 : afterTriggers.firing_depth++;
7907 tgl@sss.pgh.pa.us 5349 [ + + ]:CBC 392386 : while (afterTriggerMarkEvents(events, NULL, false))
5350 : : {
4212 rhaas@postgresql.org 5351 : 238 : CommandId firing_id = afterTriggers.firing_counter++;
5352 : :
6402 tgl@sss.pgh.pa.us 5353 [ + - ]: 238 : if (afterTriggerInvokeEvents(events, firing_id, NULL, true))
5354 : 169 : break; /* all fired */
5355 : : }
5356 : :
5357 : : /* Flush any fast-path batches accumulated by the triggers just fired. */
25 amitlan@postgresql.o 5358 :GNC 392317 : FireAfterTriggerBatchCallbacks(afterTriggers.batch_callbacks);
5359 : :
5360 : 392278 : afterTriggers.firing_depth--;
5361 : :
5362 : : /*
5363 : : * We don't bother freeing the event list or batch_callbacks, since they
5364 : : * will go away anyway (and more efficiently than via pfree) in
5365 : : * AfterTriggerEndXact.
5366 : : */
5367 : :
6567 alvherre@alvh.no-ip. 5368 [ + + ]:CBC 392278 : if (snap_pushed)
5369 : 130 : PopActiveSnapshot();
9715 JanWieck@Yahoo.com 5370 : 392278 : }
5371 : :
5372 : :
5373 : : /* ----------
5374 : : * AfterTriggerEndXact()
5375 : : *
5376 : : * The current transaction is finishing.
5377 : : *
5378 : : * Any unfired triggers are canceled so we simply throw
5379 : : * away anything we know.
5380 : : *
5381 : : * Note: it is possible for this to be called repeatedly in case of
5382 : : * error during transaction abort; therefore, do not complain if
5383 : : * already closed down.
5384 : : * ----------
5385 : : */
5386 : : void
7694 tgl@sss.pgh.pa.us 5387 : 422192 : AfterTriggerEndXact(bool isCommit)
5388 : : {
5389 : : /*
5390 : : * Forget the pending-events list.
5391 : : *
5392 : : * Since all the info is in TopTransactionContext or children thereof, we
5393 : : * don't really need to do anything to reclaim memory. However, the
5394 : : * pending-events list could be large, and so it's useful to discard it as
5395 : : * soon as possible --- especially if we are aborting because we ran out
5396 : : * of memory for the list!
5397 : : */
4212 rhaas@postgresql.org 5398 [ + + ]: 422192 : if (afterTriggers.event_cxt)
5399 : : {
5400 : 4475 : MemoryContextDelete(afterTriggers.event_cxt);
5401 : 4475 : afterTriggers.event_cxt = NULL;
5402 : 4475 : afterTriggers.events.head = NULL;
5403 : 4475 : afterTriggers.events.tail = NULL;
5404 : 4475 : afterTriggers.events.tailfree = NULL;
5405 : : }
5406 : :
5407 : : /*
5408 : : * Forget any subtransaction state as well. Since this can't be very
5409 : : * large, we let the eventual reset of TopTransactionContext free the
5410 : : * memory instead of doing it here.
5411 : : */
3153 tgl@sss.pgh.pa.us 5412 : 422192 : afterTriggers.trans_stack = NULL;
4212 rhaas@postgresql.org 5413 : 422192 : afterTriggers.maxtransdepth = 0;
5414 : :
5415 : :
5416 : : /*
5417 : : * Forget the query stack and constraint-related state information. As
5418 : : * with the subtransaction state information, we don't bother freeing the
5419 : : * memory here.
5420 : : */
5421 : 422192 : afterTriggers.query_stack = NULL;
5422 : 422192 : afterTriggers.maxquerydepth = 0;
5423 : 422192 : afterTriggers.state = NULL;
5424 : :
5425 : : /* No more afterTriggers manipulation until next transaction starts. */
5426 : 422192 : afterTriggers.query_depth = -1;
5427 : :
25 amitlan@postgresql.o 5428 :GNC 422192 : afterTriggers.firing_depth = 0;
5429 : :
5430 : 422192 : list_free_deep(afterTriggers.batch_callbacks);
5431 : 422192 : afterTriggers.batch_callbacks = NIL;
5432 : 422192 : afterTriggers.firing_batch_callbacks = false;
9715 JanWieck@Yahoo.com 5433 :CBC 422192 : }
5434 : :
5435 : : /*
5436 : : * AfterTriggerBeginSubXact()
5437 : : *
5438 : : * Start a subtransaction.
5439 : : */
5440 : : void
7907 tgl@sss.pgh.pa.us 5441 : 12646 : AfterTriggerBeginSubXact(void)
5442 : : {
7911 5443 : 12646 : int my_level = GetCurrentTransactionNestLevel();
5444 : :
5445 : : /*
5446 : : * Allocate more space in the trans_stack if needed. (Note: because the
5447 : : * minimum nest level of a subtransaction is 2, we waste the first couple
5448 : : * entries of the array; not worth the notational effort to avoid it.)
5449 : : */
4212 rhaas@postgresql.org 5450 [ + + ]: 14176 : while (my_level >= afterTriggers.maxtransdepth)
5451 : : {
5452 [ + + ]: 1530 : if (afterTriggers.maxtransdepth == 0)
5453 : : {
5454 : : /* Arbitrarily initialize for max of 8 subtransaction levels */
3153 tgl@sss.pgh.pa.us 5455 : 1486 : afterTriggers.trans_stack = (AfterTriggersTransData *)
5456 : 1486 : MemoryContextAlloc(TopTransactionContext,
5457 : : 8 * sizeof(AfterTriggersTransData));
5458 : 1486 : afterTriggers.maxtransdepth = 8;
5459 : : }
5460 : : else
5461 : : {
5462 : : /* repalloc will keep the stack in the same context */
4212 rhaas@postgresql.org 5463 : 44 : int new_alloc = afterTriggers.maxtransdepth * 2;
5464 : :
3153 tgl@sss.pgh.pa.us 5465 : 44 : afterTriggers.trans_stack = (AfterTriggersTransData *)
5466 : 44 : repalloc(afterTriggers.trans_stack,
5467 : : new_alloc * sizeof(AfterTriggersTransData));
4212 rhaas@postgresql.org 5468 : 44 : afterTriggers.maxtransdepth = new_alloc;
5469 : : }
5470 : : }
5471 : :
5472 : : /*
5473 : : * Push the current information into the stack. The SET CONSTRAINTS state
5474 : : * is not saved until/unless changed. Likewise, we don't make a
5475 : : * per-subtransaction event context until needed.
5476 : : */
3153 tgl@sss.pgh.pa.us 5477 : 12646 : afterTriggers.trans_stack[my_level].state = NULL;
5478 : 12646 : afterTriggers.trans_stack[my_level].events = afterTriggers.events;
5479 : 12646 : afterTriggers.trans_stack[my_level].query_depth = afterTriggers.query_depth;
5480 : 12646 : afterTriggers.trans_stack[my_level].firing_counter = afterTriggers.firing_counter;
7978 5481 : 12646 : }
5482 : :
5483 : : /*
5484 : : * AfterTriggerEndSubXact()
5485 : : *
5486 : : * The current subtransaction is ending.
5487 : : */
5488 : : void
7907 5489 : 12646 : AfterTriggerEndSubXact(bool isCommit)
5490 : : {
7911 5491 : 12646 : int my_level = GetCurrentTransactionNestLevel();
5492 : : SetConstraintState state;
5493 : : AfterTriggerEvent event;
5494 : : AfterTriggerEventChunk *chunk;
5495 : : CommandId subxact_firing_id;
5496 : :
5497 : : /*
5498 : : * Pop the prior state if needed.
5499 : : */
7978 5500 [ + + ]: 12646 : if (isCommit)
5501 : : {
4212 rhaas@postgresql.org 5502 [ - + ]: 7266 : Assert(my_level < afterTriggers.maxtransdepth);
5503 : : /* If we saved a prior state, we don't need it anymore */
3153 tgl@sss.pgh.pa.us 5504 : 7266 : state = afterTriggers.trans_stack[my_level].state;
7978 5505 [ + + ]: 7266 : if (state != NULL)
5506 : 4 : pfree(state);
5507 : : /* this avoids double pfree if error later: */
3153 5508 : 7266 : afterTriggers.trans_stack[my_level].state = NULL;
4212 rhaas@postgresql.org 5509 [ - + ]: 7266 : Assert(afterTriggers.query_depth ==
5510 : : afterTriggers.trans_stack[my_level].query_depth);
5511 : : }
5512 : : else
5513 : : {
5514 : : /*
5515 : : * Aborting. It is possible subxact start failed before calling
5516 : : * AfterTriggerBeginSubXact, in which case we mustn't risk touching
5517 : : * trans_stack levels that aren't there.
5518 : : */
5519 [ - + ]: 5380 : if (my_level >= afterTriggers.maxtransdepth)
5945 tgl@sss.pgh.pa.us 5520 :UBC 0 : return;
5521 : :
5522 : : /*
5523 : : * Release query-level storage for queries being aborted, and restore
5524 : : * query_depth to its pre-subxact value. This assumes that a
5525 : : * subtransaction will not add events to query levels started in a
5526 : : * earlier transaction state.
5527 : : */
3153 tgl@sss.pgh.pa.us 5528 [ + + ]:CBC 5439 : while (afterTriggers.query_depth > afterTriggers.trans_stack[my_level].query_depth)
5529 : : {
4212 rhaas@postgresql.org 5530 [ + + ]: 59 : if (afterTriggers.query_depth < afterTriggers.maxquerydepth)
3153 tgl@sss.pgh.pa.us 5531 : 17 : AfterTriggerFreeQuery(&afterTriggers.query_stack[afterTriggers.query_depth]);
4212 rhaas@postgresql.org 5532 : 59 : afterTriggers.query_depth--;
5533 : : }
5534 [ - + ]: 5380 : Assert(afterTriggers.query_depth ==
5535 : : afterTriggers.trans_stack[my_level].query_depth);
5536 : :
5537 : : /*
5538 : : * Restore the global deferred-event list to its former length,
5539 : : * discarding any events queued by the subxact.
5540 : : */
5541 : 5380 : afterTriggerRestoreEventList(&afterTriggers.events,
3153 tgl@sss.pgh.pa.us 5542 : 5380 : &afterTriggers.trans_stack[my_level].events);
5543 : :
5544 : : /*
5545 : : * Restore the trigger state. If the saved state is NULL, then this
5546 : : * subxact didn't save it, so it doesn't need restoring.
5547 : : */
5548 : 5380 : state = afterTriggers.trans_stack[my_level].state;
7978 5549 [ + + ]: 5380 : if (state != NULL)
5550 : : {
4212 rhaas@postgresql.org 5551 : 2 : pfree(afterTriggers.state);
5552 : 2 : afterTriggers.state = state;
5553 : : }
5554 : : /* this avoids double pfree if error later: */
3153 tgl@sss.pgh.pa.us 5555 : 5380 : afterTriggers.trans_stack[my_level].state = NULL;
5556 : :
5557 : : /*
5558 : : * Scan for any remaining deferred events that were marked DONE or IN
5559 : : * PROGRESS by this subxact or a child, and un-mark them. We can
5560 : : * recognize such events because they have a firing ID greater than or
5561 : : * equal to the firing_counter value we saved at subtransaction start.
5562 : : * (This essentially assumes that the current subxact includes all
5563 : : * subxacts started after it.)
5564 : : */
5565 : 5380 : subxact_firing_id = afterTriggers.trans_stack[my_level].firing_counter;
4212 rhaas@postgresql.org 5566 [ + - + - : 5408 : for_each_event_chunk(event, chunk, afterTriggers.events)
+ - + + +
+ ]
5567 : : {
6402 tgl@sss.pgh.pa.us 5568 : 14 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
5569 : :
5570 [ + + ]: 14 : if (event->ate_flags &
5571 : : (AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS))
5572 : : {
5573 [ + - ]: 2 : if (evtshared->ats_firing_id >= subxact_firing_id)
5574 : 2 : event->ate_flags &=
5575 : : ~(AFTER_TRIGGER_DONE | AFTER_TRIGGER_IN_PROGRESS);
5576 : : }
5577 : : }
5578 : : }
5579 : :
5580 : : /* Reset in case a callback threw an error while firing. */
25 amitlan@postgresql.o 5581 :GNC 12646 : afterTriggers.firing_batch_callbacks = false;
5582 : : }
5583 : :
5584 : : /*
5585 : : * Get the transition table for the given event and depending on whether we are
5586 : : * processing the old or the new tuple.
5587 : : */
5588 : : static Tuplestorestate *
1516 alvherre@alvh.no-ip. 5589 :CBC 44126 : GetAfterTriggersTransitionTable(int event,
5590 : : TupleTableSlot *oldslot,
5591 : : TupleTableSlot *newslot,
5592 : : TransitionCaptureState *transition_capture)
5593 : : {
5594 : 44126 : Tuplestorestate *tuplestore = NULL;
5595 : 44126 : bool delete_old_table = transition_capture->tcs_delete_old_table;
5596 : 44126 : bool update_old_table = transition_capture->tcs_update_old_table;
5597 : 44126 : bool update_new_table = transition_capture->tcs_update_new_table;
5598 : 44126 : bool insert_new_table = transition_capture->tcs_insert_new_table;
5599 : :
5600 : : /*
5601 : : * For INSERT events NEW should be non-NULL, for DELETE events OLD should
5602 : : * be non-NULL, whereas for UPDATE events normally both OLD and NEW are
5603 : : * non-NULL. But for UPDATE events fired for capturing transition tuples
5604 : : * during UPDATE partition-key row movement, OLD is NULL when the event is
5605 : : * for a row being inserted, whereas NEW is NULL when the event is for a
5606 : : * row being deleted.
5607 : : */
5608 [ + + + - : 44126 : Assert(!(event == TRIGGER_EVENT_DELETE && delete_old_table &&
+ - - + ]
5609 : : TupIsNull(oldslot)));
5610 [ + + + - : 44126 : Assert(!(event == TRIGGER_EVENT_INSERT && insert_new_table &&
+ - - + ]
5611 : : TupIsNull(newslot)));
5612 : :
5613 [ + + + - ]: 44126 : if (!TupIsNull(oldslot))
5614 : : {
5615 [ - + - - ]: 3619 : Assert(TupIsNull(newslot));
5616 [ + + + - ]: 3619 : if (event == TRIGGER_EVENT_DELETE && delete_old_table)
101 dean.a.rasheed@gmail 5617 : 3372 : tuplestore = transition_capture->tcs_delete_private->old_tuplestore;
1516 alvherre@alvh.no-ip. 5618 [ + - + + ]: 247 : else if (event == TRIGGER_EVENT_UPDATE && update_old_table)
101 dean.a.rasheed@gmail 5619 : 231 : tuplestore = transition_capture->tcs_update_private->old_tuplestore;
5620 : : }
1516 alvherre@alvh.no-ip. 5621 [ + - + - ]: 40507 : else if (!TupIsNull(newslot))
5622 : : {
5623 [ - + - - ]: 40507 : Assert(TupIsNull(oldslot));
5624 [ + + + - ]: 40507 : if (event == TRIGGER_EVENT_INSERT && insert_new_table)
101 dean.a.rasheed@gmail 5625 : 40260 : tuplestore = transition_capture->tcs_insert_private->new_tuplestore;
1516 alvherre@alvh.no-ip. 5626 [ + - + + ]: 247 : else if (event == TRIGGER_EVENT_UPDATE && update_new_table)
101 dean.a.rasheed@gmail 5627 : 243 : tuplestore = transition_capture->tcs_update_private->new_tuplestore;
5628 : : }
5629 : :
1516 alvherre@alvh.no-ip. 5630 : 44126 : return tuplestore;
5631 : : }
5632 : :
5633 : : /*
5634 : : * Add the given heap tuple to the given tuplestore, applying the conversion
5635 : : * map if necessary.
5636 : : *
5637 : : * If original_insert_tuple is given, we can add that tuple without conversion.
5638 : : */
5639 : : static void
5640 : 44126 : TransitionTableAddTuple(EState *estate,
5641 : : int event,
5642 : : TransitionCaptureState *transition_capture,
5643 : : ResultRelInfo *relinfo,
5644 : : TupleTableSlot *slot,
5645 : : TupleTableSlot *original_insert_tuple,
5646 : : Tuplestorestate *tuplestore)
5647 : : {
5648 : : TupleConversionMap *map;
5649 : :
5650 : : /*
5651 : : * Nothing needs to be done if we don't have a tuplestore.
5652 : : */
5653 [ + + ]: 44126 : if (tuplestore == NULL)
5654 : 20 : return;
5655 : :
5656 [ + + ]: 44106 : if (original_insert_tuple)
5657 : 96 : tuplestore_puttupleslot(tuplestore, original_insert_tuple);
5658 [ + + ]: 44010 : else if ((map = ExecGetChildToRootMap(relinfo)) != NULL)
5659 : : {
5660 : : AfterTriggersTableData *table;
5661 : : TupleTableSlot *storeslot;
5662 : :
101 dean.a.rasheed@gmail 5663 [ + + + - ]: 196 : switch (event)
5664 : : {
5665 : 8 : case TRIGGER_EVENT_INSERT:
5666 : 8 : table = transition_capture->tcs_insert_private;
5667 : 8 : break;
5668 : 164 : case TRIGGER_EVENT_UPDATE:
5669 : 164 : table = transition_capture->tcs_update_private;
5670 : 164 : break;
5671 : 24 : case TRIGGER_EVENT_DELETE:
5672 : 24 : table = transition_capture->tcs_delete_private;
5673 : 24 : break;
101 dean.a.rasheed@gmail 5674 :UBC 0 : default:
5675 [ # # ]: 0 : elog(ERROR, "invalid after-trigger event code: %d", event);
5676 : : table = NULL; /* keep compiler quiet */
5677 : : break;
5678 : : }
5679 : :
1516 alvherre@alvh.no-ip. 5680 :CBC 196 : storeslot = GetAfterTriggersStoreSlot(table, map->outdesc);
5681 : 196 : execute_attr_map_slot(map->attrMap, slot, storeslot);
5682 : 196 : tuplestore_puttupleslot(tuplestore, storeslot);
5683 : : }
5684 : : else
5685 : 43814 : tuplestore_puttupleslot(tuplestore, slot);
5686 : : }
5687 : :
5688 : : /* ----------
5689 : : * AfterTriggerEnlargeQueryState()
5690 : : *
5691 : : * Prepare the necessary state so that we can record AFTER trigger events
5692 : : * queued by a query. It is allowed to have nested queries within a
5693 : : * (sub)transaction, so we need to have separate state for each query
5694 : : * nesting level.
5695 : : * ----------
5696 : : */
5697 : : static void
4212 rhaas@postgresql.org 5698 : 4708 : AfterTriggerEnlargeQueryState(void)
5699 : : {
4000 bruce@momjian.us 5700 : 4708 : int init_depth = afterTriggers.maxquerydepth;
5701 : :
4212 rhaas@postgresql.org 5702 [ - + ]: 4708 : Assert(afterTriggers.query_depth >= afterTriggers.maxquerydepth);
5703 : :
5704 [ + - ]: 4708 : if (afterTriggers.maxquerydepth == 0)
5705 : : {
4211 5706 : 4708 : int new_alloc = Max(afterTriggers.query_depth + 1, 8);
5707 : :
3153 tgl@sss.pgh.pa.us 5708 : 4708 : afterTriggers.query_stack = (AfterTriggersQueryData *)
4212 rhaas@postgresql.org 5709 : 4708 : MemoryContextAlloc(TopTransactionContext,
5710 : : new_alloc * sizeof(AfterTriggersQueryData));
5711 : 4708 : afterTriggers.maxquerydepth = new_alloc;
5712 : : }
5713 : : else
5714 : : {
5715 : : /* repalloc will keep the stack in the same context */
4212 rhaas@postgresql.org 5716 :UBC 0 : int old_alloc = afterTriggers.maxquerydepth;
4211 5717 : 0 : int new_alloc = Max(afterTriggers.query_depth + 1,
5718 : : old_alloc * 2);
5719 : :
3153 tgl@sss.pgh.pa.us 5720 : 0 : afterTriggers.query_stack = (AfterTriggersQueryData *)
4212 rhaas@postgresql.org 5721 : 0 : repalloc(afterTriggers.query_stack,
5722 : : new_alloc * sizeof(AfterTriggersQueryData));
5723 : 0 : afterTriggers.maxquerydepth = new_alloc;
5724 : : }
5725 : :
5726 : : /* Initialize new array entries to empty */
4212 rhaas@postgresql.org 5727 [ + + ]:CBC 42372 : while (init_depth < afterTriggers.maxquerydepth)
5728 : : {
3153 tgl@sss.pgh.pa.us 5729 : 37664 : AfterTriggersQueryData *qs = &afterTriggers.query_stack[init_depth];
5730 : :
5731 : 37664 : qs->events.head = NULL;
5732 : 37664 : qs->events.tail = NULL;
5733 : 37664 : qs->events.tailfree = NULL;
5734 : 37664 : qs->fdw_tuplestore = NULL;
5735 : 37664 : qs->tables = NIL;
25 amitlan@postgresql.o 5736 :GNC 37664 : qs->batch_callbacks = NIL;
5737 : :
4212 rhaas@postgresql.org 5738 :CBC 37664 : ++init_depth;
5739 : : }
5740 : 4708 : }
5741 : :
5742 : : /*
5743 : : * Create an empty SetConstraintState with room for numalloc trigstates
5744 : : */
5745 : : static SetConstraintState
7907 tgl@sss.pgh.pa.us 5746 : 67 : SetConstraintStateCreate(int numalloc)
5747 : : {
5748 : : SetConstraintState state;
5749 : :
5750 : : /* Behave sanely with numalloc == 0 */
7978 5751 [ + + ]: 67 : if (numalloc <= 0)
5752 : 6 : numalloc = 1;
5753 : :
5754 : : /*
5755 : : * We assume that zeroing will correctly initialize the state values.
5756 : : */
5757 : : state = (SetConstraintState)
5758 : 67 : MemoryContextAllocZero(TopTransactionContext,
5759 : : offsetof(SetConstraintStateData, trigstates) +
4000 bruce@momjian.us 5760 : 67 : numalloc * sizeof(SetConstraintTriggerData));
5761 : :
7978 tgl@sss.pgh.pa.us 5762 : 67 : state->numalloc = numalloc;
5763 : :
5764 : 67 : return state;
5765 : : }
5766 : :
5767 : : /*
5768 : : * Copy a SetConstraintState
5769 : : */
5770 : : static SetConstraintState
7907 5771 : 6 : SetConstraintStateCopy(SetConstraintState origstate)
5772 : : {
5773 : : SetConstraintState state;
5774 : :
5775 : 6 : state = SetConstraintStateCreate(origstate->numstates);
5776 : :
7978 5777 : 6 : state->all_isset = origstate->all_isset;
5778 : 6 : state->all_isdeferred = origstate->all_isdeferred;
5779 : 6 : state->numstates = origstate->numstates;
5780 : 6 : memcpy(state->trigstates, origstate->trigstates,
7907 5781 : 6 : origstate->numstates * sizeof(SetConstraintTriggerData));
5782 : :
7978 5783 : 6 : return state;
5784 : : }
5785 : :
5786 : : /*
5787 : : * Add a per-trigger item to a SetConstraintState. Returns possibly-changed
5788 : : * pointer to the state object (it will change if we have to repalloc).
5789 : : */
5790 : : static SetConstraintState
7907 5791 : 228 : SetConstraintStateAddItem(SetConstraintState state,
5792 : : Oid tgoid, bool tgisdeferred)
5793 : : {
7978 5794 [ + + ]: 228 : if (state->numstates >= state->numalloc)
5795 : : {
7919 bruce@momjian.us 5796 : 20 : int newalloc = state->numalloc * 2;
5797 : :
5798 : 20 : newalloc = Max(newalloc, 8); /* in case original has size 0 */
5799 : : state = (SetConstraintState)
7978 tgl@sss.pgh.pa.us 5800 : 20 : repalloc(state,
5801 : : offsetof(SetConstraintStateData, trigstates) +
4092 5802 : 20 : newalloc * sizeof(SetConstraintTriggerData));
7978 5803 : 20 : state->numalloc = newalloc;
5804 [ - + ]: 20 : Assert(state->numstates < state->numalloc);
5805 : : }
5806 : :
7907 5807 : 228 : state->trigstates[state->numstates].sct_tgoid = tgoid;
5808 : 228 : state->trigstates[state->numstates].sct_tgisdeferred = tgisdeferred;
7978 5809 : 228 : state->numstates++;
5810 : :
5811 : 228 : return state;
5812 : : }
5813 : :
5814 : : /* ----------
5815 : : * AfterTriggerSetState()
5816 : : *
5817 : : * Execute the SET CONSTRAINTS ... utility command.
5818 : : * ----------
5819 : : */
5820 : : void
7907 5821 : 71 : AfterTriggerSetState(ConstraintsSetStmt *stmt)
5822 : : {
7911 5823 : 71 : int my_level = GetCurrentTransactionNestLevel();
5824 : :
5825 : : /* If we haven't already done so, initialize our state. */
4212 rhaas@postgresql.org 5826 [ + + ]: 71 : if (afterTriggers.state == NULL)
5827 : 61 : afterTriggers.state = SetConstraintStateCreate(8);
5828 : :
5829 : : /*
5830 : : * If in a subtransaction, and we didn't save the current state already,
5831 : : * save it so it can be restored if the subtransaction aborts.
5832 : : */
7911 tgl@sss.pgh.pa.us 5833 [ + + ]: 71 : if (my_level > 1 &&
3153 5834 [ + - ]: 6 : afterTriggers.trans_stack[my_level].state == NULL)
5835 : : {
5836 : 6 : afterTriggers.trans_stack[my_level].state =
4212 rhaas@postgresql.org 5837 : 6 : SetConstraintStateCopy(afterTriggers.state);
5838 : : }
5839 : :
5840 : : /*
5841 : : * Handle SET CONSTRAINTS ALL ...
5842 : : */
9519 bruce@momjian.us 5843 [ + + ]: 71 : if (stmt->constraints == NIL)
5844 : : {
5845 : : /*
5846 : : * Forget any previous SET CONSTRAINTS commands in this transaction.
5847 : : */
4212 rhaas@postgresql.org 5848 : 39 : afterTriggers.state->numstates = 0;
5849 : :
5850 : : /*
5851 : : * Set the per-transaction ALL state to known.
5852 : : */
5853 : 39 : afterTriggers.state->all_isset = true;
5854 : 39 : afterTriggers.state->all_isdeferred = stmt->deferred;
5855 : : }
5856 : : else
5857 : : {
5858 : : Relation conrel;
5859 : : Relation tgrel;
5952 tgl@sss.pgh.pa.us 5860 : 32 : List *conoidlist = NIL;
5861 : 32 : List *tgoidlist = NIL;
5862 : : ListCell *lc;
5863 : :
5864 : : /*
5865 : : * Handle SET CONSTRAINTS constraint-name [, ...]
5866 : : *
5867 : : * First, identify all the named constraints and make a list of their
5868 : : * OIDs. Since, unlike the SQL spec, we allow multiple constraints of
5869 : : * the same name within a schema, the specifications are not
5870 : : * necessarily unique. Our strategy is to target all matching
5871 : : * constraints within the first search-path schema that has any
5872 : : * matches, but disregard matches in schemas beyond the first match.
5873 : : * (This is a bit odd but it's the historical behavior.)
5874 : : *
5875 : : * A constraint in a partitioned table may have corresponding
5876 : : * constraints in the partitions. Grab those too.
5877 : : */
2661 andres@anarazel.de 5878 : 32 : conrel = table_open(ConstraintRelationId, AccessShareLock);
5879 : :
5952 tgl@sss.pgh.pa.us 5880 [ + - + + : 64 : foreach(lc, stmt->constraints)
+ + ]
5881 : : {
5882 : 32 : RangeVar *constraint = lfirst(lc);
5883 : : bool found;
5884 : : List *namespacelist;
5885 : : ListCell *nslc;
5886 : :
7313 bruce@momjian.us 5887 [ - + ]: 32 : if (constraint->catalogname)
5888 : : {
7313 bruce@momjian.us 5889 [ # # ]:UBC 0 : if (strcmp(constraint->catalogname, get_database_name(MyDatabaseId)) != 0)
5890 [ # # ]: 0 : ereport(ERROR,
5891 : : (errcode(ERRCODE_FEATURE_NOT_SUPPORTED),
5892 : : errmsg("cross-database references are not implemented: \"%s.%s.%s\"",
5893 : : constraint->catalogname, constraint->schemaname,
5894 : : constraint->relname)));
5895 : : }
5896 : :
5897 : : /*
5898 : : * If we're given the schema name with the constraint, look only
5899 : : * in that schema. If given a bare constraint name, use the
5900 : : * search path to find the first matching constraint.
5901 : : */
7153 bruce@momjian.us 5902 [ + + ]:CBC 32 : if (constraint->schemaname)
5903 : : {
4847 5904 : 8 : Oid namespaceId = LookupExplicitNamespace(constraint->schemaname,
5905 : : false);
5906 : :
5952 tgl@sss.pgh.pa.us 5907 : 8 : namespacelist = list_make1_oid(namespaceId);
5908 : : }
5909 : : else
5910 : : {
5911 : 24 : namespacelist = fetch_search_path(true);
5912 : : }
5913 : :
9715 JanWieck@Yahoo.com 5914 : 32 : found = false;
5952 tgl@sss.pgh.pa.us 5915 [ + - + - : 80 : foreach(nslc, namespacelist)
+ - ]
5916 : : {
5917 : 80 : Oid namespaceId = lfirst_oid(nslc);
5918 : : SysScanDesc conscan;
5919 : : ScanKeyData skey[2];
5920 : : HeapTuple tup;
5921 : :
5922 : 80 : ScanKeyInit(&skey[0],
5923 : : Anum_pg_constraint_conname,
5924 : : BTEqualStrategyNumber, F_NAMEEQ,
5925 : 80 : CStringGetDatum(constraint->relname));
5926 : 80 : ScanKeyInit(&skey[1],
5927 : : Anum_pg_constraint_connamespace,
5928 : : BTEqualStrategyNumber, F_OIDEQ,
5929 : : ObjectIdGetDatum(namespaceId));
5930 : :
5931 : 80 : conscan = systable_beginscan(conrel, ConstraintNameNspIndexId,
5932 : : true, NULL, 2, skey);
5933 : :
5934 [ + + ]: 144 : while (HeapTupleIsValid(tup = systable_getnext(conscan)))
5935 : : {
5936 : 64 : Form_pg_constraint con = (Form_pg_constraint) GETSTRUCT(tup);
5937 : :
5938 [ + - ]: 64 : if (con->condeferrable)
2723 andres@anarazel.de 5939 : 64 : conoidlist = lappend_oid(conoidlist, con->oid);
5952 tgl@sss.pgh.pa.us 5940 [ # # ]:UBC 0 : else if (stmt->deferred)
5941 [ # # ]: 0 : ereport(ERROR,
5942 : : (errcode(ERRCODE_WRONG_OBJECT_TYPE),
5943 : : errmsg("constraint \"%s\" is not deferrable",
5944 : : constraint->relname)));
7313 bruce@momjian.us 5945 :CBC 64 : found = true;
5946 : : }
5947 : :
5952 tgl@sss.pgh.pa.us 5948 : 80 : systable_endscan(conscan);
5949 : :
5950 : : /*
5951 : : * Once we've found a matching constraint we do not search
5952 : : * later parts of the search path.
5953 : : */
7313 bruce@momjian.us 5954 [ + + ]: 80 : if (found)
5955 : 32 : break;
5956 : : }
5957 : :
5952 tgl@sss.pgh.pa.us 5958 : 32 : list_free(namespacelist);
5959 : :
5960 : : /*
5961 : : * Not found ?
5962 : : */
8662 bruce@momjian.us 5963 [ - + ]: 32 : if (!found)
8325 tgl@sss.pgh.pa.us 5964 [ # # ]:UBC 0 : ereport(ERROR,
5965 : : (errcode(ERRCODE_UNDEFINED_OBJECT),
5966 : : errmsg("constraint \"%s\" does not exist",
5967 : : constraint->relname)));
5968 : : }
5969 : :
5970 : : /*
5971 : : * Scan for any possible descendants of the constraints. We append
5972 : : * whatever we find to the same list that we're scanning; this has the
5973 : : * effect that we create new scans for those, too, so if there are
5974 : : * further descendents, we'll also catch them.
5975 : : */
2965 alvherre@alvh.no-ip. 5976 [ + - + + :CBC 172 : foreach(lc, conoidlist)
+ + ]
5977 : : {
5978 : 140 : Oid parent = lfirst_oid(lc);
5979 : : ScanKeyData key;
5980 : : SysScanDesc scan;
5981 : : HeapTuple tuple;
5982 : :
5983 : 140 : ScanKeyInit(&key,
5984 : : Anum_pg_constraint_conparentid,
5985 : : BTEqualStrategyNumber, F_OIDEQ,
5986 : : ObjectIdGetDatum(parent));
5987 : :
5988 : 140 : scan = systable_beginscan(conrel, ConstraintParentIndexId, true, NULL, 1, &key);
5989 : :
5990 [ + + ]: 216 : while (HeapTupleIsValid(tuple = systable_getnext(scan)))
5991 : : {
2723 andres@anarazel.de 5992 : 76 : Form_pg_constraint con = (Form_pg_constraint) GETSTRUCT(tuple);
5993 : :
5994 : 76 : conoidlist = lappend_oid(conoidlist, con->oid);
5995 : : }
5996 : :
2965 alvherre@alvh.no-ip. 5997 : 140 : systable_endscan(scan);
5998 : : }
5999 : :
2661 andres@anarazel.de 6000 : 32 : table_close(conrel, AccessShareLock);
6001 : :
6002 : : /*
6003 : : * Now, locate the trigger(s) implementing each of these constraints,
6004 : : * and make a list of their OIDs.
6005 : : */
6006 : 32 : tgrel = table_open(TriggerRelationId, AccessShareLock);
6007 : :
5952 tgl@sss.pgh.pa.us 6008 [ + - + + : 172 : foreach(lc, conoidlist)
+ + ]
6009 : : {
6010 : 140 : Oid conoid = lfirst_oid(lc);
6011 : : ScanKeyData skey;
6012 : : SysScanDesc tgscan;
6013 : : HeapTuple htup;
6014 : :
6015 : 140 : ScanKeyInit(&skey,
6016 : : Anum_pg_trigger_tgconstraint,
6017 : : BTEqualStrategyNumber, F_OIDEQ,
6018 : : ObjectIdGetDatum(conoid));
6019 : :
6020 : 140 : tgscan = systable_beginscan(tgrel, TriggerConstraintIndexId, true,
6021 : : NULL, 1, &skey);
6022 : :
6023 [ + + ]: 572 : while (HeapTupleIsValid(htup = systable_getnext(tgscan)))
6024 : : {
6025 : 292 : Form_pg_trigger pg_trigger = (Form_pg_trigger) GETSTRUCT(htup);
6026 : :
6027 : : /*
6028 : : * Silently skip triggers that are marked as non-deferrable in
6029 : : * pg_trigger. This is not an error condition, since a
6030 : : * deferrable RI constraint may have some non-deferrable
6031 : : * actions.
6032 : : */
6033 [ + - ]: 292 : if (pg_trigger->tgdeferrable)
2723 andres@anarazel.de 6034 : 292 : tgoidlist = lappend_oid(tgoidlist, pg_trigger->oid);
6035 : : }
6036 : :
5952 tgl@sss.pgh.pa.us 6037 : 140 : systable_endscan(tgscan);
6038 : : }
6039 : :
2661 andres@anarazel.de 6040 : 32 : table_close(tgrel, AccessShareLock);
6041 : :
6042 : : /*
6043 : : * Now we can set the trigger states of individual triggers for this
6044 : : * xact.
6045 : : */
5952 tgl@sss.pgh.pa.us 6046 [ + - + + : 324 : foreach(lc, tgoidlist)
+ + ]
6047 : : {
6048 : 292 : Oid tgoid = lfirst_oid(lc);
4212 rhaas@postgresql.org 6049 : 292 : SetConstraintState state = afterTriggers.state;
7978 tgl@sss.pgh.pa.us 6050 : 292 : bool found = false;
6051 : : int i;
6052 : :
7909 6053 [ + + ]: 1632 : for (i = 0; i < state->numstates; i++)
6054 : : {
7907 6055 [ + + ]: 1404 : if (state->trigstates[i].sct_tgoid == tgoid)
6056 : : {
6057 : 64 : state->trigstates[i].sct_tgisdeferred = stmt->deferred;
9715 JanWieck@Yahoo.com 6058 : 64 : found = true;
6059 : 64 : break;
6060 : : }
6061 : : }
6062 [ + + ]: 292 : if (!found)
6063 : : {
4212 rhaas@postgresql.org 6064 : 228 : afterTriggers.state =
7907 tgl@sss.pgh.pa.us 6065 : 228 : SetConstraintStateAddItem(state, tgoid, stmt->deferred);
6066 : : }
6067 : : }
6068 : : }
6069 : :
6070 : : /*
6071 : : * SQL99 requires that when a constraint is set to IMMEDIATE, any deferred
6072 : : * checks against that constraint must be made when the SET CONSTRAINTS
6073 : : * command is executed -- i.e. the effects of the SET CONSTRAINTS command
6074 : : * apply retroactively. We've updated the constraints state, so scan the
6075 : : * list of previously deferred events to fire any that have now become
6076 : : * immediate.
6077 : : *
6078 : : * Obviously, if this was SET ... DEFERRED then it can't have converted
6079 : : * any unfired events to immediate, so we need do nothing in that case.
6080 : : */
6081 [ + + ]: 71 : if (!stmt->deferred)
6082 : : {
4212 rhaas@postgresql.org 6083 : 26 : AfterTriggerEventList *events = &afterTriggers.events;
6352 tgl@sss.pgh.pa.us 6084 : 26 : bool snapshot_set = false;
6085 : :
25 amitlan@postgresql.o 6086 :GNC 26 : afterTriggers.firing_depth++;
6838 tgl@sss.pgh.pa.us 6087 [ + + ]:CBC 26 : while (afterTriggerMarkEvents(events, NULL, true))
6088 : : {
4212 rhaas@postgresql.org 6089 : 14 : CommandId firing_id = afterTriggers.firing_counter++;
6090 : :
6091 : : /*
6092 : : * Make sure a snapshot has been established in case trigger
6093 : : * functions need one. Note that we avoid setting a snapshot if
6094 : : * we don't find at least one trigger that has to be fired now.
6095 : : * This is so that BEGIN; SET CONSTRAINTS ...; SET TRANSACTION
6096 : : * ISOLATION LEVEL SERIALIZABLE; ... works properly. (If we are
6097 : : * at the start of a transaction it's not possible for any trigger
6098 : : * events to be queued yet.)
6099 : : */
6352 tgl@sss.pgh.pa.us 6100 [ + - ]: 14 : if (!snapshot_set)
6101 : : {
6102 : 14 : PushActiveSnapshot(GetTransactionSnapshot());
6103 : 14 : snapshot_set = true;
6104 : : }
6105 : :
6106 : : /*
6107 : : * We can delete fired events if we are at top transaction level,
6108 : : * but we'd better not if inside a subtransaction, since the
6109 : : * subtransaction could later get rolled back.
6110 : : */
6402 tgl@sss.pgh.pa.us 6111 [ + - ]:GBC 10 : if (afterTriggerInvokeEvents(events, firing_id, NULL,
6402 tgl@sss.pgh.pa.us 6112 :CBC 14 : !IsSubTransaction()))
6402 tgl@sss.pgh.pa.us 6113 :GBC 10 : break; /* all fired */
6114 : : }
6115 : :
6116 : : /*
6117 : : * Flush any fast-path batches accumulated by the triggers just fired.
6118 : : */
25 amitlan@postgresql.o 6119 :GNC 22 : FireAfterTriggerBatchCallbacks(afterTriggers.batch_callbacks);
6120 : 16 : afterTriggers.firing_depth--;
6121 : 16 : list_free_deep(afterTriggers.batch_callbacks);
6122 : 16 : afterTriggers.batch_callbacks = NIL;
6123 : :
6352 tgl@sss.pgh.pa.us 6124 [ + + ]:CBC 16 : if (snapshot_set)
6352 tgl@sss.pgh.pa.us 6125 :GBC 4 : PopActiveSnapshot();
6126 : : }
9715 JanWieck@Yahoo.com 6127 :CBC 61 : }
6128 : :
6129 : : /* ----------
6130 : : * AfterTriggerPendingOnRel()
6131 : : * Test to see if there are any pending after-trigger events for rel.
6132 : : *
6133 : : * This is used by TRUNCATE, CLUSTER, ALTER TABLE, etc to detect whether
6134 : : * it is unsafe to perform major surgery on a relation. Note that only
6135 : : * local pending events are examined. We assume that having exclusive lock
6136 : : * on a rel guarantees there are no unserviced events in other backends ---
6137 : : * but having a lock does not prevent there being such events in our own.
6138 : : *
6139 : : * In some scenarios it'd be reasonable to remove pending events (more
6140 : : * specifically, mark them DONE by the current subxact) but without a lot
6141 : : * of knowledge of the trigger semantics we can't do this in general.
6142 : : * ----------
6143 : : */
6144 : : bool
6698 tgl@sss.pgh.pa.us 6145 : 93333 : AfterTriggerPendingOnRel(Oid relid)
6146 : : {
6147 : : AfterTriggerEvent event;
6148 : : AfterTriggerEventChunk *chunk;
6149 : : int depth;
6150 : :
6151 : : /* Scan queued events */
4212 rhaas@postgresql.org 6152 [ + - + - : 93357 : for_each_event_chunk(event, chunk, afterTriggers.events)
+ - + + +
+ ]
6153 : : {
6402 tgl@sss.pgh.pa.us 6154 : 24 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
6155 : :
6156 : : /*
6157 : : * We can ignore completed events. (Even if a DONE flag is rolled
6158 : : * back by subxact abort, it's OK because the effects of the TRUNCATE
6159 : : * or whatever must get rolled back too.)
6160 : : */
6161 [ - + ]: 24 : if (event->ate_flags & AFTER_TRIGGER_DONE)
7183 tgl@sss.pgh.pa.us 6162 :UBC 0 : continue;
6163 : :
6402 tgl@sss.pgh.pa.us 6164 [ + + ]:CBC 24 : if (evtshared->ats_relid == relid)
6698 6165 : 12 : return true;
6166 : : }
6167 : :
6168 : : /*
6169 : : * Also scan events queued by incomplete queries. This could only matter
6170 : : * if TRUNCATE/etc is executed by a function or trigger within an updating
6171 : : * query on the same relation, which is pretty perverse, but let's check.
6172 : : */
4194 rhaas@postgresql.org 6173 [ - + - - ]: 93321 : for (depth = 0; depth <= afterTriggers.query_depth && depth < afterTriggers.maxquerydepth; depth++)
6174 : : {
3153 tgl@sss.pgh.pa.us 6175 [ # # # # :UBC 0 : for_each_event_chunk(event, chunk, afterTriggers.query_stack[depth].events)
# # # # #
# ]
6176 : : {
6402 6177 : 0 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
6178 : :
6179 [ # # ]: 0 : if (event->ate_flags & AFTER_TRIGGER_DONE)
7183 6180 : 0 : continue;
6181 : :
6402 6182 [ # # ]: 0 : if (evtshared->ats_relid == relid)
6698 6183 : 0 : return true;
6184 : : }
6185 : : }
6186 : :
6698 tgl@sss.pgh.pa.us 6187 :CBC 93321 : return false;
6188 : : }
6189 : :
6190 : : /* ----------
6191 : : * AfterTriggerSaveEvent()
6192 : : *
6193 : : * Called by ExecA[RS]...Triggers() to queue up the triggers that should
6194 : : * be fired for an event.
6195 : : *
6196 : : * NOTE: this is called whenever there are any triggers associated with
6197 : : * the event (even if they are disabled). This function decides which
6198 : : * triggers actually need to be queued. It is also called after each row,
6199 : : * even if there are no triggers for that event, if there are any AFTER
6200 : : * STATEMENT triggers for the statement which use transition tables, so that
6201 : : * the transition tuplestores can be built. Furthermore, if the transition
6202 : : * capture is happening for UPDATEd rows being moved to another partition due
6203 : : * to the partition-key being changed, then this function is called once when
6204 : : * the row is deleted (to capture OLD row), and once when the row is inserted
6205 : : * into another partition (to capture NEW row). This is done separately because
6206 : : * DELETE and INSERT happen on different tables.
6207 : : *
6208 : : * Transition tuplestores are built now, rather than when events are pulled
6209 : : * off of the queue because AFTER ROW triggers are allowed to select from the
6210 : : * transition tables for the statement.
6211 : : *
6212 : : * This contains special support to queue the update events for the case where
6213 : : * a partitioned table undergoing a cross-partition update may have foreign
6214 : : * keys pointing into it. Normally, a partitioned table's row triggers are
6215 : : * not fired because the leaf partition(s) which are modified as a result of
6216 : : * the operation on the partitioned table contain the same triggers which are
6217 : : * fired instead. But that general scheme can cause problematic behavior with
6218 : : * foreign key triggers during cross-partition updates, which are implemented
6219 : : * as DELETE on the source partition followed by INSERT into the destination
6220 : : * partition. Specifically, firing DELETE triggers would lead to the wrong
6221 : : * foreign key action to be enforced considering that the original command is
6222 : : * UPDATE; in this case, this function is called with relinfo as the
6223 : : * partitioned table, and src_partinfo and dst_partinfo referring to the
6224 : : * source and target leaf partitions, respectively.
6225 : : *
6226 : : * is_crosspart_update is true either when a DELETE event is fired on the
6227 : : * source partition (which is to be ignored) or an UPDATE event is fired on
6228 : : * the root partitioned table.
6229 : : * ----------
6230 : : */
6231 : : static void
6010 6232 : 452928 : AfterTriggerSaveEvent(EState *estate, ResultRelInfo *relinfo,
6233 : : ResultRelInfo *src_partinfo,
6234 : : ResultRelInfo *dst_partinfo,
6235 : : int event, bool row_trigger,
6236 : : TupleTableSlot *oldslot, TupleTableSlot *newslot,
6237 : : List *recheckIndexes, Bitmapset *modifiedCols,
6238 : : TransitionCaptureState *transition_capture,
6239 : : bool is_crosspart_update)
6240 : : {
9104 6241 : 452928 : Relation rel = relinfo->ri_RelationDesc;
6242 : 452928 : TriggerDesc *trigdesc = relinfo->ri_TrigDesc;
6243 : : AfterTriggerEventData new_event;
6244 : : AfterTriggerSharedData new_shared;
3153 6245 : 452928 : char relkind = rel->rd_rel->relkind;
6246 : : int tgtype_event;
6247 : : int tgtype_level;
6248 : : int i;
4426 noah@leadboat.com 6249 : 452928 : Tuplestorestate *fdw_tuplestore = NULL;
6250 : :
6251 : : /*
6252 : : * Check state. We use a normal test not Assert because it is possible to
6253 : : * reach here in the wrong state given misconfigured RI triggers, in
6254 : : * particular deferring a cascade action trigger.
6255 : : */
4212 rhaas@postgresql.org 6256 [ - + ]: 452928 : if (afterTriggers.query_depth < 0)
6034 tgl@sss.pgh.pa.us 6257 [ # # ]:UBC 0 : elog(ERROR, "AfterTriggerSaveEvent() called outside of query");
6258 : :
6259 : : /* Be sure we have enough space to record events at this query depth. */
4212 rhaas@postgresql.org 6260 [ + + ]:CBC 452928 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
6261 : 4166 : AfterTriggerEnlargeQueryState();
6262 : :
6263 : : /*
6264 : : * If the directly named relation has any triggers with transition tables,
6265 : : * then we need to capture transition tuples.
6266 : : */
3233 rhodiumtoad@postgres 6267 [ + + + + ]: 452928 : if (row_trigger && transition_capture != NULL)
6268 : : {
1499 alvherre@alvh.no-ip. 6269 : 43911 : TupleTableSlot *original_insert_tuple = transition_capture->tcs_original_insert_tuple;
6270 : :
6271 : : /*
6272 : : * Capture the old tuple in the appropriate transition table based on
6273 : : * the event.
6274 : : */
1516 6275 [ + + + + ]: 43911 : if (!TupIsNull(oldslot))
6276 : : {
6277 : : Tuplestorestate *old_tuplestore;
6278 : :
6279 : 3619 : old_tuplestore = GetAfterTriggersTransitionTable(event,
6280 : : oldslot,
6281 : : NULL,
6282 : : transition_capture);
101 dean.a.rasheed@gmail 6283 : 3619 : TransitionTableAddTuple(estate, event, transition_capture, relinfo,
6284 : : oldslot, NULL, old_tuplestore);
6285 : : }
6286 : :
6287 : : /*
6288 : : * Capture the new tuple in the appropriate transition table based on
6289 : : * the event.
6290 : : */
1516 alvherre@alvh.no-ip. 6291 [ + + + - ]: 43911 : if (!TupIsNull(newslot))
6292 : : {
6293 : : Tuplestorestate *new_tuplestore;
6294 : :
6295 : 40507 : new_tuplestore = GetAfterTriggersTransitionTable(event,
6296 : : NULL,
6297 : : newslot,
6298 : : transition_capture);
101 dean.a.rasheed@gmail 6299 : 40507 : TransitionTableAddTuple(estate, event, transition_capture, relinfo,
6300 : : newslot, original_insert_tuple, new_tuplestore);
6301 : : }
6302 : :
6303 : : /*
6304 : : * If transition tables are the only reason we're here, return. As
6305 : : * mentioned above, we can also be here during update tuple routing in
6306 : : * presence of transition tables, in which case this function is
6307 : : * called separately for OLD and NEW, so we expect exactly one of them
6308 : : * to be NULL.
6309 : : */
3233 rhodiumtoad@postgres 6310 [ + + + + ]: 43911 : if (trigdesc == NULL ||
6311 [ + + + + ]: 43751 : (event == TRIGGER_EVENT_DELETE && !trigdesc->trig_delete_after_row) ||
3240 tgl@sss.pgh.pa.us 6312 [ + + + + ]: 40431 : (event == TRIGGER_EVENT_INSERT && !trigdesc->trig_insert_after_row) ||
3028 rhaas@postgresql.org 6313 [ + + + + ]: 263 : (event == TRIGGER_EVENT_UPDATE && !trigdesc->trig_update_after_row) ||
2625 andres@anarazel.de 6314 [ + - - + : 32 : (event == TRIGGER_EVENT_UPDATE && (TupIsNull(oldslot) ^ TupIsNull(newslot))))
+ - - + -
+ ]
3469 kgrittn@postgresql.o 6315 : 43807 : return;
6316 : : }
6317 : :
6318 : : /*
6319 : : * We normally don't see partitioned tables here for row level triggers
6320 : : * except in the special case of a cross-partition update. In that case,
6321 : : * nodeModifyTable.c:ExecCrossPartitionUpdateForeignKey() calls here to
6322 : : * queue an update event on the root target partitioned table, also
6323 : : * passing the source and destination partitions and their tuples.
6324 : : */
1507 alvherre@alvh.no-ip. 6325 [ + + + + : 409121 : Assert(!row_trigger ||
+ - + - +
- - + ]
6326 : : rel->rd_rel->relkind != RELKIND_PARTITIONED_TABLE ||
6327 : : (is_crosspart_update &&
6328 : : TRIGGER_FIRED_BY_UPDATE(event) &&
6329 : : src_partinfo != NULL && dst_partinfo != NULL));
6330 : :
6331 : : /*
6332 : : * Validate the event code and collect the associated tuple CTIDs.
6333 : : *
6334 : : * The event code will be used both as a bitmask and an array offset, so
6335 : : * validation is important to make sure we don't walk off the edge of our
6336 : : * arrays.
6337 : : *
6338 : : * Also, if we're considering statement-level triggers, check whether we
6339 : : * already queued a set of them for this event, and cancel the prior set
6340 : : * if so. This preserves the behavior that statement-level triggers fire
6341 : : * just once per statement and fire after row-level triggers.
6342 : : */
6402 tgl@sss.pgh.pa.us 6343 [ + + + + : 409121 : switch (event)
- ]
6344 : : {
6345 : 405666 : case TRIGGER_EVENT_INSERT:
5686 6346 : 405666 : tgtype_event = TRIGGER_TYPE_INSERT;
6402 6347 [ + + ]: 405666 : if (row_trigger)
6348 : : {
2625 andres@anarazel.de 6349 [ - + ]: 405317 : Assert(oldslot == NULL);
6350 [ - + ]: 405317 : Assert(newslot != NULL);
6351 : 405317 : ItemPointerCopy(&(newslot->tts_tid), &(new_event.ate_ctid1));
6402 tgl@sss.pgh.pa.us 6352 : 405317 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6353 : : }
6354 : : else
6355 : : {
2625 andres@anarazel.de 6356 [ - + ]: 349 : Assert(oldslot == NULL);
6357 [ - + ]: 349 : Assert(newslot == NULL);
6402 tgl@sss.pgh.pa.us 6358 : 349 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6359 : 349 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
3153 6360 : 349 : cancel_prior_stmt_triggers(RelationGetRelid(rel),
6361 : : CMD_INSERT, event);
6362 : : }
6402 6363 : 405666 : break;
6364 : 976 : case TRIGGER_EVENT_DELETE:
5686 6365 : 976 : tgtype_event = TRIGGER_TYPE_DELETE;
6402 6366 [ + + ]: 976 : if (row_trigger)
6367 : : {
2625 andres@anarazel.de 6368 [ - + ]: 811 : Assert(oldslot != NULL);
6369 [ - + ]: 811 : Assert(newslot == NULL);
6370 : 811 : ItemPointerCopy(&(oldslot->tts_tid), &(new_event.ate_ctid1));
6402 tgl@sss.pgh.pa.us 6371 : 811 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6372 : : }
6373 : : else
6374 : : {
2625 andres@anarazel.de 6375 [ - + ]: 165 : Assert(oldslot == NULL);
6376 [ - + ]: 165 : Assert(newslot == NULL);
6402 tgl@sss.pgh.pa.us 6377 : 165 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6378 : 165 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
3153 6379 : 165 : cancel_prior_stmt_triggers(RelationGetRelid(rel),
6380 : : CMD_DELETE, event);
6381 : : }
6402 6382 : 976 : break;
6383 : 2474 : case TRIGGER_EVENT_UPDATE:
5686 6384 : 2474 : tgtype_event = TRIGGER_TYPE_UPDATE;
6402 6385 [ + + ]: 2474 : if (row_trigger)
6386 : : {
2625 andres@anarazel.de 6387 [ - + ]: 2197 : Assert(oldslot != NULL);
6388 [ - + ]: 2197 : Assert(newslot != NULL);
6389 : 2197 : ItemPointerCopy(&(oldslot->tts_tid), &(new_event.ate_ctid1));
6390 : 2197 : ItemPointerCopy(&(newslot->tts_tid), &(new_event.ate_ctid2));
6391 : :
6392 : : /*
6393 : : * Also remember the OIDs of partitions to fetch these tuples
6394 : : * out of later in AfterTriggerExecute().
6395 : : */
1507 alvherre@alvh.no-ip. 6396 [ + + ]: 2197 : if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6397 : : {
6398 [ + - - + ]: 192 : Assert(src_partinfo != NULL && dst_partinfo != NULL);
6399 : 192 : new_event.ate_src_part =
6400 : 192 : RelationGetRelid(src_partinfo->ri_RelationDesc);
6401 : 192 : new_event.ate_dst_part =
6402 : 192 : RelationGetRelid(dst_partinfo->ri_RelationDesc);
6403 : : }
6404 : : }
6405 : : else
6406 : : {
2625 andres@anarazel.de 6407 [ - + ]: 277 : Assert(oldslot == NULL);
6408 [ - + ]: 277 : Assert(newslot == NULL);
6402 tgl@sss.pgh.pa.us 6409 : 277 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6410 : 277 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
3153 6411 : 277 : cancel_prior_stmt_triggers(RelationGetRelid(rel),
6412 : : CMD_UPDATE, event);
6413 : : }
6402 6414 : 2474 : break;
6415 : 5 : case TRIGGER_EVENT_TRUNCATE:
5686 6416 : 5 : tgtype_event = TRIGGER_TYPE_TRUNCATE;
2625 andres@anarazel.de 6417 [ - + ]: 5 : Assert(oldslot == NULL);
6418 [ - + ]: 5 : Assert(newslot == NULL);
6402 tgl@sss.pgh.pa.us 6419 : 5 : ItemPointerSetInvalid(&(new_event.ate_ctid1));
6420 : 5 : ItemPointerSetInvalid(&(new_event.ate_ctid2));
6421 : 5 : break;
6402 tgl@sss.pgh.pa.us 6422 :UBC 0 : default:
6423 [ # # ]: 0 : elog(ERROR, "invalid after-trigger event code: %d", event);
6424 : : tgtype_event = 0; /* keep compiler quiet */
6425 : : break;
6426 : : }
6427 : :
6428 : : /* Determine flags */
4426 noah@leadboat.com 6429 [ + + + + ]:CBC 409121 : if (!(relkind == RELKIND_FOREIGN_TABLE && row_trigger))
6430 : : {
1507 alvherre@alvh.no-ip. 6431 [ + + + + ]: 409093 : if (row_trigger && event == TRIGGER_EVENT_UPDATE)
6432 : : {
6433 [ + + ]: 2187 : if (relkind == RELKIND_PARTITIONED_TABLE)
6434 : 192 : new_event.ate_flags = AFTER_TRIGGER_CP_UPDATE;
6435 : : else
6436 : 1995 : new_event.ate_flags = AFTER_TRIGGER_2CTID;
6437 : : }
6438 : : else
6439 : 406906 : new_event.ate_flags = AFTER_TRIGGER_1CTID;
6440 : : }
6441 : :
6442 : : /* else, we'll initialize ate_flags for each trigger */
6443 : :
5686 tgl@sss.pgh.pa.us 6444 : 409121 : tgtype_level = (row_trigger ? TRIGGER_TYPE_ROW : TRIGGER_TYPE_STATEMENT);
6445 : :
6446 : : /*
6447 : : * Must convert/copy the source and destination partition tuples into the
6448 : : * root partitioned table's format/slot, because the processing in the
6449 : : * loop below expects both oldslot and newslot tuples to be in that form.
6450 : : */
1507 alvherre@alvh.no-ip. 6451 [ + + + + ]: 409121 : if (row_trigger && rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6452 : : {
6453 : : TupleTableSlot *rootslot;
6454 : : TupleConversionMap *map;
6455 : :
6456 : 192 : rootslot = ExecGetTriggerOldSlot(estate, relinfo);
6457 : 192 : map = ExecGetChildToRootMap(src_partinfo);
6458 [ + + ]: 192 : if (map)
6459 : 24 : oldslot = execute_attr_map_slot(map->attrMap,
6460 : : oldslot,
6461 : : rootslot);
6462 : : else
6463 : 168 : oldslot = ExecCopySlot(rootslot, oldslot);
6464 : :
6465 : 192 : rootslot = ExecGetTriggerNewSlot(estate, relinfo);
6466 : 192 : map = ExecGetChildToRootMap(dst_partinfo);
6467 [ + + ]: 192 : if (map)
6468 : 28 : newslot = execute_attr_map_slot(map->attrMap,
6469 : : newslot,
6470 : : rootslot);
6471 : : else
6472 : 164 : newslot = ExecCopySlot(rootslot, newslot);
6473 : : }
6474 : :
5686 tgl@sss.pgh.pa.us 6475 [ + + ]: 1639400 : for (i = 0; i < trigdesc->numtriggers; i++)
6476 : : {
6477 : 1230279 : Trigger *trigger = &trigdesc->triggers[i];
6478 : :
6479 [ + + ]: 1230279 : if (!TRIGGER_TYPE_MATCHES(trigger->tgtype,
6480 : : tgtype_level,
6481 : : TRIGGER_TYPE_AFTER,
6482 : : tgtype_event))
6483 : 619189 : continue;
6010 6484 [ + + ]: 611090 : if (!TriggerEnabled(estate, relinfo, trigger, event,
6485 : : modifiedCols, oldslot, newslot))
6047 6486 : 273 : continue;
6487 : :
4426 noah@leadboat.com 6488 [ + + + + ]: 610817 : if (relkind == RELKIND_FOREIGN_TABLE && row_trigger)
6489 : : {
6490 [ + + ]: 29 : if (fdw_tuplestore == NULL)
6491 : : {
3153 tgl@sss.pgh.pa.us 6492 : 25 : fdw_tuplestore = GetCurrentFDWTuplestore();
4426 noah@leadboat.com 6493 : 25 : new_event.ate_flags = AFTER_TRIGGER_FDW_FETCH;
6494 : : }
6495 : : else
6496 : : /* subsequent event for the same tuple */
6497 : 4 : new_event.ate_flags = AFTER_TRIGGER_FDW_REUSE;
6498 : : }
6499 : :
6500 : : /*
6501 : : * If the trigger is a foreign key enforcement trigger, there are
6502 : : * certain cases where we can skip queueing the event because we can
6503 : : * tell by inspection that the FK constraint will still pass. There
6504 : : * are also some cases during cross-partition updates of a partitioned
6505 : : * table where queuing the event can be skipped.
6506 : : */
2847 peter_e@gmx.net 6507 [ + + + + ]: 610817 : if (TRIGGER_FIRED_BY_UPDATE(event) || TRIGGER_FIRED_BY_DELETE(event))
6508 : : {
7645 neilc@samurai.com 6509 [ + + + - ]: 4354 : switch (RI_FKey_trigger_type(trigger->tgfoid))
6510 : : {
6511 : 1701 : case RI_TRIGGER_PK:
6512 : :
6513 : : /*
6514 : : * For cross-partitioned updates of partitioned PK table,
6515 : : * skip the event fired by the component delete on the
6516 : : * source leaf partition unless the constraint originates
6517 : : * in the partition itself (!tgisclone), because the
6518 : : * update event that will be fired on the root
6519 : : * (partitioned) target table will be used to perform the
6520 : : * necessary foreign key enforcement action.
6521 : : */
1507 alvherre@alvh.no-ip. 6522 [ + + ]: 1701 : if (is_crosspart_update &&
6523 [ + + ]: 338 : TRIGGER_FIRED_BY_DELETE(event) &&
6524 [ + + ]: 178 : trigger->tgisclone)
6525 : 168 : continue;
6526 : :
6527 : : /* Update or delete on trigger's PK table */
5068 tgl@sss.pgh.pa.us 6528 [ + + ]: 1533 : if (!RI_FKey_pk_upd_check_required(trigger, rel,
6529 : : oldslot, newslot))
6530 : : {
6531 : : /* skip queuing this event */
7645 neilc@samurai.com 6532 : 292 : continue;
6533 : : }
7907 tgl@sss.pgh.pa.us 6534 : 1241 : break;
6535 : :
7645 neilc@samurai.com 6536 : 748 : case RI_TRIGGER_FK:
6537 : :
6538 : : /*
6539 : : * Update on trigger's FK table. We can skip the update
6540 : : * event fired on a partitioned table during a
6541 : : * cross-partition of that table, because the insert event
6542 : : * that is fired on the destination leaf partition would
6543 : : * suffice to perform the necessary foreign key check.
6544 : : * Moreover, RI_FKey_fk_upd_check_required() expects to be
6545 : : * passed a tuple that contains system attributes, most of
6546 : : * which are not present in the virtual slot belonging to
6547 : : * a partitioned table.
6548 : : */
1507 alvherre@alvh.no-ip. 6549 [ + + ]: 748 : if (rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE ||
6550 [ + + ]: 668 : !RI_FKey_fk_upd_check_required(trigger, rel,
6551 : : oldslot, newslot))
6552 : : {
6553 : : /* skip queuing this event */
7645 neilc@samurai.com 6554 : 442 : continue;
6555 : : }
7907 tgl@sss.pgh.pa.us 6556 : 306 : break;
6557 : :
7645 neilc@samurai.com 6558 : 1905 : case RI_TRIGGER_NONE:
6559 : :
6560 : : /*
6561 : : * Not an FK trigger. No need to queue the update event
6562 : : * fired during a cross-partitioned update of a
6563 : : * partitioned table, because the same row trigger must be
6564 : : * present in the leaf partition(s) that are affected as
6565 : : * part of this update and the events fired on them are
6566 : : * queued instead.
6567 : : */
1507 alvherre@alvh.no-ip. 6568 [ + + ]: 1905 : if (row_trigger &&
6569 [ + + ]: 1439 : rel->rd_rel->relkind == RELKIND_PARTITIONED_TABLE)
6570 : 20 : continue;
7645 neilc@samurai.com 6571 : 1885 : break;
6572 : : }
6573 : : }
6574 : :
6575 : : /*
6576 : : * If the trigger is a deferred unique constraint check trigger, only
6577 : : * queue it if the unique constraint was potentially violated, which
6578 : : * we know from index insertion time.
6579 : : */
6124 tgl@sss.pgh.pa.us 6580 [ + + ]: 609895 : if (trigger->tgfoid == F_UNIQUE_KEY_RECHECK)
6581 : : {
6582 [ + + ]: 139 : if (!list_member_oid(recheckIndexes, trigger->tgconstrindid))
6583 : 58 : continue; /* Uniqueness definitely not violated */
6584 : : }
6585 : :
6586 : : /*
6587 : : * Fill in event structure and add it to the current query's queue.
6588 : : * Note we set ats_table to NULL whenever this trigger doesn't use
6589 : : * transition tables, to improve sharability of the shared event data.
6590 : : */
6402 6591 : 609837 : new_shared.ats_event =
7907 6592 : 1219674 : (event & TRIGGER_EVENT_OPMASK) |
6593 [ + + ]: 609837 : (row_trigger ? TRIGGER_EVENT_ROW : 0) |
6594 [ + + ]: 609837 : (trigger->tgdeferrable ? AFTER_TRIGGER_DEFERRABLE : 0) |
6595 [ + + ]: 609837 : (trigger->tginitdeferred ? AFTER_TRIGGER_INITDEFERRED : 0);
6402 6596 : 609837 : new_shared.ats_tgoid = trigger->tgoid;
6597 : 609837 : new_shared.ats_relid = RelationGetRelid(rel);
467 6598 : 609837 : new_shared.ats_rolid = GetUserId();
6402 6599 : 609837 : new_shared.ats_firing_id = 0;
3153 6600 [ + + + + : 609837 : if ((trigger->tgoldtable || trigger->tgnewtable) &&
+ - ]
6601 : : transition_capture != NULL)
6602 : : {
101 dean.a.rasheed@gmail 6603 [ + + + - ]: 487 : switch (event)
6604 : : {
6605 : 192 : case TRIGGER_EVENT_INSERT:
6606 : 192 : new_shared.ats_table = transition_capture->tcs_insert_private;
6607 : 192 : break;
6608 : 155 : case TRIGGER_EVENT_UPDATE:
6609 : 155 : new_shared.ats_table = transition_capture->tcs_update_private;
6610 : 155 : break;
6611 : 140 : case TRIGGER_EVENT_DELETE:
6612 : 140 : new_shared.ats_table = transition_capture->tcs_delete_private;
6613 : 140 : break;
101 dean.a.rasheed@gmail 6614 :UBC 0 : default:
6615 : : /* Must be TRUNCATE, see switch above */
6616 : 0 : new_shared.ats_table = NULL;
6617 : 0 : break;
6618 : : }
6619 : : }
6620 : : else
3153 tgl@sss.pgh.pa.us 6621 :CBC 609350 : new_shared.ats_table = NULL;
468 6622 : 609837 : new_shared.ats_modifiedcols = modifiedCols;
6623 : :
3153 6624 : 609837 : afterTriggerAddEvent(&afterTriggers.query_stack[afterTriggers.query_depth].events,
6625 : : &new_event, &new_shared);
6626 : : }
6627 : :
6628 : : /*
6629 : : * Finally, spool any foreign tuple(s). The tuplestore squashes them to
6630 : : * minimal tuples, so this loses any system columns. The executor lost
6631 : : * those columns before us, for an unrelated reason, so this is fine.
6632 : : */
4426 noah@leadboat.com 6633 [ + + ]: 409121 : if (fdw_tuplestore)
6634 : : {
2625 andres@anarazel.de 6635 [ + + ]: 25 : if (oldslot != NULL)
6636 : 16 : tuplestore_puttupleslot(fdw_tuplestore, oldslot);
6637 [ + + ]: 25 : if (newslot != NULL)
6638 : 18 : tuplestore_puttupleslot(fdw_tuplestore, newslot);
6639 : : }
6640 : : }
6641 : :
6642 : : /*
6643 : : * Detect whether we already queued BEFORE STATEMENT triggers for the given
6644 : : * relation + operation, and set the flag so the next call will report "true".
6645 : : */
6646 : : static bool
3152 tgl@sss.pgh.pa.us 6647 : 387 : before_stmt_triggers_fired(Oid relid, CmdType cmdType)
6648 : : {
6649 : : bool result;
6650 : : AfterTriggersTableData *table;
6651 : :
6652 : : /* Check state, like AfterTriggerSaveEvent. */
6653 [ - + ]: 387 : if (afterTriggers.query_depth < 0)
3152 tgl@sss.pgh.pa.us 6654 [ # # ]:UBC 0 : elog(ERROR, "before_stmt_triggers_fired() called outside of query");
6655 : :
6656 : : /* Be sure we have enough space to record events at this query depth. */
3152 tgl@sss.pgh.pa.us 6657 [ + + ]:CBC 387 : if (afterTriggers.query_depth >= afterTriggers.maxquerydepth)
6658 : 227 : AfterTriggerEnlargeQueryState();
6659 : :
6660 : : /*
6661 : : * We keep this state in the AfterTriggersTableData that also holds
6662 : : * transition tables for the relation + operation. In this way, if we are
6663 : : * forced to make a new set of transition tables because more tuples get
6664 : : * entered after we've already fired triggers, we will allow a new set of
6665 : : * statement triggers to get queued.
6666 : : */
6667 : 387 : table = GetAfterTriggersTableData(relid, cmdType);
6668 : 387 : result = table->before_trig_done;
6669 : 387 : table->before_trig_done = true;
6670 : 387 : return result;
6671 : : }
6672 : :
6673 : : /*
6674 : : * If we previously queued a set of AFTER STATEMENT triggers for the given
6675 : : * relation + operation, and they've not been fired yet, cancel them. The
6676 : : * caller will queue a fresh set that's after any row-level triggers that may
6677 : : * have been queued by the current sub-statement, preserving (as much as
6678 : : * possible) the property that AFTER ROW triggers fire before AFTER STATEMENT
6679 : : * triggers, and that the latter only fire once. This deals with the
6680 : : * situation where several FK enforcement triggers sequentially queue triggers
6681 : : * for the same table into the same trigger query level. We can't fully
6682 : : * prevent odd behavior though: if there are AFTER ROW triggers taking
6683 : : * transition tables, we don't want to change the transition tables once the
6684 : : * first such trigger has seen them. In such a case, any additional events
6685 : : * will result in creating new transition tables and allowing new firings of
6686 : : * statement triggers.
6687 : : *
6688 : : * This also saves the current event list location so that a later invocation
6689 : : * of this function can cheaply find the triggers we're about to queue and
6690 : : * cancel them.
6691 : : */
6692 : : static void
3153 6693 : 791 : cancel_prior_stmt_triggers(Oid relid, CmdType cmdType, int tgevent)
6694 : : {
6695 : : AfterTriggersTableData *table;
6696 : 791 : AfterTriggersQueryData *qs = &afterTriggers.query_stack[afterTriggers.query_depth];
6697 : :
6698 : : /*
6699 : : * We keep this state in the AfterTriggersTableData that also holds
6700 : : * transition tables for the relation + operation. In this way, if we are
6701 : : * forced to make a new set of transition tables because more tuples get
6702 : : * entered after we've already fired triggers, we will allow a new set of
6703 : : * statement triggers to get queued without canceling the old ones.
6704 : : */
6705 : 791 : table = GetAfterTriggersTableData(relid, cmdType);
6706 : :
3152 6707 [ + + ]: 791 : if (table->after_trig_done)
6708 : : {
6709 : : /*
6710 : : * We want to start scanning from the tail location that existed just
6711 : : * before we inserted any statement triggers. But the events list
6712 : : * might've been entirely empty then, in which case scan from the
6713 : : * current head.
6714 : : */
6715 : : AfterTriggerEvent event;
6716 : : AfterTriggerEventChunk *chunk;
6717 : :
6718 [ + + ]: 48 : if (table->after_trig_events.tail)
6719 : : {
6720 : 40 : chunk = table->after_trig_events.tail;
6721 : 40 : event = (AfterTriggerEvent) table->after_trig_events.tailfree;
6722 : : }
6723 : : else
6724 : : {
3153 6725 : 8 : chunk = qs->events.head;
6726 : 8 : event = NULL;
6727 : : }
6728 : :
6729 [ + + ]: 72 : for_each_chunk_from(chunk)
6730 : : {
6731 [ + + ]: 48 : if (event == NULL)
6732 : 8 : event = (AfterTriggerEvent) CHUNK_DATA_START(chunk);
6733 [ + - + - : 100 : for_each_event_from(event, chunk)
+ + ]
6734 : : {
6735 : 76 : AfterTriggerShared evtshared = GetTriggerSharedData(event);
6736 : :
6737 : : /*
6738 : : * Exit loop when we reach events that aren't AS triggers for
6739 : : * the target relation.
6740 : : */
6741 [ - + ]: 76 : if (evtshared->ats_relid != relid)
3153 tgl@sss.pgh.pa.us 6742 :UBC 0 : goto done;
3153 tgl@sss.pgh.pa.us 6743 [ - + ]:CBC 76 : if ((evtshared->ats_event & TRIGGER_EVENT_OPMASK) != tgevent)
3153 tgl@sss.pgh.pa.us 6744 :UBC 0 : goto done;
3153 tgl@sss.pgh.pa.us 6745 [ + + ]:CBC 76 : if (!TRIGGER_FIRED_FOR_STATEMENT(evtshared->ats_event))
6746 : 24 : goto done;
6747 [ - + ]: 52 : if (!TRIGGER_FIRED_AFTER(evtshared->ats_event))
3153 tgl@sss.pgh.pa.us 6748 :UBC 0 : goto done;
6749 : : /* OK, mark it DONE */
3153 tgl@sss.pgh.pa.us 6750 :CBC 52 : event->ate_flags &= ~AFTER_TRIGGER_IN_PROGRESS;
6751 [ + - ]: 52 : event->ate_flags |= AFTER_TRIGGER_DONE;
6752 : : }
6753 : : /* signal we must reinitialize event ptr for next chunk */
6754 : 24 : event = NULL;
6755 : : }
6756 : : }
6757 : 767 : done:
6758 : :
6759 : : /* In any case, save current insertion point for next time */
3152 6760 : 791 : table->after_trig_done = true;
6761 : 791 : table->after_trig_events = qs->events;
3153 6762 : 791 : }
6763 : :
6764 : : /*
6765 : : * GUC assign_hook for session_replication_role
6766 : : */
6767 : : void
1330 6768 : 1892 : assign_session_replication_role(int newval, void *extra)
6769 : : {
6770 : : /*
6771 : : * Must flush the plan cache when changing replication role; but don't
6772 : : * flush unnecessarily.
6773 : : */
6774 [ + + ]: 1892 : if (SessionReplicationRole != newval)
6775 : 606 : ResetPlanCache();
6776 : 1892 : }
6777 : :
6778 : : /*
6779 : : * SQL function pg_trigger_depth()
6780 : : */
6781 : : Datum
5214 alvherre@alvh.no-ip. 6782 : 84 : pg_trigger_depth(PG_FUNCTION_ARGS)
6783 : : {
6784 : 84 : PG_RETURN_INT32(MyTriggerDepth);
6785 : : }
6786 : :
6787 : : /*
6788 : : * Check whether a trigger modified a virtual generated column and replace the
6789 : : * value with null if so.
6790 : : *
6791 : : * We need to check this so that we don't end up storing a non-null value in a
6792 : : * virtual generated column.
6793 : : *
6794 : : * We don't need to check for stored generated columns, since those will be
6795 : : * overwritten later anyway.
6796 : : */
6797 : : static HeapTuple
452 peter@eisentraut.org 6798 : 1268 : check_modified_virtual_generated(TupleDesc tupdesc, HeapTuple tuple)
6799 : : {
6800 [ + + + + ]: 1268 : if (!(tupdesc->constr && tupdesc->constr->has_generated_virtual))
6801 : 1257 : return tuple;
6802 : :
6803 [ + + ]: 39 : for (int i = 0; i < tupdesc->natts; i++)
6804 : : {
6805 [ + + ]: 28 : if (TupleDescAttr(tupdesc, i)->attgenerated == ATTRIBUTE_GENERATED_VIRTUAL)
6806 : : {
6807 [ + + ]: 11 : if (!heap_attisnull(tuple, i + 1, tupdesc))
6808 : : {
6809 : 8 : int replCol = i + 1;
6810 : 8 : Datum replValue = 0;
6811 : 8 : bool replIsnull = true;
6812 : :
6813 : 8 : tuple = heap_modify_tuple_by_cols(tuple, tupdesc, 1, &replCol, &replValue, &replIsnull);
6814 : : }
6815 : : }
6816 : : }
6817 : :
6818 : 11 : return tuple;
6819 : : }
6820 : :
6821 : : /*
6822 : : * RegisterAfterTriggerBatchCallback
6823 : : * Register a function to be called when the current trigger-firing
6824 : : * batch completes.
6825 : : *
6826 : : * Must be called from within a trigger function's execution context
6827 : : * (i.e., while afterTriggers state is active).
6828 : : *
6829 : : * The callback list is cleared after invocation, so the caller must
6830 : : * re-register for each new batch if needed.
6831 : : */
6832 : : void
32 amitlan@postgresql.o 6833 :GNC 1496 : RegisterAfterTriggerBatchCallback(AfterTriggerBatchCallback callback,
6834 : : void *arg)
6835 : : {
6836 : : AfterTriggerCallbackItem *item;
6837 : : MemoryContext oldcxt;
6838 : :
6839 : : /*
6840 : : * Allocate in TopTransactionContext so the item survives for the duration
6841 : : * of the batch, which may span multiple trigger invocations.
6842 : : *
6843 : : * Must be called while afterTriggers is active; callbacks registered
6844 : : * outside a trigger-firing context would never fire.
6845 : : */
25 6846 [ - + ]: 1496 : Assert(afterTriggers.firing_depth > 0);
6847 [ - + ]: 1496 : Assert(!afterTriggers.firing_batch_callbacks);
32 6848 : 1496 : oldcxt = MemoryContextSwitchTo(TopTransactionContext);
6849 : 1496 : item = palloc(sizeof(AfterTriggerCallbackItem));
6850 : 1496 : item->callback = callback;
6851 : 1496 : item->arg = arg;
25 6852 [ + + ]: 1496 : if (afterTriggers.query_depth >= 0)
6853 : : {
6854 : 1420 : AfterTriggersQueryData *qs =
6855 : 1420 : &afterTriggers.query_stack[afterTriggers.query_depth];
6856 : :
6857 : 1420 : qs->batch_callbacks = lappend(qs->batch_callbacks, item);
6858 : : }
6859 : : else
6860 : 76 : afterTriggers.batch_callbacks =
6861 : 76 : lappend(afterTriggers.batch_callbacks, item);
32 6862 : 1496 : MemoryContextSwitchTo(oldcxt);
6863 : 1496 : }
6864 : :
6865 : : /*
6866 : : * FireAfterTriggerBatchCallbacks
6867 : : * Invoke all callbacks in the given list.
6868 : : *
6869 : : * Memory cleanup of the list and its items is handled by the caller
6870 : : * (AfterTriggerFreeQuery for query-level callbacks, AfterTriggerEndXact
6871 : : * for top-level deferred callbacks).
6872 : : */
6873 : : static void
25 6874 : 397775 : FireAfterTriggerBatchCallbacks(List *callbacks)
6875 : : {
6876 : : ListCell *lc;
6877 : :
6878 [ - + ]: 397775 : Assert(afterTriggers.firing_depth > 0);
6879 : 397775 : afterTriggers.firing_batch_callbacks = true;
6880 [ + + + + : 398978 : foreach(lc, callbacks)
+ + ]
6881 : : {
32 6882 : 1496 : AfterTriggerCallbackItem *item = lfirst(lc);
6883 : :
6884 : 1496 : item->callback(item->arg);
6885 : : }
25 6886 : 397482 : afterTriggers.firing_batch_callbacks = false;
32 6887 : 397482 : }
6888 : :
6889 : : /*
6890 : : * AfterTriggerIsActive
6891 : : * Returns true if we're inside the after-trigger framework where
6892 : : * registered batch callbacks will actually be invoked.
6893 : : *
6894 : : * This is false during validateForeignKeyConstraint(), which calls
6895 : : * RI trigger functions directly outside the after-trigger framework.
6896 : : */
6897 : : bool
6898 : 603689 : AfterTriggerIsActive(void)
6899 : : {
25 6900 : 603689 : return afterTriggers.firing_depth > 0;
6901 : : }
|