Line data Source code
1 : /*-------------------------------------------------------------------------
2 : *
3 : * catcache.c
4 : * System catalog cache for tuples matching a key.
5 : *
6 : * Portions Copyright (c) 1996-2024, PostgreSQL Global Development Group
7 : * Portions Copyright (c) 1994, Regents of the University of California
8 : *
9 : *
10 : * IDENTIFICATION
11 : * src/backend/utils/cache/catcache.c
12 : *
13 : *-------------------------------------------------------------------------
14 : */
15 : #include "postgres.h"
16 :
17 : #include "access/genam.h"
18 : #include "access/heaptoast.h"
19 : #include "access/relscan.h"
20 : #include "access/table.h"
21 : #include "access/xact.h"
22 : #include "catalog/pg_collation.h"
23 : #include "catalog/pg_type.h"
24 : #include "common/hashfn.h"
25 : #include "common/pg_prng.h"
26 : #include "miscadmin.h"
27 : #include "port/pg_bitutils.h"
28 : #ifdef CATCACHE_STATS
29 : #include "storage/ipc.h" /* for on_proc_exit */
30 : #endif
31 : #include "storage/lmgr.h"
32 : #include "utils/builtins.h"
33 : #include "utils/catcache.h"
34 : #include "utils/datum.h"
35 : #include "utils/fmgroids.h"
36 : #include "utils/inval.h"
37 : #include "utils/memutils.h"
38 : #include "utils/rel.h"
39 : #include "utils/resowner.h"
40 : #include "utils/syscache.h"
41 :
42 :
43 : /* #define CACHEDEBUG */ /* turns DEBUG elogs on */
44 :
45 : /*
46 : * Given a hash value and the size of the hash table, find the bucket
47 : * in which the hash value belongs. Since the hash table must contain
48 : * a power-of-2 number of elements, this is a simple bitmask.
49 : */
50 : #define HASH_INDEX(h, sz) ((Index) ((h) & ((sz) - 1)))
51 :
52 :
53 : /*
54 : * variables, macros and other stuff
55 : */
56 :
57 : #ifdef CACHEDEBUG
58 : #define CACHE_elog(...) elog(__VA_ARGS__)
59 : #else
60 : #define CACHE_elog(...)
61 : #endif
62 :
63 : /* Cache management header --- pointer is NULL until created */
64 : static CatCacheHeader *CacheHdr = NULL;
65 :
66 : static inline HeapTuple SearchCatCacheInternal(CatCache *cache,
67 : int nkeys,
68 : Datum v1, Datum v2,
69 : Datum v3, Datum v4);
70 :
71 : static pg_noinline HeapTuple SearchCatCacheMiss(CatCache *cache,
72 : int nkeys,
73 : uint32 hashValue,
74 : Index hashIndex,
75 : Datum v1, Datum v2,
76 : Datum v3, Datum v4);
77 :
78 : static uint32 CatalogCacheComputeHashValue(CatCache *cache, int nkeys,
79 : Datum v1, Datum v2, Datum v3, Datum v4);
80 : static uint32 CatalogCacheComputeTupleHashValue(CatCache *cache, int nkeys,
81 : HeapTuple tuple);
82 : static inline bool CatalogCacheCompareTuple(const CatCache *cache, int nkeys,
83 : const Datum *cachekeys,
84 : const Datum *searchkeys);
85 :
86 : #ifdef CATCACHE_STATS
87 : static void CatCachePrintStats(int code, Datum arg);
88 : #endif
89 : static void CatCacheRemoveCTup(CatCache *cache, CatCTup *ct);
90 : static void CatCacheRemoveCList(CatCache *cache, CatCList *cl);
91 : static void RehashCatCache(CatCache *cp);
92 : static void RehashCatCacheLists(CatCache *cp);
93 : static void CatalogCacheInitializeCache(CatCache *cache);
94 : static CatCTup *CatalogCacheCreateEntry(CatCache *cache,
95 : HeapTuple ntp, SysScanDesc scandesc,
96 : Datum *arguments,
97 : uint32 hashValue, Index hashIndex);
98 :
99 : static void ReleaseCatCacheWithOwner(HeapTuple tuple, ResourceOwner resowner);
100 : static void ReleaseCatCacheListWithOwner(CatCList *list, ResourceOwner resowner);
101 : static void CatCacheFreeKeys(TupleDesc tupdesc, int nkeys, int *attnos,
102 : Datum *keys);
103 : static void CatCacheCopyKeys(TupleDesc tupdesc, int nkeys, int *attnos,
104 : Datum *srckeys, Datum *dstkeys);
105 :
106 :
107 : /*
108 : * internal support functions
109 : */
110 :
111 : /* ResourceOwner callbacks to hold catcache references */
112 :
113 : static void ResOwnerReleaseCatCache(Datum res);
114 : static char *ResOwnerPrintCatCache(Datum res);
115 : static void ResOwnerReleaseCatCacheList(Datum res);
116 : static char *ResOwnerPrintCatCacheList(Datum res);
117 :
118 : static const ResourceOwnerDesc catcache_resowner_desc =
119 : {
120 : /* catcache references */
121 : .name = "catcache reference",
122 : .release_phase = RESOURCE_RELEASE_AFTER_LOCKS,
123 : .release_priority = RELEASE_PRIO_CATCACHE_REFS,
124 : .ReleaseResource = ResOwnerReleaseCatCache,
125 : .DebugPrint = ResOwnerPrintCatCache
126 : };
127 :
128 : static const ResourceOwnerDesc catlistref_resowner_desc =
129 : {
130 : /* catcache-list pins */
131 : .name = "catcache list reference",
132 : .release_phase = RESOURCE_RELEASE_AFTER_LOCKS,
133 : .release_priority = RELEASE_PRIO_CATCACHE_LIST_REFS,
134 : .ReleaseResource = ResOwnerReleaseCatCacheList,
135 : .DebugPrint = ResOwnerPrintCatCacheList
136 : };
137 :
138 : /* Convenience wrappers over ResourceOwnerRemember/Forget */
139 : static inline void
140 69499428 : ResourceOwnerRememberCatCacheRef(ResourceOwner owner, HeapTuple tuple)
141 : {
142 69499428 : ResourceOwnerRemember(owner, PointerGetDatum(tuple), &catcache_resowner_desc);
143 69499428 : }
144 : static inline void
145 69489338 : ResourceOwnerForgetCatCacheRef(ResourceOwner owner, HeapTuple tuple)
146 : {
147 69489338 : ResourceOwnerForget(owner, PointerGetDatum(tuple), &catcache_resowner_desc);
148 69489338 : }
149 : static inline void
150 3061002 : ResourceOwnerRememberCatCacheListRef(ResourceOwner owner, CatCList *list)
151 : {
152 3061002 : ResourceOwnerRemember(owner, PointerGetDatum(list), &catlistref_resowner_desc);
153 3061002 : }
154 : static inline void
155 3060966 : ResourceOwnerForgetCatCacheListRef(ResourceOwner owner, CatCList *list)
156 : {
157 3060966 : ResourceOwnerForget(owner, PointerGetDatum(list), &catlistref_resowner_desc);
158 3060966 : }
159 :
160 :
161 : /*
162 : * Hash and equality functions for system types that are used as cache key
163 : * fields. In some cases, we just call the regular SQL-callable functions for
164 : * the appropriate data type, but that tends to be a little slow, and the
165 : * speed of these functions is performance-critical. Therefore, for data
166 : * types that frequently occur as catcache keys, we hard-code the logic here.
167 : * Avoiding the overhead of DirectFunctionCallN(...) is a substantial win, and
168 : * in certain cases (like int4) we can adopt a faster hash algorithm as well.
169 : */
170 :
171 : static bool
172 4561910 : chareqfast(Datum a, Datum b)
173 : {
174 4561910 : return DatumGetChar(a) == DatumGetChar(b);
175 : }
176 :
177 : static uint32
178 5214710 : charhashfast(Datum datum)
179 : {
180 5214710 : return murmurhash32((int32) DatumGetChar(datum));
181 : }
182 :
183 : static bool
184 3379454 : nameeqfast(Datum a, Datum b)
185 : {
186 3379454 : char *ca = NameStr(*DatumGetName(a));
187 3379454 : char *cb = NameStr(*DatumGetName(b));
188 :
189 3379454 : return strncmp(ca, cb, NAMEDATALEN) == 0;
190 : }
191 :
192 : static uint32
193 7510158 : namehashfast(Datum datum)
194 : {
195 7510158 : char *key = NameStr(*DatumGetName(datum));
196 :
197 7510158 : return hash_any((unsigned char *) key, strlen(key));
198 : }
199 :
200 : static bool
201 7061540 : int2eqfast(Datum a, Datum b)
202 : {
203 7061540 : return DatumGetInt16(a) == DatumGetInt16(b);
204 : }
205 :
206 : static uint32
207 9617722 : int2hashfast(Datum datum)
208 : {
209 9617722 : return murmurhash32((int32) DatumGetInt16(datum));
210 : }
211 :
212 : static bool
213 81859226 : int4eqfast(Datum a, Datum b)
214 : {
215 81859226 : return DatumGetInt32(a) == DatumGetInt32(b);
216 : }
217 :
218 : static uint32
219 95000170 : int4hashfast(Datum datum)
220 : {
221 95000170 : return murmurhash32((int32) DatumGetInt32(datum));
222 : }
223 :
224 : static bool
225 166 : texteqfast(Datum a, Datum b)
226 : {
227 : /*
228 : * The use of DEFAULT_COLLATION_OID is fairly arbitrary here. We just
229 : * want to take the fast "deterministic" path in texteq().
230 : */
231 166 : return DatumGetBool(DirectFunctionCall2Coll(texteq, DEFAULT_COLLATION_OID, a, b));
232 : }
233 :
234 : static uint32
235 3306 : texthashfast(Datum datum)
236 : {
237 : /* analogously here as in texteqfast() */
238 3306 : return DatumGetInt32(DirectFunctionCall1Coll(hashtext, DEFAULT_COLLATION_OID, datum));
239 : }
240 :
241 : static bool
242 2526 : oidvectoreqfast(Datum a, Datum b)
243 : {
244 2526 : return DatumGetBool(DirectFunctionCall2(oidvectoreq, a, b));
245 : }
246 :
247 : static uint32
248 324220 : oidvectorhashfast(Datum datum)
249 : {
250 324220 : return DatumGetInt32(DirectFunctionCall1(hashoidvector, datum));
251 : }
252 :
253 : /* Lookup support functions for a type. */
254 : static void
255 883966 : GetCCHashEqFuncs(Oid keytype, CCHashFN *hashfunc, RegProcedure *eqfunc, CCFastEqualFN *fasteqfunc)
256 : {
257 883966 : switch (keytype)
258 : {
259 12060 : case BOOLOID:
260 12060 : *hashfunc = charhashfast;
261 12060 : *fasteqfunc = chareqfast;
262 12060 : *eqfunc = F_BOOLEQ;
263 12060 : break;
264 16266 : case CHAROID:
265 16266 : *hashfunc = charhashfast;
266 16266 : *fasteqfunc = chareqfast;
267 16266 : *eqfunc = F_CHAREQ;
268 16266 : break;
269 166274 : case NAMEOID:
270 166274 : *hashfunc = namehashfast;
271 166274 : *fasteqfunc = nameeqfast;
272 166274 : *eqfunc = F_NAMEEQ;
273 166274 : break;
274 53334 : case INT2OID:
275 53334 : *hashfunc = int2hashfast;
276 53334 : *fasteqfunc = int2eqfast;
277 53334 : *eqfunc = F_INT2EQ;
278 53334 : break;
279 11698 : case INT4OID:
280 11698 : *hashfunc = int4hashfast;
281 11698 : *fasteqfunc = int4eqfast;
282 11698 : *eqfunc = F_INT4EQ;
283 11698 : break;
284 5324 : case TEXTOID:
285 5324 : *hashfunc = texthashfast;
286 5324 : *fasteqfunc = texteqfast;
287 5324 : *eqfunc = F_TEXTEQ;
288 5324 : break;
289 606136 : case OIDOID:
290 : case REGPROCOID:
291 : case REGPROCEDUREOID:
292 : case REGOPEROID:
293 : case REGOPERATOROID:
294 : case REGCLASSOID:
295 : case REGTYPEOID:
296 : case REGCOLLATIONOID:
297 : case REGCONFIGOID:
298 : case REGDICTIONARYOID:
299 : case REGROLEOID:
300 : case REGNAMESPACEOID:
301 606136 : *hashfunc = int4hashfast;
302 606136 : *fasteqfunc = int4eqfast;
303 606136 : *eqfunc = F_OIDEQ;
304 606136 : break;
305 12874 : case OIDVECTOROID:
306 12874 : *hashfunc = oidvectorhashfast;
307 12874 : *fasteqfunc = oidvectoreqfast;
308 12874 : *eqfunc = F_OIDVECTOREQ;
309 12874 : break;
310 0 : default:
311 0 : elog(FATAL, "type %u not supported as catcache key", keytype);
312 : *hashfunc = NULL; /* keep compiler quiet */
313 :
314 : *eqfunc = InvalidOid;
315 : break;
316 : }
317 883966 : }
318 :
319 : /*
320 : * CatalogCacheComputeHashValue
321 : *
322 : * Compute the hash value associated with a given set of lookup keys
323 : */
324 : static uint32
325 83600356 : CatalogCacheComputeHashValue(CatCache *cache, int nkeys,
326 : Datum v1, Datum v2, Datum v3, Datum v4)
327 : {
328 83600356 : uint32 hashValue = 0;
329 : uint32 oneHash;
330 83600356 : CCHashFN *cc_hashfunc = cache->cc_hashfunc;
331 :
332 : CACHE_elog(DEBUG2, "CatalogCacheComputeHashValue %s %d %p",
333 : cache->cc_relname, nkeys, cache);
334 :
335 83600356 : switch (nkeys)
336 : {
337 3922094 : case 4:
338 3922094 : oneHash = (cc_hashfunc[3]) (v4);
339 3922094 : hashValue ^= pg_rotate_left32(oneHash, 24);
340 : /* FALLTHROUGH */
341 9814576 : case 3:
342 9814576 : oneHash = (cc_hashfunc[2]) (v3);
343 9814576 : hashValue ^= pg_rotate_left32(oneHash, 16);
344 : /* FALLTHROUGH */
345 20333260 : case 2:
346 20333260 : oneHash = (cc_hashfunc[1]) (v2);
347 20333260 : hashValue ^= pg_rotate_left32(oneHash, 8);
348 : /* FALLTHROUGH */
349 83600356 : case 1:
350 83600356 : oneHash = (cc_hashfunc[0]) (v1);
351 83600356 : hashValue ^= oneHash;
352 83600356 : break;
353 0 : default:
354 0 : elog(FATAL, "wrong number of hash keys: %d", nkeys);
355 : break;
356 : }
357 :
358 83600356 : return hashValue;
359 : }
360 :
361 : /*
362 : * CatalogCacheComputeTupleHashValue
363 : *
364 : * Compute the hash value associated with a given tuple to be cached
365 : */
366 : static uint32
367 5850528 : CatalogCacheComputeTupleHashValue(CatCache *cache, int nkeys, HeapTuple tuple)
368 : {
369 5850528 : Datum v1 = 0,
370 5850528 : v2 = 0,
371 5850528 : v3 = 0,
372 5850528 : v4 = 0;
373 5850528 : bool isNull = false;
374 5850528 : int *cc_keyno = cache->cc_keyno;
375 5850528 : TupleDesc cc_tupdesc = cache->cc_tupdesc;
376 :
377 : /* Now extract key fields from tuple, insert into scankey */
378 5850528 : switch (nkeys)
379 : {
380 399344 : case 4:
381 399344 : v4 = fastgetattr(tuple,
382 399344 : cc_keyno[3],
383 : cc_tupdesc,
384 : &isNull);
385 : Assert(!isNull);
386 : /* FALLTHROUGH */
387 1077076 : case 3:
388 1077076 : v3 = fastgetattr(tuple,
389 1077076 : cc_keyno[2],
390 : cc_tupdesc,
391 : &isNull);
392 : Assert(!isNull);
393 : /* FALLTHROUGH */
394 4418242 : case 2:
395 4418242 : v2 = fastgetattr(tuple,
396 4418242 : cc_keyno[1],
397 : cc_tupdesc,
398 : &isNull);
399 : Assert(!isNull);
400 : /* FALLTHROUGH */
401 5850528 : case 1:
402 5850528 : v1 = fastgetattr(tuple,
403 : cc_keyno[0],
404 : cc_tupdesc,
405 : &isNull);
406 : Assert(!isNull);
407 5850528 : break;
408 0 : default:
409 0 : elog(FATAL, "wrong number of hash keys: %d", nkeys);
410 : break;
411 : }
412 :
413 5850528 : return CatalogCacheComputeHashValue(cache, nkeys, v1, v2, v3, v4);
414 : }
415 :
416 : /*
417 : * CatalogCacheCompareTuple
418 : *
419 : * Compare a tuple to the passed arguments.
420 : */
421 : static inline bool
422 72587594 : CatalogCacheCompareTuple(const CatCache *cache, int nkeys,
423 : const Datum *cachekeys,
424 : const Datum *searchkeys)
425 : {
426 72587594 : const CCFastEqualFN *cc_fastequal = cache->cc_fastequal;
427 : int i;
428 :
429 169452416 : for (i = 0; i < nkeys; i++)
430 : {
431 96864822 : if (!(cc_fastequal[i]) (cachekeys[i], searchkeys[i]))
432 0 : return false;
433 : }
434 72587594 : return true;
435 : }
436 :
437 :
438 : #ifdef CATCACHE_STATS
439 :
440 : static void
441 : CatCachePrintStats(int code, Datum arg)
442 : {
443 : slist_iter iter;
444 : long cc_searches = 0;
445 : long cc_hits = 0;
446 : long cc_neg_hits = 0;
447 : long cc_newloads = 0;
448 : long cc_invals = 0;
449 : long cc_nlists = 0;
450 : long cc_lsearches = 0;
451 : long cc_lhits = 0;
452 :
453 : slist_foreach(iter, &CacheHdr->ch_caches)
454 : {
455 : CatCache *cache = slist_container(CatCache, cc_next, iter.cur);
456 :
457 : if (cache->cc_ntup == 0 && cache->cc_searches == 0)
458 : continue; /* don't print unused caches */
459 : elog(DEBUG2, "catcache %s/%u: %d tup, %ld srch, %ld+%ld=%ld hits, %ld+%ld=%ld loads, %ld invals, %d lists, %ld lsrch, %ld lhits",
460 : cache->cc_relname,
461 : cache->cc_indexoid,
462 : cache->cc_ntup,
463 : cache->cc_searches,
464 : cache->cc_hits,
465 : cache->cc_neg_hits,
466 : cache->cc_hits + cache->cc_neg_hits,
467 : cache->cc_newloads,
468 : cache->cc_searches - cache->cc_hits - cache->cc_neg_hits - cache->cc_newloads,
469 : cache->cc_searches - cache->cc_hits - cache->cc_neg_hits,
470 : cache->cc_invals,
471 : cache->cc_nlist,
472 : cache->cc_lsearches,
473 : cache->cc_lhits);
474 : cc_searches += cache->cc_searches;
475 : cc_hits += cache->cc_hits;
476 : cc_neg_hits += cache->cc_neg_hits;
477 : cc_newloads += cache->cc_newloads;
478 : cc_invals += cache->cc_invals;
479 : cc_nlists += cache->cc_nlist;
480 : cc_lsearches += cache->cc_lsearches;
481 : cc_lhits += cache->cc_lhits;
482 : }
483 : elog(DEBUG2, "catcache totals: %d tup, %ld srch, %ld+%ld=%ld hits, %ld+%ld=%ld loads, %ld invals, %ld lists, %ld lsrch, %ld lhits",
484 : CacheHdr->ch_ntup,
485 : cc_searches,
486 : cc_hits,
487 : cc_neg_hits,
488 : cc_hits + cc_neg_hits,
489 : cc_newloads,
490 : cc_searches - cc_hits - cc_neg_hits - cc_newloads,
491 : cc_searches - cc_hits - cc_neg_hits,
492 : cc_invals,
493 : cc_nlists,
494 : cc_lsearches,
495 : cc_lhits);
496 : }
497 : #endif /* CATCACHE_STATS */
498 :
499 :
500 : /*
501 : * CatCacheRemoveCTup
502 : *
503 : * Unlink and delete the given cache entry
504 : *
505 : * NB: if it is a member of a CatCList, the CatCList is deleted too.
506 : * Both the cache entry and the list had better have zero refcount.
507 : */
508 : static void
509 1417782 : CatCacheRemoveCTup(CatCache *cache, CatCTup *ct)
510 : {
511 : Assert(ct->refcount == 0);
512 : Assert(ct->my_cache == cache);
513 :
514 1417782 : if (ct->c_list)
515 : {
516 : /*
517 : * The cleanest way to handle this is to call CatCacheRemoveCList,
518 : * which will recurse back to me, and the recursive call will do the
519 : * work. Set the "dead" flag to make sure it does recurse.
520 : */
521 0 : ct->dead = true;
522 0 : CatCacheRemoveCList(cache, ct->c_list);
523 0 : return; /* nothing left to do */
524 : }
525 :
526 : /* delink from linked list */
527 1417782 : dlist_delete(&ct->cache_elem);
528 :
529 : /*
530 : * Free keys when we're dealing with a negative entry, normal entries just
531 : * point into tuple, allocated together with the CatCTup.
532 : */
533 1417782 : if (ct->negative)
534 390214 : CatCacheFreeKeys(cache->cc_tupdesc, cache->cc_nkeys,
535 390214 : cache->cc_keyno, ct->keys);
536 :
537 1417782 : pfree(ct);
538 :
539 1417782 : --cache->cc_ntup;
540 1417782 : --CacheHdr->ch_ntup;
541 : }
542 :
543 : /*
544 : * CatCacheRemoveCList
545 : *
546 : * Unlink and delete the given cache list entry
547 : *
548 : * NB: any dead member entries that become unreferenced are deleted too.
549 : */
550 : static void
551 112118 : CatCacheRemoveCList(CatCache *cache, CatCList *cl)
552 : {
553 : int i;
554 :
555 : Assert(cl->refcount == 0);
556 : Assert(cl->my_cache == cache);
557 :
558 : /* delink from member tuples */
559 374728 : for (i = cl->n_members; --i >= 0;)
560 : {
561 262610 : CatCTup *ct = cl->members[i];
562 :
563 : Assert(ct->c_list == cl);
564 262610 : ct->c_list = NULL;
565 : /* if the member is dead and now has no references, remove it */
566 262610 : if (
567 : #ifndef CATCACHE_FORCE_RELEASE
568 262610 : ct->dead &&
569 : #endif
570 144 : ct->refcount == 0)
571 144 : CatCacheRemoveCTup(cache, ct);
572 : }
573 :
574 : /* delink from linked list */
575 112118 : dlist_delete(&cl->cache_elem);
576 :
577 : /* free associated column data */
578 112118 : CatCacheFreeKeys(cache->cc_tupdesc, cl->nkeys,
579 112118 : cache->cc_keyno, cl->keys);
580 :
581 112118 : pfree(cl);
582 :
583 112118 : --cache->cc_nlist;
584 112118 : }
585 :
586 :
587 : /*
588 : * CatCacheInvalidate
589 : *
590 : * Invalidate entries in the specified cache, given a hash value.
591 : *
592 : * We delete cache entries that match the hash value, whether positive
593 : * or negative. We don't care whether the invalidation is the result
594 : * of a tuple insertion or a deletion.
595 : *
596 : * We used to try to match positive cache entries by TID, but that is
597 : * unsafe after a VACUUM FULL on a system catalog: an inval event could
598 : * be queued before VACUUM FULL, and then processed afterwards, when the
599 : * target tuple that has to be invalidated has a different TID than it
600 : * did when the event was created. So now we just compare hash values and
601 : * accept the small risk of unnecessary invalidations due to false matches.
602 : *
603 : * This routine is only quasi-public: it should only be used by inval.c.
604 : */
605 : void
606 18264032 : CatCacheInvalidate(CatCache *cache, uint32 hashValue)
607 : {
608 : Index hashIndex;
609 : dlist_mutable_iter iter;
610 :
611 : CACHE_elog(DEBUG2, "CatCacheInvalidate: called");
612 :
613 : /*
614 : * We don't bother to check whether the cache has finished initialization
615 : * yet; if not, there will be no entries in it so no problem.
616 : */
617 :
618 : /*
619 : * Invalidate *all* CatCLists in this cache; it's too hard to tell which
620 : * searches might still be correct, so just zap 'em all.
621 : */
622 21582816 : for (int i = 0; i < cache->cc_nlbuckets; i++)
623 : {
624 3318784 : dlist_head *bucket = &cache->cc_lbucket[i];
625 :
626 3425394 : dlist_foreach_modify(iter, bucket)
627 : {
628 106610 : CatCList *cl = dlist_container(CatCList, cache_elem, iter.cur);
629 :
630 106610 : if (cl->refcount > 0)
631 144 : cl->dead = true;
632 : else
633 106466 : CatCacheRemoveCList(cache, cl);
634 : }
635 : }
636 :
637 : /*
638 : * inspect the proper hash bucket for tuple matches
639 : */
640 18264032 : hashIndex = HASH_INDEX(hashValue, cache->cc_nbuckets);
641 24829976 : dlist_foreach_modify(iter, &cache->cc_bucket[hashIndex])
642 : {
643 6565944 : CatCTup *ct = dlist_container(CatCTup, cache_elem, iter.cur);
644 :
645 6565944 : if (hashValue == ct->hash_value)
646 : {
647 1211454 : if (ct->refcount > 0 ||
648 1196414 : (ct->c_list && ct->c_list->refcount > 0))
649 : {
650 15184 : ct->dead = true;
651 : /* list, if any, was marked dead above */
652 15184 : Assert(ct->c_list == NULL || ct->c_list->dead);
653 : }
654 : else
655 1196270 : CatCacheRemoveCTup(cache, ct);
656 : CACHE_elog(DEBUG2, "CatCacheInvalidate: invalidated");
657 : #ifdef CATCACHE_STATS
658 : cache->cc_invals++;
659 : #endif
660 : /* could be multiple matches, so keep looking! */
661 : }
662 : }
663 18264032 : }
664 :
665 : /* ----------------------------------------------------------------
666 : * public functions
667 : * ----------------------------------------------------------------
668 : */
669 :
670 :
671 : /*
672 : * Standard routine for creating cache context if it doesn't exist yet
673 : *
674 : * There are a lot of places (probably far more than necessary) that check
675 : * whether CacheMemoryContext exists yet and want to create it if not.
676 : * We centralize knowledge of exactly how to create it here.
677 : */
678 : void
679 25896 : CreateCacheMemoryContext(void)
680 : {
681 : /*
682 : * Purely for paranoia, check that context doesn't exist; caller probably
683 : * did so already.
684 : */
685 25896 : if (!CacheMemoryContext)
686 25896 : CacheMemoryContext = AllocSetContextCreate(TopMemoryContext,
687 : "CacheMemoryContext",
688 : ALLOCSET_DEFAULT_SIZES);
689 25896 : }
690 :
691 :
692 : /*
693 : * ResetCatalogCache
694 : *
695 : * Reset one catalog cache to empty.
696 : *
697 : * This is not very efficient if the target cache is nearly empty.
698 : * However, it shouldn't need to be efficient; we don't invoke it often.
699 : */
700 : static void
701 335996 : ResetCatalogCache(CatCache *cache)
702 : {
703 : dlist_mutable_iter iter;
704 : int i;
705 :
706 : /* Remove each list in this cache, or at least mark it dead */
707 379068 : for (i = 0; i < cache->cc_nlbuckets; i++)
708 : {
709 43072 : dlist_head *bucket = &cache->cc_lbucket[i];
710 :
711 48718 : dlist_foreach_modify(iter, bucket)
712 : {
713 5646 : CatCList *cl = dlist_container(CatCList, cache_elem, iter.cur);
714 :
715 5646 : if (cl->refcount > 0)
716 0 : cl->dead = true;
717 : else
718 5646 : CatCacheRemoveCList(cache, cl);
719 : }
720 : }
721 :
722 : /* Remove each tuple in this cache, or at least mark it dead */
723 10260544 : for (i = 0; i < cache->cc_nbuckets; i++)
724 : {
725 9924548 : dlist_head *bucket = &cache->cc_bucket[i];
726 :
727 10130882 : dlist_foreach_modify(iter, bucket)
728 : {
729 206334 : CatCTup *ct = dlist_container(CatCTup, cache_elem, iter.cur);
730 :
731 206334 : if (ct->refcount > 0 ||
732 206328 : (ct->c_list && ct->c_list->refcount > 0))
733 : {
734 6 : ct->dead = true;
735 : /* list, if any, was marked dead above */
736 6 : Assert(ct->c_list == NULL || ct->c_list->dead);
737 : }
738 : else
739 206328 : CatCacheRemoveCTup(cache, ct);
740 : #ifdef CATCACHE_STATS
741 : cache->cc_invals++;
742 : #endif
743 : }
744 : }
745 335996 : }
746 :
747 : /*
748 : * ResetCatalogCaches
749 : *
750 : * Reset all caches when a shared cache inval event forces it
751 : */
752 : void
753 4038 : ResetCatalogCaches(void)
754 : {
755 : slist_iter iter;
756 :
757 : CACHE_elog(DEBUG2, "ResetCatalogCaches called");
758 :
759 339192 : slist_foreach(iter, &CacheHdr->ch_caches)
760 : {
761 335154 : CatCache *cache = slist_container(CatCache, cc_next, iter.cur);
762 :
763 335154 : ResetCatalogCache(cache);
764 : }
765 :
766 : CACHE_elog(DEBUG2, "end of ResetCatalogCaches call");
767 4038 : }
768 :
769 : /*
770 : * CatalogCacheFlushCatalog
771 : *
772 : * Flush all catcache entries that came from the specified system catalog.
773 : * This is needed after VACUUM FULL/CLUSTER on the catalog, since the
774 : * tuples very likely now have different TIDs than before. (At one point
775 : * we also tried to force re-execution of CatalogCacheInitializeCache for
776 : * the cache(s) on that catalog. This is a bad idea since it leads to all
777 : * kinds of trouble if a cache flush occurs while loading cache entries.
778 : * We now avoid the need to do it by copying cc_tupdesc out of the relcache,
779 : * rather than relying on the relcache to keep a tupdesc for us. Of course
780 : * this assumes the tupdesc of a cachable system table will not change...)
781 : */
782 : void
783 674 : CatalogCacheFlushCatalog(Oid catId)
784 : {
785 : slist_iter iter;
786 :
787 : CACHE_elog(DEBUG2, "CatalogCacheFlushCatalog called for %u", catId);
788 :
789 56616 : slist_foreach(iter, &CacheHdr->ch_caches)
790 : {
791 55942 : CatCache *cache = slist_container(CatCache, cc_next, iter.cur);
792 :
793 : /* Does this cache store tuples of the target catalog? */
794 55942 : if (cache->cc_reloid == catId)
795 : {
796 : /* Yes, so flush all its contents */
797 842 : ResetCatalogCache(cache);
798 :
799 : /* Tell inval.c to call syscache callbacks for this cache */
800 842 : CallSyscacheCallbacks(cache->id, 0);
801 : }
802 : }
803 :
804 : CACHE_elog(DEBUG2, "end of CatalogCacheFlushCatalog call");
805 674 : }
806 :
807 : /*
808 : * InitCatCache
809 : *
810 : * This allocates and initializes a cache for a system catalog relation.
811 : * Actually, the cache is only partially initialized to avoid opening the
812 : * relation. The relation will be opened and the rest of the cache
813 : * structure initialized on the first access.
814 : */
815 : #ifdef CACHEDEBUG
816 : #define InitCatCache_DEBUG2 \
817 : do { \
818 : elog(DEBUG2, "InitCatCache: rel=%u ind=%u id=%d nkeys=%d size=%d", \
819 : cp->cc_reloid, cp->cc_indexoid, cp->id, \
820 : cp->cc_nkeys, cp->cc_nbuckets); \
821 : } while(0)
822 : #else
823 : #define InitCatCache_DEBUG2
824 : #endif
825 :
826 : CatCache *
827 2149368 : InitCatCache(int id,
828 : Oid reloid,
829 : Oid indexoid,
830 : int nkeys,
831 : const int *key,
832 : int nbuckets)
833 : {
834 : CatCache *cp;
835 : MemoryContext oldcxt;
836 : int i;
837 :
838 : /*
839 : * nbuckets is the initial number of hash buckets to use in this catcache.
840 : * It will be enlarged later if it becomes too full.
841 : *
842 : * nbuckets must be a power of two. We check this via Assert rather than
843 : * a full runtime check because the values will be coming from constant
844 : * tables.
845 : *
846 : * If you're confused by the power-of-two check, see comments in
847 : * bitmapset.c for an explanation.
848 : */
849 : Assert(nbuckets > 0 && (nbuckets & -nbuckets) == nbuckets);
850 :
851 : /*
852 : * first switch to the cache context so our allocations do not vanish at
853 : * the end of a transaction
854 : */
855 2149368 : if (!CacheMemoryContext)
856 0 : CreateCacheMemoryContext();
857 :
858 2149368 : oldcxt = MemoryContextSwitchTo(CacheMemoryContext);
859 :
860 : /*
861 : * if first time through, initialize the cache group header
862 : */
863 2149368 : if (CacheHdr == NULL)
864 : {
865 25896 : CacheHdr = (CatCacheHeader *) palloc(sizeof(CatCacheHeader));
866 25896 : slist_init(&CacheHdr->ch_caches);
867 25896 : CacheHdr->ch_ntup = 0;
868 : #ifdef CATCACHE_STATS
869 : /* set up to dump stats at backend exit */
870 : on_proc_exit(CatCachePrintStats, 0);
871 : #endif
872 : }
873 :
874 : /*
875 : * Allocate a new cache structure, aligning to a cacheline boundary
876 : *
877 : * Note: we rely on zeroing to initialize all the dlist headers correctly
878 : */
879 2149368 : cp = (CatCache *) palloc_aligned(sizeof(CatCache), PG_CACHE_LINE_SIZE,
880 : MCXT_ALLOC_ZERO);
881 2149368 : cp->cc_bucket = palloc0(nbuckets * sizeof(dlist_head));
882 :
883 : /*
884 : * Many catcaches never receive any list searches. Therefore, we don't
885 : * allocate the cc_lbuckets till we get a list search.
886 : */
887 2149368 : cp->cc_lbucket = NULL;
888 :
889 : /*
890 : * initialize the cache's relation information for the relation
891 : * corresponding to this cache, and initialize some of the new cache's
892 : * other internal fields. But don't open the relation yet.
893 : */
894 2149368 : cp->id = id;
895 2149368 : cp->cc_relname = "(not known yet)";
896 2149368 : cp->cc_reloid = reloid;
897 2149368 : cp->cc_indexoid = indexoid;
898 2149368 : cp->cc_relisshared = false; /* temporary */
899 2149368 : cp->cc_tupdesc = (TupleDesc) NULL;
900 2149368 : cp->cc_ntup = 0;
901 2149368 : cp->cc_nlist = 0;
902 2149368 : cp->cc_nbuckets = nbuckets;
903 2149368 : cp->cc_nlbuckets = 0;
904 2149368 : cp->cc_nkeys = nkeys;
905 5645328 : for (i = 0; i < nkeys; ++i)
906 : {
907 : Assert(AttributeNumberIsValid(key[i]));
908 3495960 : cp->cc_keyno[i] = key[i];
909 : }
910 :
911 : /*
912 : * new cache is initialized as far as we can go for now. print some
913 : * debugging information, if appropriate.
914 : */
915 : InitCatCache_DEBUG2;
916 :
917 : /*
918 : * add completed cache to top of group header's list
919 : */
920 2149368 : slist_push_head(&CacheHdr->ch_caches, &cp->cc_next);
921 :
922 : /*
923 : * back to the old context before we return...
924 : */
925 2149368 : MemoryContextSwitchTo(oldcxt);
926 :
927 2149368 : return cp;
928 : }
929 :
930 : /*
931 : * Enlarge a catcache, doubling the number of buckets.
932 : */
933 : static void
934 5016 : RehashCatCache(CatCache *cp)
935 : {
936 : dlist_head *newbucket;
937 : int newnbuckets;
938 : int i;
939 :
940 5016 : elog(DEBUG1, "rehashing catalog cache id %d for %s; %d tups, %d buckets",
941 : cp->id, cp->cc_relname, cp->cc_ntup, cp->cc_nbuckets);
942 :
943 : /* Allocate a new, larger, hash table. */
944 5016 : newnbuckets = cp->cc_nbuckets * 2;
945 5016 : newbucket = (dlist_head *) MemoryContextAllocZero(CacheMemoryContext, newnbuckets * sizeof(dlist_head));
946 :
947 : /* Move all entries from old hash table to new. */
948 420792 : for (i = 0; i < cp->cc_nbuckets; i++)
949 : {
950 : dlist_mutable_iter iter;
951 :
952 1252344 : dlist_foreach_modify(iter, &cp->cc_bucket[i])
953 : {
954 836568 : CatCTup *ct = dlist_container(CatCTup, cache_elem, iter.cur);
955 836568 : int hashIndex = HASH_INDEX(ct->hash_value, newnbuckets);
956 :
957 836568 : dlist_delete(iter.cur);
958 836568 : dlist_push_head(&newbucket[hashIndex], &ct->cache_elem);
959 : }
960 : }
961 :
962 : /* Switch to the new array. */
963 5016 : pfree(cp->cc_bucket);
964 5016 : cp->cc_nbuckets = newnbuckets;
965 5016 : cp->cc_bucket = newbucket;
966 5016 : }
967 :
968 : /*
969 : * Enlarge a catcache's list storage, doubling the number of buckets.
970 : */
971 : static void
972 1036 : RehashCatCacheLists(CatCache *cp)
973 : {
974 : dlist_head *newbucket;
975 : int newnbuckets;
976 : int i;
977 :
978 1036 : elog(DEBUG1, "rehashing catalog cache id %d for %s; %d lists, %d buckets",
979 : cp->id, cp->cc_relname, cp->cc_nlist, cp->cc_nlbuckets);
980 :
981 : /* Allocate a new, larger, hash table. */
982 1036 : newnbuckets = cp->cc_nlbuckets * 2;
983 1036 : newbucket = (dlist_head *) MemoryContextAllocZero(CacheMemoryContext, newnbuckets * sizeof(dlist_head));
984 :
985 : /* Move all entries from old hash table to new. */
986 31916 : for (i = 0; i < cp->cc_nlbuckets; i++)
987 : {
988 : dlist_mutable_iter iter;
989 :
990 93676 : dlist_foreach_modify(iter, &cp->cc_lbucket[i])
991 : {
992 62796 : CatCList *cl = dlist_container(CatCList, cache_elem, iter.cur);
993 62796 : int hashIndex = HASH_INDEX(cl->hash_value, newnbuckets);
994 :
995 62796 : dlist_delete(iter.cur);
996 62796 : dlist_push_head(&newbucket[hashIndex], &cl->cache_elem);
997 : }
998 : }
999 :
1000 : /* Switch to the new array. */
1001 1036 : pfree(cp->cc_lbucket);
1002 1036 : cp->cc_nlbuckets = newnbuckets;
1003 1036 : cp->cc_lbucket = newbucket;
1004 1036 : }
1005 :
1006 : /*
1007 : * CatalogCacheInitializeCache
1008 : *
1009 : * This function does final initialization of a catcache: obtain the tuple
1010 : * descriptor and set up the hash and equality function links. We assume
1011 : * that the relcache entry can be opened at this point!
1012 : */
1013 : #ifdef CACHEDEBUG
1014 : #define CatalogCacheInitializeCache_DEBUG1 \
1015 : elog(DEBUG2, "CatalogCacheInitializeCache: cache @%p rel=%u", cache, \
1016 : cache->cc_reloid)
1017 :
1018 : #define CatalogCacheInitializeCache_DEBUG2 \
1019 : do { \
1020 : if (cache->cc_keyno[i] > 0) { \
1021 : elog(DEBUG2, "CatalogCacheInitializeCache: load %d/%d w/%d, %u", \
1022 : i+1, cache->cc_nkeys, cache->cc_keyno[i], \
1023 : TupleDescAttr(tupdesc, cache->cc_keyno[i] - 1)->atttypid); \
1024 : } else { \
1025 : elog(DEBUG2, "CatalogCacheInitializeCache: load %d/%d w/%d", \
1026 : i+1, cache->cc_nkeys, cache->cc_keyno[i]); \
1027 : } \
1028 : } while(0)
1029 : #else
1030 : #define CatalogCacheInitializeCache_DEBUG1
1031 : #define CatalogCacheInitializeCache_DEBUG2
1032 : #endif
1033 :
1034 : static void
1035 554388 : CatalogCacheInitializeCache(CatCache *cache)
1036 : {
1037 : Relation relation;
1038 : MemoryContext oldcxt;
1039 : TupleDesc tupdesc;
1040 : int i;
1041 :
1042 : CatalogCacheInitializeCache_DEBUG1;
1043 :
1044 554388 : relation = table_open(cache->cc_reloid, AccessShareLock);
1045 :
1046 : /*
1047 : * switch to the cache context so our allocations do not vanish at the end
1048 : * of a transaction
1049 : */
1050 : Assert(CacheMemoryContext != NULL);
1051 :
1052 554384 : oldcxt = MemoryContextSwitchTo(CacheMemoryContext);
1053 :
1054 : /*
1055 : * copy the relcache's tuple descriptor to permanent cache storage
1056 : */
1057 554384 : tupdesc = CreateTupleDescCopyConstr(RelationGetDescr(relation));
1058 :
1059 : /*
1060 : * save the relation's name and relisshared flag, too (cc_relname is used
1061 : * only for debugging purposes)
1062 : */
1063 554384 : cache->cc_relname = pstrdup(RelationGetRelationName(relation));
1064 554384 : cache->cc_relisshared = RelationGetForm(relation)->relisshared;
1065 :
1066 : /*
1067 : * return to the caller's memory context and close the rel
1068 : */
1069 554384 : MemoryContextSwitchTo(oldcxt);
1070 :
1071 554384 : table_close(relation, AccessShareLock);
1072 :
1073 : CACHE_elog(DEBUG2, "CatalogCacheInitializeCache: %s, %d keys",
1074 : cache->cc_relname, cache->cc_nkeys);
1075 :
1076 : /*
1077 : * initialize cache's key information
1078 : */
1079 1438350 : for (i = 0; i < cache->cc_nkeys; ++i)
1080 : {
1081 : Oid keytype;
1082 : RegProcedure eqfunc;
1083 :
1084 : CatalogCacheInitializeCache_DEBUG2;
1085 :
1086 883966 : if (cache->cc_keyno[i] > 0)
1087 : {
1088 883966 : Form_pg_attribute attr = TupleDescAttr(tupdesc,
1089 : cache->cc_keyno[i] - 1);
1090 :
1091 883966 : keytype = attr->atttypid;
1092 : /* cache key columns should always be NOT NULL */
1093 : Assert(attr->attnotnull);
1094 : }
1095 : else
1096 : {
1097 0 : if (cache->cc_keyno[i] < 0)
1098 0 : elog(FATAL, "sys attributes are not supported in caches");
1099 0 : keytype = OIDOID;
1100 : }
1101 :
1102 883966 : GetCCHashEqFuncs(keytype,
1103 : &cache->cc_hashfunc[i],
1104 : &eqfunc,
1105 : &cache->cc_fastequal[i]);
1106 :
1107 : /*
1108 : * Do equality-function lookup (we assume this won't need a catalog
1109 : * lookup for any supported type)
1110 : */
1111 883966 : fmgr_info_cxt(eqfunc,
1112 : &cache->cc_skey[i].sk_func,
1113 : CacheMemoryContext);
1114 :
1115 : /* Initialize sk_attno suitably for HeapKeyTest() and heap scans */
1116 883966 : cache->cc_skey[i].sk_attno = cache->cc_keyno[i];
1117 :
1118 : /* Fill in sk_strategy as well --- always standard equality */
1119 883966 : cache->cc_skey[i].sk_strategy = BTEqualStrategyNumber;
1120 883966 : cache->cc_skey[i].sk_subtype = InvalidOid;
1121 : /* If a catcache key requires a collation, it must be C collation */
1122 883966 : cache->cc_skey[i].sk_collation = C_COLLATION_OID;
1123 :
1124 : CACHE_elog(DEBUG2, "CatalogCacheInitializeCache %s %d %p",
1125 : cache->cc_relname, i, cache);
1126 : }
1127 :
1128 : /*
1129 : * mark this cache fully initialized
1130 : */
1131 554384 : cache->cc_tupdesc = tupdesc;
1132 554384 : }
1133 :
1134 : /*
1135 : * InitCatCachePhase2 -- external interface for CatalogCacheInitializeCache
1136 : *
1137 : * One reason to call this routine is to ensure that the relcache has
1138 : * created entries for all the catalogs and indexes referenced by catcaches.
1139 : * Therefore, provide an option to open the index as well as fixing the
1140 : * cache itself. An exception is the indexes on pg_am, which we don't use
1141 : * (cf. IndexScanOK).
1142 : */
1143 : void
1144 203430 : InitCatCachePhase2(CatCache *cache, bool touch_index)
1145 : {
1146 203430 : if (cache->cc_tupdesc == NULL)
1147 188574 : CatalogCacheInitializeCache(cache);
1148 :
1149 203428 : if (touch_index &&
1150 178910 : cache->id != AMOID &&
1151 176754 : cache->id != AMNAME)
1152 : {
1153 : Relation idesc;
1154 :
1155 : /*
1156 : * We must lock the underlying catalog before opening the index to
1157 : * avoid deadlock, since index_open could possibly result in reading
1158 : * this same catalog, and if anyone else is exclusive-locking this
1159 : * catalog and index they'll be doing it in that order.
1160 : */
1161 174598 : LockRelationOid(cache->cc_reloid, AccessShareLock);
1162 174598 : idesc = index_open(cache->cc_indexoid, AccessShareLock);
1163 :
1164 : /*
1165 : * While we've got the index open, let's check that it's unique (and
1166 : * not just deferrable-unique, thank you very much). This is just to
1167 : * catch thinkos in definitions of new catcaches, so we don't worry
1168 : * about the pg_am indexes not getting tested.
1169 : */
1170 : Assert(idesc->rd_index->indisunique &&
1171 : idesc->rd_index->indimmediate);
1172 :
1173 174598 : index_close(idesc, AccessShareLock);
1174 174598 : UnlockRelationOid(cache->cc_reloid, AccessShareLock);
1175 : }
1176 203428 : }
1177 :
1178 :
1179 : /*
1180 : * IndexScanOK
1181 : *
1182 : * This function checks for tuples that will be fetched by
1183 : * IndexSupportInitialize() during relcache initialization for
1184 : * certain system indexes that support critical syscaches.
1185 : * We can't use an indexscan to fetch these, else we'll get into
1186 : * infinite recursion. A plain heap scan will work, however.
1187 : * Once we have completed relcache initialization (signaled by
1188 : * criticalRelcachesBuilt), we don't have to worry anymore.
1189 : *
1190 : * Similarly, during backend startup we have to be able to use the
1191 : * pg_authid, pg_auth_members and pg_database syscaches for
1192 : * authentication even if we don't yet have relcache entries for those
1193 : * catalogs' indexes.
1194 : */
1195 : static bool
1196 5021178 : IndexScanOK(CatCache *cache, ScanKey cur_skey)
1197 : {
1198 5021178 : switch (cache->id)
1199 : {
1200 420336 : case INDEXRELID:
1201 :
1202 : /*
1203 : * Rather than tracking exactly which indexes have to be loaded
1204 : * before we can use indexscans (which changes from time to time),
1205 : * just force all pg_index searches to be heap scans until we've
1206 : * built the critical relcaches.
1207 : */
1208 420336 : if (!criticalRelcachesBuilt)
1209 27546 : return false;
1210 392790 : break;
1211 :
1212 48254 : case AMOID:
1213 : case AMNAME:
1214 :
1215 : /*
1216 : * Always do heap scans in pg_am, because it's so small there's
1217 : * not much point in an indexscan anyway. We *must* do this when
1218 : * initially building critical relcache entries, but we might as
1219 : * well just always do it.
1220 : */
1221 48254 : return false;
1222 :
1223 89932 : case AUTHNAME:
1224 : case AUTHOID:
1225 : case AUTHMEMMEMROLE:
1226 : case DATABASEOID:
1227 :
1228 : /*
1229 : * Protect authentication lookups occurring before relcache has
1230 : * collected entries for shared indexes.
1231 : */
1232 89932 : if (!criticalSharedRelcachesBuilt)
1233 3544 : return false;
1234 86388 : break;
1235 :
1236 4462656 : default:
1237 4462656 : break;
1238 : }
1239 :
1240 : /* Normal case, allow index scan */
1241 4941834 : return true;
1242 : }
1243 :
1244 : /*
1245 : * SearchCatCache
1246 : *
1247 : * This call searches a system cache for a tuple, opening the relation
1248 : * if necessary (on the first access to a particular cache).
1249 : *
1250 : * The result is NULL if not found, or a pointer to a HeapTuple in
1251 : * the cache. The caller must not modify the tuple, and must call
1252 : * ReleaseCatCache() when done with it.
1253 : *
1254 : * The search key values should be expressed as Datums of the key columns'
1255 : * datatype(s). (Pass zeroes for any unused parameters.) As a special
1256 : * exception, the passed-in key for a NAME column can be just a C string;
1257 : * the caller need not go to the trouble of converting it to a fully
1258 : * null-padded NAME.
1259 : */
1260 : HeapTuple
1261 4751022 : SearchCatCache(CatCache *cache,
1262 : Datum v1,
1263 : Datum v2,
1264 : Datum v3,
1265 : Datum v4)
1266 : {
1267 4751022 : return SearchCatCacheInternal(cache, cache->cc_nkeys, v1, v2, v3, v4);
1268 : }
1269 :
1270 :
1271 : /*
1272 : * SearchCatCacheN() are SearchCatCache() versions for a specific number of
1273 : * arguments. The compiler can inline the body and unroll loops, making them a
1274 : * bit faster than SearchCatCache().
1275 : */
1276 :
1277 : HeapTuple
1278 57215158 : SearchCatCache1(CatCache *cache,
1279 : Datum v1)
1280 : {
1281 57215158 : return SearchCatCacheInternal(cache, 1, v1, 0, 0, 0);
1282 : }
1283 :
1284 :
1285 : HeapTuple
1286 4553358 : SearchCatCache2(CatCache *cache,
1287 : Datum v1, Datum v2)
1288 : {
1289 4553358 : return SearchCatCacheInternal(cache, 2, v1, v2, 0, 0);
1290 : }
1291 :
1292 :
1293 : HeapTuple
1294 4506382 : SearchCatCache3(CatCache *cache,
1295 : Datum v1, Datum v2, Datum v3)
1296 : {
1297 4506382 : return SearchCatCacheInternal(cache, 3, v1, v2, v3, 0);
1298 : }
1299 :
1300 :
1301 : HeapTuple
1302 3521852 : SearchCatCache4(CatCache *cache,
1303 : Datum v1, Datum v2, Datum v3, Datum v4)
1304 : {
1305 3521852 : return SearchCatCacheInternal(cache, 4, v1, v2, v3, v4);
1306 : }
1307 :
1308 : /*
1309 : * Work-horse for SearchCatCache/SearchCatCacheN.
1310 : */
1311 : static inline HeapTuple
1312 74547772 : SearchCatCacheInternal(CatCache *cache,
1313 : int nkeys,
1314 : Datum v1,
1315 : Datum v2,
1316 : Datum v3,
1317 : Datum v4)
1318 : {
1319 : Datum arguments[CATCACHE_MAXKEYS];
1320 : uint32 hashValue;
1321 : Index hashIndex;
1322 : dlist_iter iter;
1323 : dlist_head *bucket;
1324 : CatCTup *ct;
1325 :
1326 : /* Make sure we're in an xact, even if this ends up being a cache hit */
1327 : Assert(IsTransactionState());
1328 :
1329 : Assert(cache->cc_nkeys == nkeys);
1330 :
1331 : /*
1332 : * one-time startup overhead for each cache
1333 : */
1334 74547772 : if (unlikely(cache->cc_tupdesc == NULL))
1335 306418 : CatalogCacheInitializeCache(cache);
1336 :
1337 : #ifdef CATCACHE_STATS
1338 : cache->cc_searches++;
1339 : #endif
1340 :
1341 : /* Initialize local parameter array */
1342 74547770 : arguments[0] = v1;
1343 74547770 : arguments[1] = v2;
1344 74547770 : arguments[2] = v3;
1345 74547770 : arguments[3] = v4;
1346 :
1347 : /*
1348 : * find the hash bucket in which to look for the tuple
1349 : */
1350 74547770 : hashValue = CatalogCacheComputeHashValue(cache, nkeys, v1, v2, v3, v4);
1351 74547770 : hashIndex = HASH_INDEX(hashValue, cache->cc_nbuckets);
1352 :
1353 : /*
1354 : * scan the hash bucket until we find a match or exhaust our tuples
1355 : *
1356 : * Note: it's okay to use dlist_foreach here, even though we modify the
1357 : * dlist within the loop, because we don't continue the loop afterwards.
1358 : */
1359 74547770 : bucket = &cache->cc_bucket[hashIndex];
1360 79451194 : dlist_foreach(iter, bucket)
1361 : {
1362 74688928 : ct = dlist_container(CatCTup, cache_elem, iter.cur);
1363 :
1364 74688928 : if (ct->dead)
1365 2 : continue; /* ignore dead entries */
1366 :
1367 74688926 : if (ct->hash_value != hashValue)
1368 4903422 : continue; /* quickly skip entry if wrong hash val */
1369 :
1370 69785504 : if (!CatalogCacheCompareTuple(cache, nkeys, ct->keys, arguments))
1371 0 : continue;
1372 :
1373 : /*
1374 : * We found a match in the cache. Move it to the front of the list
1375 : * for its hashbucket, in order to speed subsequent searches. (The
1376 : * most frequently accessed elements in any hashbucket will tend to be
1377 : * near the front of the hashbucket's list.)
1378 : */
1379 69785504 : dlist_move_head(bucket, &ct->cache_elem);
1380 :
1381 : /*
1382 : * If it's a positive entry, bump its refcount and return it. If it's
1383 : * negative, we can report failure to the caller.
1384 : */
1385 69785504 : if (!ct->negative)
1386 : {
1387 66198960 : ResourceOwnerEnlarge(CurrentResourceOwner);
1388 66198960 : ct->refcount++;
1389 66198960 : ResourceOwnerRememberCatCacheRef(CurrentResourceOwner, &ct->tuple);
1390 :
1391 : CACHE_elog(DEBUG2, "SearchCatCache(%s): found in bucket %d",
1392 : cache->cc_relname, hashIndex);
1393 :
1394 : #ifdef CATCACHE_STATS
1395 : cache->cc_hits++;
1396 : #endif
1397 :
1398 66198960 : return &ct->tuple;
1399 : }
1400 : else
1401 : {
1402 : CACHE_elog(DEBUG2, "SearchCatCache(%s): found neg entry in bucket %d",
1403 : cache->cc_relname, hashIndex);
1404 :
1405 : #ifdef CATCACHE_STATS
1406 : cache->cc_neg_hits++;
1407 : #endif
1408 :
1409 3586544 : return NULL;
1410 : }
1411 : }
1412 :
1413 4762266 : return SearchCatCacheMiss(cache, nkeys, hashValue, hashIndex, v1, v2, v3, v4);
1414 : }
1415 :
1416 : /*
1417 : * Search the actual catalogs, rather than the cache.
1418 : *
1419 : * This is kept separate from SearchCatCacheInternal() to keep the fast-path
1420 : * as small as possible. To avoid that effort being undone by a helpful
1421 : * compiler, try to explicitly forbid inlining.
1422 : */
1423 : static pg_noinline HeapTuple
1424 4762266 : SearchCatCacheMiss(CatCache *cache,
1425 : int nkeys,
1426 : uint32 hashValue,
1427 : Index hashIndex,
1428 : Datum v1,
1429 : Datum v2,
1430 : Datum v3,
1431 : Datum v4)
1432 : {
1433 : ScanKeyData cur_skey[CATCACHE_MAXKEYS];
1434 : Relation relation;
1435 : SysScanDesc scandesc;
1436 : HeapTuple ntp;
1437 : CatCTup *ct;
1438 : bool stale;
1439 : Datum arguments[CATCACHE_MAXKEYS];
1440 :
1441 : /* Initialize local parameter array */
1442 4762266 : arguments[0] = v1;
1443 4762266 : arguments[1] = v2;
1444 4762266 : arguments[2] = v3;
1445 4762266 : arguments[3] = v4;
1446 :
1447 : /*
1448 : * Tuple was not found in cache, so we have to try to retrieve it directly
1449 : * from the relation. If found, we will add it to the cache; if not
1450 : * found, we will add a negative cache entry instead.
1451 : *
1452 : * NOTE: it is possible for recursive cache lookups to occur while reading
1453 : * the relation --- for example, due to shared-cache-inval messages being
1454 : * processed during table_open(). This is OK. It's even possible for one
1455 : * of those lookups to find and enter the very same tuple we are trying to
1456 : * fetch here. If that happens, we will enter a second copy of the tuple
1457 : * into the cache. The first copy will never be referenced again, and
1458 : * will eventually age out of the cache, so there's no functional problem.
1459 : * This case is rare enough that it's not worth expending extra cycles to
1460 : * detect.
1461 : *
1462 : * Another case, which we *must* handle, is that the tuple could become
1463 : * outdated during CatalogCacheCreateEntry's attempt to detoast it (since
1464 : * AcceptInvalidationMessages can run during TOAST table access). We do
1465 : * not want to return already-stale catcache entries, so we loop around
1466 : * and do the table scan again if that happens.
1467 : */
1468 4762266 : relation = table_open(cache->cc_reloid, AccessShareLock);
1469 :
1470 : do
1471 : {
1472 : /*
1473 : * Ok, need to make a lookup in the relation, copy the scankey and
1474 : * fill out any per-call fields. (We must re-do this when retrying,
1475 : * because systable_beginscan scribbles on the scankey.)
1476 : */
1477 4762266 : memcpy(cur_skey, cache->cc_skey, sizeof(ScanKeyData) * nkeys);
1478 4762266 : cur_skey[0].sk_argument = v1;
1479 4762266 : cur_skey[1].sk_argument = v2;
1480 4762266 : cur_skey[2].sk_argument = v3;
1481 4762266 : cur_skey[3].sk_argument = v4;
1482 :
1483 4762266 : scandesc = systable_beginscan(relation,
1484 : cache->cc_indexoid,
1485 4762266 : IndexScanOK(cache, cur_skey),
1486 : NULL,
1487 : nkeys,
1488 : cur_skey);
1489 :
1490 4762266 : ct = NULL;
1491 4762266 : stale = false;
1492 :
1493 4762266 : while (HeapTupleIsValid(ntp = systable_getnext(scandesc)))
1494 : {
1495 3300468 : ct = CatalogCacheCreateEntry(cache, ntp, scandesc, NULL,
1496 : hashValue, hashIndex);
1497 : /* upon failure, we must start the scan over */
1498 3300468 : if (ct == NULL)
1499 : {
1500 0 : stale = true;
1501 0 : break;
1502 : }
1503 : /* immediately set the refcount to 1 */
1504 3300468 : ResourceOwnerEnlarge(CurrentResourceOwner);
1505 3300468 : ct->refcount++;
1506 3300468 : ResourceOwnerRememberCatCacheRef(CurrentResourceOwner, &ct->tuple);
1507 3300468 : break; /* assume only one match */
1508 : }
1509 :
1510 4762262 : systable_endscan(scandesc);
1511 4762262 : } while (stale);
1512 :
1513 4762262 : table_close(relation, AccessShareLock);
1514 :
1515 : /*
1516 : * If tuple was not found, we need to build a negative cache entry
1517 : * containing a fake tuple. The fake tuple has the correct key columns,
1518 : * but nulls everywhere else.
1519 : *
1520 : * In bootstrap mode, we don't build negative entries, because the cache
1521 : * invalidation mechanism isn't alive and can't clear them if the tuple
1522 : * gets created later. (Bootstrap doesn't do UPDATEs, so it doesn't need
1523 : * cache inval for that.)
1524 : */
1525 4762262 : if (ct == NULL)
1526 : {
1527 1461794 : if (IsBootstrapProcessingMode())
1528 43440 : return NULL;
1529 :
1530 1418354 : ct = CatalogCacheCreateEntry(cache, NULL, NULL, arguments,
1531 : hashValue, hashIndex);
1532 :
1533 : /* Creating a negative cache entry shouldn't fail */
1534 : Assert(ct != NULL);
1535 :
1536 : CACHE_elog(DEBUG2, "SearchCatCache(%s): Contains %d/%d tuples",
1537 : cache->cc_relname, cache->cc_ntup, CacheHdr->ch_ntup);
1538 : CACHE_elog(DEBUG2, "SearchCatCache(%s): put neg entry in bucket %d",
1539 : cache->cc_relname, hashIndex);
1540 :
1541 : /*
1542 : * We are not returning the negative entry to the caller, so leave its
1543 : * refcount zero.
1544 : */
1545 :
1546 1418354 : return NULL;
1547 : }
1548 :
1549 : CACHE_elog(DEBUG2, "SearchCatCache(%s): Contains %d/%d tuples",
1550 : cache->cc_relname, cache->cc_ntup, CacheHdr->ch_ntup);
1551 : CACHE_elog(DEBUG2, "SearchCatCache(%s): put in bucket %d",
1552 : cache->cc_relname, hashIndex);
1553 :
1554 : #ifdef CATCACHE_STATS
1555 : cache->cc_newloads++;
1556 : #endif
1557 :
1558 3300468 : return &ct->tuple;
1559 : }
1560 :
1561 : /*
1562 : * ReleaseCatCache
1563 : *
1564 : * Decrement the reference count of a catcache entry (releasing the
1565 : * hold grabbed by a successful SearchCatCache).
1566 : *
1567 : * NOTE: if compiled with -DCATCACHE_FORCE_RELEASE then catcache entries
1568 : * will be freed as soon as their refcount goes to zero. In combination
1569 : * with aset.c's CLOBBER_FREED_MEMORY option, this provides a good test
1570 : * to catch references to already-released catcache entries.
1571 : */
1572 : void
1573 69489338 : ReleaseCatCache(HeapTuple tuple)
1574 : {
1575 69489338 : ReleaseCatCacheWithOwner(tuple, CurrentResourceOwner);
1576 69489338 : }
1577 :
1578 : static void
1579 69499428 : ReleaseCatCacheWithOwner(HeapTuple tuple, ResourceOwner resowner)
1580 : {
1581 69499428 : CatCTup *ct = (CatCTup *) (((char *) tuple) -
1582 : offsetof(CatCTup, tuple));
1583 :
1584 : /* Safety checks to ensure we were handed a cache entry */
1585 : Assert(ct->ct_magic == CT_MAGIC);
1586 : Assert(ct->refcount > 0);
1587 :
1588 69499428 : ct->refcount--;
1589 69499428 : if (resowner)
1590 69489338 : ResourceOwnerForgetCatCacheRef(CurrentResourceOwner, &ct->tuple);
1591 :
1592 69499428 : if (
1593 : #ifndef CATCACHE_FORCE_RELEASE
1594 69499428 : ct->dead &&
1595 : #endif
1596 15152 : ct->refcount == 0 &&
1597 15040 : (ct->c_list == NULL || ct->c_list->refcount == 0))
1598 15040 : CatCacheRemoveCTup(ct->my_cache, ct);
1599 69499428 : }
1600 :
1601 :
1602 : /*
1603 : * GetCatCacheHashValue
1604 : *
1605 : * Compute the hash value for a given set of search keys.
1606 : *
1607 : * The reason for exposing this as part of the API is that the hash value is
1608 : * exposed in cache invalidation operations, so there are places outside the
1609 : * catcache code that need to be able to compute the hash values.
1610 : */
1611 : uint32
1612 141056 : GetCatCacheHashValue(CatCache *cache,
1613 : Datum v1,
1614 : Datum v2,
1615 : Datum v3,
1616 : Datum v4)
1617 : {
1618 : /*
1619 : * one-time startup overhead for each cache
1620 : */
1621 141056 : if (cache->cc_tupdesc == NULL)
1622 21358 : CatalogCacheInitializeCache(cache);
1623 :
1624 : /*
1625 : * calculate the hash value
1626 : */
1627 141056 : return CatalogCacheComputeHashValue(cache, cache->cc_nkeys, v1, v2, v3, v4);
1628 : }
1629 :
1630 :
1631 : /*
1632 : * SearchCatCacheList
1633 : *
1634 : * Generate a list of all tuples matching a partial key (that is,
1635 : * a key specifying just the first K of the cache's N key columns).
1636 : *
1637 : * It doesn't make any sense to specify all of the cache's key columns
1638 : * here: since the key is unique, there could be at most one match, so
1639 : * you ought to use SearchCatCache() instead. Hence this function takes
1640 : * one fewer Datum argument than SearchCatCache() does.
1641 : *
1642 : * The caller must not modify the list object or the pointed-to tuples,
1643 : * and must call ReleaseCatCacheList() when done with the list.
1644 : */
1645 : CatCList *
1646 3061002 : SearchCatCacheList(CatCache *cache,
1647 : int nkeys,
1648 : Datum v1,
1649 : Datum v2,
1650 : Datum v3)
1651 : {
1652 3061002 : Datum v4 = 0; /* dummy last-column value */
1653 : Datum arguments[CATCACHE_MAXKEYS];
1654 : uint32 lHashValue;
1655 : Index lHashIndex;
1656 : dlist_iter iter;
1657 : dlist_head *lbucket;
1658 : CatCList *cl;
1659 : CatCTup *ct;
1660 : List *volatile ctlist;
1661 : ListCell *ctlist_item;
1662 : int nmembers;
1663 : bool ordered;
1664 : HeapTuple ntp;
1665 : MemoryContext oldcxt;
1666 : int i;
1667 :
1668 : /*
1669 : * one-time startup overhead for each cache
1670 : */
1671 3061002 : if (unlikely(cache->cc_tupdesc == NULL))
1672 30136 : CatalogCacheInitializeCache(cache);
1673 :
1674 : Assert(nkeys > 0 && nkeys < cache->cc_nkeys);
1675 :
1676 : #ifdef CATCACHE_STATS
1677 : cache->cc_lsearches++;
1678 : #endif
1679 :
1680 : /* Initialize local parameter array */
1681 3061002 : arguments[0] = v1;
1682 3061002 : arguments[1] = v2;
1683 3061002 : arguments[2] = v3;
1684 3061002 : arguments[3] = v4;
1685 :
1686 : /*
1687 : * If we haven't previously done a list search in this cache, create the
1688 : * bucket header array; otherwise, consider whether it's time to enlarge
1689 : * it.
1690 : */
1691 3061002 : if (cache->cc_lbucket == NULL)
1692 : {
1693 : /* Arbitrary initial size --- must be a power of 2 */
1694 34074 : int nbuckets = 16;
1695 :
1696 34074 : cache->cc_lbucket = (dlist_head *)
1697 34074 : MemoryContextAllocZero(CacheMemoryContext,
1698 : nbuckets * sizeof(dlist_head));
1699 : /* Don't set cc_nlbuckets if we get OOM allocating cc_lbucket */
1700 34074 : cache->cc_nlbuckets = nbuckets;
1701 : }
1702 : else
1703 : {
1704 : /*
1705 : * If the hash table has become too full, enlarge the buckets array.
1706 : * Quite arbitrarily, we enlarge when fill factor > 2.
1707 : */
1708 3026928 : if (cache->cc_nlist > cache->cc_nlbuckets * 2)
1709 1036 : RehashCatCacheLists(cache);
1710 : }
1711 :
1712 : /*
1713 : * Find the hash bucket in which to look for the CatCList.
1714 : */
1715 3061002 : lHashValue = CatalogCacheComputeHashValue(cache, nkeys, v1, v2, v3, v4);
1716 3061002 : lHashIndex = HASH_INDEX(lHashValue, cache->cc_nlbuckets);
1717 :
1718 : /*
1719 : * scan the items until we find a match or exhaust our list
1720 : *
1721 : * Note: it's okay to use dlist_foreach here, even though we modify the
1722 : * dlist within the loop, because we don't continue the loop afterwards.
1723 : */
1724 3061002 : lbucket = &cache->cc_lbucket[lHashIndex];
1725 3362638 : dlist_foreach(iter, lbucket)
1726 : {
1727 3103726 : cl = dlist_container(CatCList, cache_elem, iter.cur);
1728 :
1729 3103726 : if (cl->dead)
1730 0 : continue; /* ignore dead entries */
1731 :
1732 3103726 : if (cl->hash_value != lHashValue)
1733 301636 : continue; /* quickly skip entry if wrong hash val */
1734 :
1735 : /*
1736 : * see if the cached list matches our key.
1737 : */
1738 2802090 : if (cl->nkeys != nkeys)
1739 0 : continue;
1740 :
1741 2802090 : if (!CatalogCacheCompareTuple(cache, nkeys, cl->keys, arguments))
1742 0 : continue;
1743 :
1744 : /*
1745 : * We found a matching list. Move the list to the front of the list
1746 : * for its hashbucket, so as to speed subsequent searches. (We do not
1747 : * move the members to the fronts of their hashbucket lists, however,
1748 : * since there's no point in that unless they are searched for
1749 : * individually.)
1750 : */
1751 2802090 : dlist_move_head(lbucket, &cl->cache_elem);
1752 :
1753 : /* Bump the list's refcount and return it */
1754 2802090 : ResourceOwnerEnlarge(CurrentResourceOwner);
1755 2802090 : cl->refcount++;
1756 2802090 : ResourceOwnerRememberCatCacheListRef(CurrentResourceOwner, cl);
1757 :
1758 : CACHE_elog(DEBUG2, "SearchCatCacheList(%s): found list",
1759 : cache->cc_relname);
1760 :
1761 : #ifdef CATCACHE_STATS
1762 : cache->cc_lhits++;
1763 : #endif
1764 :
1765 2802090 : return cl;
1766 : }
1767 :
1768 : /*
1769 : * List was not found in cache, so we have to build it by reading the
1770 : * relation. For each matching tuple found in the relation, use an
1771 : * existing cache entry if possible, else build a new one.
1772 : *
1773 : * We have to bump the member refcounts temporarily to ensure they won't
1774 : * get dropped from the cache while loading other members. We use a PG_TRY
1775 : * block to ensure we can undo those refcounts if we get an error before
1776 : * we finish constructing the CatCList. ctlist must be valid throughout
1777 : * the PG_TRY block.
1778 : */
1779 258912 : ctlist = NIL;
1780 :
1781 258912 : PG_TRY();
1782 : {
1783 : ScanKeyData cur_skey[CATCACHE_MAXKEYS];
1784 : Relation relation;
1785 : SysScanDesc scandesc;
1786 : bool stale;
1787 :
1788 258912 : relation = table_open(cache->cc_reloid, AccessShareLock);
1789 :
1790 : do
1791 : {
1792 : /*
1793 : * Ok, need to make a lookup in the relation, copy the scankey and
1794 : * fill out any per-call fields. (We must re-do this when
1795 : * retrying, because systable_beginscan scribbles on the scankey.)
1796 : */
1797 258912 : memcpy(cur_skey, cache->cc_skey, sizeof(ScanKeyData) * cache->cc_nkeys);
1798 258912 : cur_skey[0].sk_argument = v1;
1799 258912 : cur_skey[1].sk_argument = v2;
1800 258912 : cur_skey[2].sk_argument = v3;
1801 258912 : cur_skey[3].sk_argument = v4;
1802 :
1803 517824 : scandesc = systable_beginscan(relation,
1804 : cache->cc_indexoid,
1805 258912 : IndexScanOK(cache, cur_skey),
1806 : NULL,
1807 : nkeys,
1808 : cur_skey);
1809 :
1810 : /* The list will be ordered iff we are doing an index scan */
1811 258912 : ordered = (scandesc->irel != NULL);
1812 :
1813 258912 : stale = false;
1814 :
1815 1074730 : while (HeapTupleIsValid(ntp = systable_getnext(scandesc)))
1816 : {
1817 : uint32 hashValue;
1818 : Index hashIndex;
1819 815818 : bool found = false;
1820 : dlist_head *bucket;
1821 :
1822 : /*
1823 : * See if there's an entry for this tuple already.
1824 : */
1825 815818 : ct = NULL;
1826 815818 : hashValue = CatalogCacheComputeTupleHashValue(cache, cache->cc_nkeys, ntp);
1827 815818 : hashIndex = HASH_INDEX(hashValue, cache->cc_nbuckets);
1828 :
1829 815818 : bucket = &cache->cc_bucket[hashIndex];
1830 1124162 : dlist_foreach(iter, bucket)
1831 : {
1832 428742 : ct = dlist_container(CatCTup, cache_elem, iter.cur);
1833 :
1834 428742 : if (ct->dead || ct->negative)
1835 856 : continue; /* ignore dead and negative entries */
1836 :
1837 427886 : if (ct->hash_value != hashValue)
1838 291012 : continue; /* quickly skip entry if wrong hash val */
1839 :
1840 136874 : if (!ItemPointerEquals(&(ct->tuple.t_self), &(ntp->t_self)))
1841 0 : continue; /* not same tuple */
1842 :
1843 : /*
1844 : * Found a match, but can't use it if it belongs to
1845 : * another list already
1846 : */
1847 136874 : if (ct->c_list)
1848 16476 : continue;
1849 :
1850 120398 : found = true;
1851 120398 : break; /* A-OK */
1852 : }
1853 :
1854 815818 : if (!found)
1855 : {
1856 : /* We didn't find a usable entry, so make a new one */
1857 695420 : ct = CatalogCacheCreateEntry(cache, ntp, scandesc, NULL,
1858 : hashValue, hashIndex);
1859 : /* upon failure, we must start the scan over */
1860 695420 : if (ct == NULL)
1861 : {
1862 : /*
1863 : * Release refcounts on any items we already had. We
1864 : * dare not try to free them if they're now
1865 : * unreferenced, since an error while doing that would
1866 : * result in the PG_CATCH below doing extra refcount
1867 : * decrements. Besides, we'll likely re-adopt those
1868 : * items in the next iteration, so it's not worth
1869 : * complicating matters to try to get rid of them.
1870 : */
1871 0 : foreach(ctlist_item, ctlist)
1872 : {
1873 0 : ct = (CatCTup *) lfirst(ctlist_item);
1874 : Assert(ct->c_list == NULL);
1875 : Assert(ct->refcount > 0);
1876 0 : ct->refcount--;
1877 : }
1878 : /* Reset ctlist in preparation for new try */
1879 0 : ctlist = NIL;
1880 0 : stale = true;
1881 0 : break;
1882 : }
1883 : }
1884 :
1885 : /* Careful here: add entry to ctlist, then bump its refcount */
1886 : /* This way leaves state correct if lappend runs out of memory */
1887 815818 : ctlist = lappend(ctlist, ct);
1888 815818 : ct->refcount++;
1889 : }
1890 :
1891 258912 : systable_endscan(scandesc);
1892 258912 : } while (stale);
1893 :
1894 258912 : table_close(relation, AccessShareLock);
1895 :
1896 : /* Make sure the resource owner has room to remember this entry. */
1897 258912 : ResourceOwnerEnlarge(CurrentResourceOwner);
1898 :
1899 : /* Now we can build the CatCList entry. */
1900 258912 : oldcxt = MemoryContextSwitchTo(CacheMemoryContext);
1901 258912 : nmembers = list_length(ctlist);
1902 : cl = (CatCList *)
1903 258912 : palloc(offsetof(CatCList, members) + nmembers * sizeof(CatCTup *));
1904 :
1905 : /* Extract key values */
1906 258912 : CatCacheCopyKeys(cache->cc_tupdesc, nkeys, cache->cc_keyno,
1907 258912 : arguments, cl->keys);
1908 258912 : MemoryContextSwitchTo(oldcxt);
1909 :
1910 : /*
1911 : * We are now past the last thing that could trigger an elog before we
1912 : * have finished building the CatCList and remembering it in the
1913 : * resource owner. So it's OK to fall out of the PG_TRY, and indeed
1914 : * we'd better do so before we start marking the members as belonging
1915 : * to the list.
1916 : */
1917 : }
1918 0 : PG_CATCH();
1919 : {
1920 0 : foreach(ctlist_item, ctlist)
1921 : {
1922 0 : ct = (CatCTup *) lfirst(ctlist_item);
1923 : Assert(ct->c_list == NULL);
1924 : Assert(ct->refcount > 0);
1925 0 : ct->refcount--;
1926 0 : if (
1927 : #ifndef CATCACHE_FORCE_RELEASE
1928 0 : ct->dead &&
1929 : #endif
1930 0 : ct->refcount == 0 &&
1931 0 : (ct->c_list == NULL || ct->c_list->refcount == 0))
1932 0 : CatCacheRemoveCTup(cache, ct);
1933 : }
1934 :
1935 0 : PG_RE_THROW();
1936 : }
1937 258912 : PG_END_TRY();
1938 :
1939 258912 : cl->cl_magic = CL_MAGIC;
1940 258912 : cl->my_cache = cache;
1941 258912 : cl->refcount = 0; /* for the moment */
1942 258912 : cl->dead = false;
1943 258912 : cl->ordered = ordered;
1944 258912 : cl->nkeys = nkeys;
1945 258912 : cl->hash_value = lHashValue;
1946 258912 : cl->n_members = nmembers;
1947 :
1948 258912 : i = 0;
1949 1074730 : foreach(ctlist_item, ctlist)
1950 : {
1951 815818 : cl->members[i++] = ct = (CatCTup *) lfirst(ctlist_item);
1952 : Assert(ct->c_list == NULL);
1953 815818 : ct->c_list = cl;
1954 : /* release the temporary refcount on the member */
1955 : Assert(ct->refcount > 0);
1956 815818 : ct->refcount--;
1957 : /* mark list dead if any members already dead */
1958 815818 : if (ct->dead)
1959 0 : cl->dead = true;
1960 : }
1961 : Assert(i == nmembers);
1962 :
1963 : /*
1964 : * Add the CatCList to the appropriate bucket, and count it.
1965 : */
1966 258912 : dlist_push_head(lbucket, &cl->cache_elem);
1967 :
1968 258912 : cache->cc_nlist++;
1969 :
1970 : /* Finally, bump the list's refcount and return it */
1971 258912 : cl->refcount++;
1972 258912 : ResourceOwnerRememberCatCacheListRef(CurrentResourceOwner, cl);
1973 :
1974 : CACHE_elog(DEBUG2, "SearchCatCacheList(%s): made list of %d members",
1975 : cache->cc_relname, nmembers);
1976 :
1977 258912 : return cl;
1978 : }
1979 :
1980 : /*
1981 : * ReleaseCatCacheList
1982 : *
1983 : * Decrement the reference count of a catcache list.
1984 : */
1985 : void
1986 3060966 : ReleaseCatCacheList(CatCList *list)
1987 : {
1988 3060966 : ReleaseCatCacheListWithOwner(list, CurrentResourceOwner);
1989 3060966 : }
1990 :
1991 : static void
1992 3061002 : ReleaseCatCacheListWithOwner(CatCList *list, ResourceOwner resowner)
1993 : {
1994 : /* Safety checks to ensure we were handed a cache entry */
1995 : Assert(list->cl_magic == CL_MAGIC);
1996 : Assert(list->refcount > 0);
1997 3061002 : list->refcount--;
1998 3061002 : if (resowner)
1999 3060966 : ResourceOwnerForgetCatCacheListRef(CurrentResourceOwner, list);
2000 :
2001 3061002 : if (
2002 : #ifndef CATCACHE_FORCE_RELEASE
2003 3061002 : list->dead &&
2004 : #endif
2005 6 : list->refcount == 0)
2006 6 : CatCacheRemoveCList(list->my_cache, list);
2007 3061002 : }
2008 :
2009 :
2010 : /*
2011 : * CatalogCacheCreateEntry
2012 : * Create a new CatCTup entry, copying the given HeapTuple and other
2013 : * supplied data into it. The new entry initially has refcount 0.
2014 : *
2015 : * To create a normal cache entry, ntp must be the HeapTuple just fetched
2016 : * from scandesc, and "arguments" is not used. To create a negative cache
2017 : * entry, pass NULL for ntp and scandesc; then "arguments" is the cache
2018 : * keys to use. In either case, hashValue/hashIndex are the hash values
2019 : * computed from the cache keys.
2020 : *
2021 : * Returns NULL if we attempt to detoast the tuple and observe that it
2022 : * became stale. (This cannot happen for a negative entry.) Caller must
2023 : * retry the tuple lookup in that case.
2024 : */
2025 : static CatCTup *
2026 5414242 : CatalogCacheCreateEntry(CatCache *cache, HeapTuple ntp, SysScanDesc scandesc,
2027 : Datum *arguments,
2028 : uint32 hashValue, Index hashIndex)
2029 : {
2030 : CatCTup *ct;
2031 : HeapTuple dtp;
2032 : MemoryContext oldcxt;
2033 :
2034 5414242 : if (ntp)
2035 : {
2036 : int i;
2037 :
2038 : /*
2039 : * The visibility recheck below essentially never fails during our
2040 : * regression tests, and there's no easy way to force it to fail for
2041 : * testing purposes. To ensure we have test coverage for the retry
2042 : * paths in our callers, make debug builds randomly fail about 0.1% of
2043 : * the times through this code path, even when there's no toasted
2044 : * fields.
2045 : */
2046 : #ifdef USE_ASSERT_CHECKING
2047 : if (pg_prng_uint32(&pg_global_prng_state) <= (PG_UINT32_MAX / 1000))
2048 : return NULL;
2049 : #endif
2050 :
2051 : /*
2052 : * If there are any out-of-line toasted fields in the tuple, expand
2053 : * them in-line. This saves cycles during later use of the catcache
2054 : * entry, and also protects us against the possibility of the toast
2055 : * tuples being freed before we attempt to fetch them, in case of
2056 : * something using a slightly stale catcache entry.
2057 : */
2058 3995888 : if (HeapTupleHasExternal(ntp))
2059 : {
2060 3184 : dtp = toast_flatten_tuple(ntp, cache->cc_tupdesc);
2061 :
2062 : /*
2063 : * The tuple could become stale while we are doing toast table
2064 : * access (since AcceptInvalidationMessages can run then), so we
2065 : * must recheck its visibility afterwards.
2066 : */
2067 3184 : if (!systable_recheck_tuple(scandesc, ntp))
2068 : {
2069 0 : heap_freetuple(dtp);
2070 0 : return NULL;
2071 : }
2072 : }
2073 : else
2074 3992704 : dtp = ntp;
2075 :
2076 : /* Allocate memory for CatCTup and the cached tuple in one go */
2077 3995888 : oldcxt = MemoryContextSwitchTo(CacheMemoryContext);
2078 :
2079 3995888 : ct = (CatCTup *) palloc(sizeof(CatCTup) +
2080 3995888 : MAXIMUM_ALIGNOF + dtp->t_len);
2081 3995888 : ct->tuple.t_len = dtp->t_len;
2082 3995888 : ct->tuple.t_self = dtp->t_self;
2083 3995888 : ct->tuple.t_tableOid = dtp->t_tableOid;
2084 3995888 : ct->tuple.t_data = (HeapTupleHeader)
2085 3995888 : MAXALIGN(((char *) ct) + sizeof(CatCTup));
2086 : /* copy tuple contents */
2087 3995888 : memcpy((char *) ct->tuple.t_data,
2088 3995888 : (const char *) dtp->t_data,
2089 3995888 : dtp->t_len);
2090 3995888 : MemoryContextSwitchTo(oldcxt);
2091 :
2092 3995888 : if (dtp != ntp)
2093 3184 : heap_freetuple(dtp);
2094 :
2095 : /* extract keys - they'll point into the tuple if not by-value */
2096 11697058 : for (i = 0; i < cache->cc_nkeys; i++)
2097 : {
2098 : Datum atp;
2099 : bool isnull;
2100 :
2101 7701170 : atp = heap_getattr(&ct->tuple,
2102 : cache->cc_keyno[i],
2103 : cache->cc_tupdesc,
2104 : &isnull);
2105 : Assert(!isnull);
2106 7701170 : ct->keys[i] = atp;
2107 : }
2108 : }
2109 : else
2110 : {
2111 : /* Set up keys for a negative cache entry */
2112 1418354 : oldcxt = MemoryContextSwitchTo(CacheMemoryContext);
2113 1418354 : ct = (CatCTup *) palloc(sizeof(CatCTup));
2114 :
2115 : /*
2116 : * Store keys - they'll point into separately allocated memory if not
2117 : * by-value.
2118 : */
2119 1418354 : CatCacheCopyKeys(cache->cc_tupdesc, cache->cc_nkeys, cache->cc_keyno,
2120 1418354 : arguments, ct->keys);
2121 1418354 : MemoryContextSwitchTo(oldcxt);
2122 : }
2123 :
2124 : /*
2125 : * Finish initializing the CatCTup header, and add it to the cache's
2126 : * linked list and counts.
2127 : */
2128 5414242 : ct->ct_magic = CT_MAGIC;
2129 5414242 : ct->my_cache = cache;
2130 5414242 : ct->c_list = NULL;
2131 5414242 : ct->refcount = 0; /* for the moment */
2132 5414242 : ct->dead = false;
2133 5414242 : ct->negative = (ntp == NULL);
2134 5414242 : ct->hash_value = hashValue;
2135 :
2136 5414242 : dlist_push_head(&cache->cc_bucket[hashIndex], &ct->cache_elem);
2137 :
2138 5414242 : cache->cc_ntup++;
2139 5414242 : CacheHdr->ch_ntup++;
2140 :
2141 : /*
2142 : * If the hash table has become too full, enlarge the buckets array. Quite
2143 : * arbitrarily, we enlarge when fill factor > 2.
2144 : */
2145 5414242 : if (cache->cc_ntup > cache->cc_nbuckets * 2)
2146 5016 : RehashCatCache(cache);
2147 :
2148 5414242 : return ct;
2149 : }
2150 :
2151 : /*
2152 : * Helper routine that frees keys stored in the keys array.
2153 : */
2154 : static void
2155 502332 : CatCacheFreeKeys(TupleDesc tupdesc, int nkeys, int *attnos, Datum *keys)
2156 : {
2157 : int i;
2158 :
2159 1544676 : for (i = 0; i < nkeys; i++)
2160 : {
2161 1042344 : int attnum = attnos[i];
2162 : Form_pg_attribute att;
2163 :
2164 : /* system attribute are not supported in caches */
2165 : Assert(attnum > 0);
2166 :
2167 1042344 : att = TupleDescAttr(tupdesc, attnum - 1);
2168 :
2169 1042344 : if (!att->attbyval)
2170 437302 : pfree(DatumGetPointer(keys[i]));
2171 : }
2172 502332 : }
2173 :
2174 : /*
2175 : * Helper routine that copies the keys in the srckeys array into the dstkeys
2176 : * one, guaranteeing that the datums are fully allocated in the current memory
2177 : * context.
2178 : */
2179 : static void
2180 1677266 : CatCacheCopyKeys(TupleDesc tupdesc, int nkeys, int *attnos,
2181 : Datum *srckeys, Datum *dstkeys)
2182 : {
2183 : int i;
2184 :
2185 : /*
2186 : * XXX: memory and lookup performance could possibly be improved by
2187 : * storing all keys in one allocation.
2188 : */
2189 :
2190 5249044 : for (i = 0; i < nkeys; i++)
2191 : {
2192 3571778 : int attnum = attnos[i];
2193 3571778 : Form_pg_attribute att = TupleDescAttr(tupdesc, attnum - 1);
2194 3571778 : Datum src = srckeys[i];
2195 : NameData srcname;
2196 :
2197 : /*
2198 : * Must be careful in case the caller passed a C string where a NAME
2199 : * is wanted: convert the given argument to a correctly padded NAME.
2200 : * Otherwise the memcpy() done by datumCopy() could fall off the end
2201 : * of memory.
2202 : */
2203 3571778 : if (att->atttypid == NAMEOID)
2204 : {
2205 720190 : namestrcpy(&srcname, DatumGetCString(src));
2206 720190 : src = NameGetDatum(&srcname);
2207 : }
2208 :
2209 3571778 : dstkeys[i] = datumCopy(src,
2210 3571778 : att->attbyval,
2211 3571778 : att->attlen);
2212 : }
2213 1677266 : }
2214 :
2215 : /*
2216 : * PrepareToInvalidateCacheTuple()
2217 : *
2218 : * This is part of a rather subtle chain of events, so pay attention:
2219 : *
2220 : * When a tuple is inserted or deleted, it cannot be flushed from the
2221 : * catcaches immediately, for reasons explained at the top of cache/inval.c.
2222 : * Instead we have to add entry(s) for the tuple to a list of pending tuple
2223 : * invalidations that will be done at the end of the command or transaction.
2224 : *
2225 : * The lists of tuples that need to be flushed are kept by inval.c. This
2226 : * routine is a helper routine for inval.c. Given a tuple belonging to
2227 : * the specified relation, find all catcaches it could be in, compute the
2228 : * correct hash value for each such catcache, and call the specified
2229 : * function to record the cache id and hash value in inval.c's lists.
2230 : * SysCacheInvalidate will be called later, if appropriate,
2231 : * using the recorded information.
2232 : *
2233 : * For an insert or delete, tuple is the target tuple and newtuple is NULL.
2234 : * For an update, we are called just once, with tuple being the old tuple
2235 : * version and newtuple the new version. We should make two list entries
2236 : * if the tuple's hash value changed, but only one if it didn't.
2237 : *
2238 : * Note that it is irrelevant whether the given tuple is actually loaded
2239 : * into the catcache at the moment. Even if it's not there now, it might
2240 : * be by the end of the command, or there might be a matching negative entry
2241 : * to flush --- or other backends' caches might have such entries --- so
2242 : * we have to make list entries to flush it later.
2243 : *
2244 : * Also note that it's not an error if there are no catcaches for the
2245 : * specified relation. inval.c doesn't know exactly which rels have
2246 : * catcaches --- it will call this routine for any tuple that's in a
2247 : * system relation.
2248 : */
2249 : void
2250 2570270 : PrepareToInvalidateCacheTuple(Relation relation,
2251 : HeapTuple tuple,
2252 : HeapTuple newtuple,
2253 : void (*function) (int, uint32, Oid))
2254 : {
2255 : slist_iter iter;
2256 : Oid reloid;
2257 :
2258 : CACHE_elog(DEBUG2, "PrepareToInvalidateCacheTuple: called");
2259 :
2260 : /*
2261 : * sanity checks
2262 : */
2263 : Assert(RelationIsValid(relation));
2264 : Assert(HeapTupleIsValid(tuple));
2265 : Assert(PointerIsValid(function));
2266 : Assert(CacheHdr != NULL);
2267 :
2268 2570270 : reloid = RelationGetRelid(relation);
2269 :
2270 : /* ----------------
2271 : * for each cache
2272 : * if the cache contains tuples from the specified relation
2273 : * compute the tuple's hash value(s) in this cache,
2274 : * and call the passed function to register the information.
2275 : * ----------------
2276 : */
2277 :
2278 215902680 : slist_foreach(iter, &CacheHdr->ch_caches)
2279 : {
2280 213332410 : CatCache *ccp = slist_container(CatCache, cc_next, iter.cur);
2281 : uint32 hashvalue;
2282 : Oid dbid;
2283 :
2284 213332410 : if (ccp->cc_reloid != reloid)
2285 208660274 : continue;
2286 :
2287 : /* Just in case cache hasn't finished initialization yet... */
2288 4672136 : if (ccp->cc_tupdesc == NULL)
2289 7902 : CatalogCacheInitializeCache(ccp);
2290 :
2291 4672136 : hashvalue = CatalogCacheComputeTupleHashValue(ccp, ccp->cc_nkeys, tuple);
2292 4672136 : dbid = ccp->cc_relisshared ? (Oid) 0 : MyDatabaseId;
2293 :
2294 4672136 : (*function) (ccp->id, hashvalue, dbid);
2295 :
2296 4672136 : if (newtuple)
2297 : {
2298 : uint32 newhashvalue;
2299 :
2300 362574 : newhashvalue = CatalogCacheComputeTupleHashValue(ccp, ccp->cc_nkeys, newtuple);
2301 :
2302 362574 : if (newhashvalue != hashvalue)
2303 5712 : (*function) (ccp->id, newhashvalue, dbid);
2304 : }
2305 : }
2306 2570270 : }
2307 :
2308 : /* ResourceOwner callbacks */
2309 :
2310 : static void
2311 10090 : ResOwnerReleaseCatCache(Datum res)
2312 : {
2313 10090 : ReleaseCatCacheWithOwner((HeapTuple) DatumGetPointer(res), NULL);
2314 10090 : }
2315 :
2316 : static char *
2317 0 : ResOwnerPrintCatCache(Datum res)
2318 : {
2319 0 : HeapTuple tuple = (HeapTuple) DatumGetPointer(res);
2320 0 : CatCTup *ct = (CatCTup *) (((char *) tuple) -
2321 : offsetof(CatCTup, tuple));
2322 :
2323 : /* Safety check to ensure we were handed a cache entry */
2324 : Assert(ct->ct_magic == CT_MAGIC);
2325 :
2326 0 : return psprintf("cache %s (%d), tuple %u/%u has count %d",
2327 0 : ct->my_cache->cc_relname, ct->my_cache->id,
2328 0 : ItemPointerGetBlockNumber(&(tuple->t_self)),
2329 0 : ItemPointerGetOffsetNumber(&(tuple->t_self)),
2330 : ct->refcount);
2331 : }
2332 :
2333 : static void
2334 36 : ResOwnerReleaseCatCacheList(Datum res)
2335 : {
2336 36 : ReleaseCatCacheListWithOwner((CatCList *) DatumGetPointer(res), NULL);
2337 36 : }
2338 :
2339 : static char *
2340 0 : ResOwnerPrintCatCacheList(Datum res)
2341 : {
2342 0 : CatCList *list = (CatCList *) DatumGetPointer(res);
2343 :
2344 0 : return psprintf("cache %s (%d), list %p has count %d",
2345 0 : list->my_cache->cc_relname, list->my_cache->id,
2346 : list, list->refcount);
2347 : }
|