Handle-with-cache.c

// Improved get_handle() with double-check UserProfile* get_user_profile_handle_safe(int user_id) { pthread_mutex_lock(&cache_lock); CacheEntry *entry = g_hash_table_lookup(handle_cache, &user_id); if (entry) { entry->ref_count++; pthread_mutex_unlock(&cache_lock); return entry->profile; } pthread_mutex_unlock(&cache_lock); // Load outside lock UserProfile *profile = load_user_profile_from_disk(user_id);

// Find the entry for this profile (simplified; real code needs reverse mapping) GHashTableIter iter; gpointer key, value; g_hash_table_iter_init(&iter, handle_cache); while (g_hash_table_iter_next(&iter, &key, &value)) { CacheEntry *entry = value; if (entry->profile == profile) { entry->ref_count--; if (entry->ref_count == 0) { // Last reference - we could evict immediately or mark as stale printf("No more references to user %d, marking for eviction\n", *(int*)key); } break; } } handle-with-cache.c

pthread_mutex_unlock(&cache_lock); } A cache without eviction is a memory leak. handle-with-cache.c should implement a policy like LRU (Least Recently Used) or TTL (Time To Live) . A "handle" is an opaque pointer or identifier

The module handle-with-cache.c exemplifies a classic design pattern: the . A "handle" is an opaque pointer or identifier to a resource, and the cache stores recently accessed handles to avoid redundant initialization or I/O operations. However, it introduces a race condition where two

pthread_mutex_unlock(&cache_lock); return profile; }

pthread_mutex_unlock(&cache_lock); } The cache_lock mutex protects the hash table, but note that get_handle() releases the lock during the actual load_user_profile_from_disk() call. This is crucial to avoid blocking all threads during I/O. However, it introduces a race condition where two threads might simultaneously miss the cache and both load the same resource.