@@ -10,6 +10,7 @@
struct xfs_inode;
struct xfs_buftarg;
struct xfs_da_geometry;
+struct btcache;
/*
* Define a user-level mount structure with all we need
@@ -155,6 +156,8 @@ typedef struct xfs_perag {
/* reference count */
uint8_t pagf_refcount_level;
+
+ struct btcache *pag_buf_hash;
} xfs_perag_t;
static inline struct xfs_ag_resv *
@@ -425,3 +425,236 @@ xfs_buf_associate_memory(
bp->b_length = BTOBB(len);
return 0;
}
+
+/*
+ * Buffer cache hash implementation
+ */
+
+struct btcache *
+btc_init(
+ unsigned int hashsize)
+{
+ struct btcache *btc;
+ unsigned int i, maxcount;
+
+ maxcount = hashsize * HASH_CACHE_RATIO;
+
+ btc = malloc(sizeof(struct btcache));
+ if (!btc)
+ return NULL;
+ btc->hash = calloc(hashsize, sizeof(struct btcache_hash));
+ if (!btc->hash) {
+ free(btc);
+ return NULL;
+ }
+
+ atomic_set(&btc->count, 0);
+ btc->max = 0;
+ btc->hits = 0;
+ btc->misses = 0;
+ btc->maxcount = maxcount;
+ btc->hashsize = hashsize;
+ btc->hashshift = libxfs_highbit32(hashsize);
+ pthread_mutex_init(&btc->lock, NULL);
+
+ for (i = 0; i < hashsize; i++) {
+ list_head_init(&btc->hash[i].chain);
+ btc->hash[i].count = 0;
+ pthread_mutex_init(&btc->hash[i].lock, NULL);
+ }
+
+ return btc;
+}
+
+void
+btc_destroy(
+ struct btcache *btc)
+{
+ unsigned int i;
+
+ if (!btc)
+ return;
+
+ for (i = 0; i < btc->hashsize; i++) {
+ list_head_destroy(&btc->hash[i].chain);
+ pthread_mutex_destroy(&btc->hash[i].lock);
+ }
+ pthread_mutex_destroy(&btc->lock);
+ free(btc->hash);
+ free(btc);
+}
+
+
+#define HASH_REPORT (3 * HASH_CACHE_RATIO)
+void
+btc_report_ag(
+ FILE *fp,
+ const char *name,
+ xfs_agnumber_t agno,
+ struct btcache *btc)
+{
+ int i;
+ unsigned long count, index, total;
+ unsigned long hash_bucket_lengths[HASH_REPORT + 2];
+
+ if ((btc->hits + btc->misses) == 0)
+ return;
+
+ /* report btc summary */
+ fprintf(fp, "%8u|\t%9u\t%9u\t%8u\t%8u\t%8llu\t%8llu\t%5.2f\n",
+ agno,
+ btc->maxcount,
+ btc->max,
+ atomic_read(&btc->count),
+ btc->hashsize,
+ btc->hits,
+ btc->misses,
+ (double)btc->hits * 100 /
+ (btc->hits + btc->misses)
+ );
+
+ /* report hash bucket lengths */
+ memset(hash_bucket_lengths, 0, sizeof(hash_bucket_lengths));
+
+ for (i = 0; i < btc->hashsize; i++) {
+ count = btc->hash[i].count;
+ if (count > HASH_REPORT)
+ index = HASH_REPORT + 1;
+ else
+ index = count;
+ hash_bucket_lengths[index]++;
+ }
+
+ total = 0;
+ for (i = 0; i < HASH_REPORT + 1; i++) {
+ total += i * hash_bucket_lengths[i];
+ if (hash_bucket_lengths[i] == 0)
+ continue;
+ fprintf(fp, "Hash buckets with %2d entries %6ld (%3ld%%)\n",
+ i, hash_bucket_lengths[i],
+ (i * hash_bucket_lengths[i] * 100) /
+ atomic_read(&btc->count));
+ }
+ if (hash_bucket_lengths[i]) /* last report bucket is the overflow bucket */
+ fprintf(fp, "Hash buckets with >%2d entries %6ld (%3ld%%)\n",
+ i - 1, hash_bucket_lengths[i],
+ ((btc->count - total) * 100) /
+ atomic_read(&btc->count));
+}
+
+void
+btc_report(
+ FILE *fp,
+ const char *name,
+ struct xfs_mount *mp)
+{
+ int i;
+
+ if (!mp)
+ return;
+
+ fprintf(fp, "%s: Per-AG summary\n", name);
+ fprintf(fp, "AG\t|\t\tEntries\t\t|\t\tHash Table\n");
+ fprintf(fp, "\t|\tSupported\tUtilised\tActive\tSize\tHits\tMisses\tRatio\n");
+ for (i = 0; i < mp->m_sb.sb_agcount; i++) {
+ struct xfs_perag *pag = xfs_perag_get(mp, i);
+
+ btc_report_ag(fp, name, i, pag->pag_buf_hash);
+
+ xfs_perag_put(pag);
+ }
+}
+
+/* 2^63 + 2^61 - 2^57 + 2^54 - 2^51 - 2^18 + 1 */
+#define GOLDEN_RATIO_PRIME 0x9e37fffffffc0001UL
+#define CACHE_LINE_SIZE 64
+static unsigned int
+btchash(xfs_daddr_t blkno, unsigned int hashsize, unsigned int hashshift)
+{
+ uint64_t hashval = blkno;
+ uint64_t tmp;
+
+ tmp = hashval ^ (GOLDEN_RATIO_PRIME + hashval) / CACHE_LINE_SIZE;
+ tmp = tmp ^ ((tmp ^ GOLDEN_RATIO_PRIME) >> hashshift);
+ return tmp % hashsize;
+}
+
+struct xfs_buf *
+btc_node_find(
+ struct btcache *btc,
+ struct xfs_buf_map *map)
+{
+ struct xfs_buf *bp = NULL;
+ struct btcache_hash *hash;
+ struct list_head *head;
+ unsigned int hashidx;
+
+
+ hashidx = btchash(map->bm_bn, btc->hashsize, btc->hashshift);
+ hash = btc->hash + hashidx;
+ head = &hash->chain;
+
+ pthread_mutex_lock(&hash->lock);
+ list_for_each_entry(bp, head, b_hash) {
+ if (bp->b_bn != map->bm_bn)
+ continue;
+
+ if (bp->b_length != map->bm_len) {
+ /*
+ * found a block number match. If the range doesn't
+ * match, the only way this is allowed is if the buffer
+ * in the btc is stale and the transaction that made
+ * it stale has not yet committed. i.e. we are
+ * reallocating a busy extent. Skip this buffer and
+ * continue searching for an exact match.
+ */
+ ASSERT(bp->b_flags & XBF_STALE);
+ continue;
+ }
+ btc->hits++;
+ pthread_mutex_unlock(&hash->lock);
+ return bp;
+
+ }
+ btc->misses++;
+ pthread_mutex_unlock(&hash->lock);
+ return NULL;
+}
+
+void
+btc_node_insert(
+ struct btcache *btc,
+ struct xfs_buf *bp)
+{
+ struct btcache_hash *hash;
+ struct list_head *head;
+ unsigned int hashidx;
+
+ hashidx = btchash(bp->b_bn, btc->hashsize, btc->hashshift);
+ hash = btc->hash + hashidx;
+ head = &hash->chain;
+
+ pthread_mutex_lock(&hash->lock);
+ list_add(&bp->b_hash, head);
+ hash->count++;
+ atomic_inc(&btc->count);
+ pthread_mutex_unlock(&hash->lock);
+}
+
+void
+btc_node_remove(
+ struct btcache *btc,
+ struct xfs_buf *bp)
+{
+ struct btcache_hash *hash;
+ unsigned int hashidx;
+
+ hashidx = btchash(bp->b_bn, btc->hashsize, btc->hashshift);
+ hash = btc->hash + hashidx;
+
+ pthread_mutex_lock(&hash->lock);
+ list_del(&bp->b_hash);
+ hash->count--;
+ atomic_dec(&btc->count);
+ pthread_mutex_unlock(&hash->lock);
+}
@@ -44,9 +44,10 @@ struct xfs_buf_ops {
struct xfs_buf {
struct cache_node b_node;
- unsigned int b_flags;
+ struct list_head b_hash; /* will replace b_node */
xfs_daddr_t b_bn;
unsigned int b_length;
+ unsigned int b_flags;
struct xfs_buftarg *b_target;
pthread_mutex_t b_lock;
pthread_t b_holder;
@@ -13,6 +13,10 @@ struct xfs_buf_ops;
#define XFS_BUF_DADDR_NULL ((xfs_daddr_t) (-1LL))
+struct xfs_buf;
+struct xfs_buf_map;
+struct xfs_mount;
+
/*
* The xfs_buftarg contains 2 notions of "sector size" -
*
@@ -101,4 +105,45 @@ int xfs_bwrite(struct xfs_buf *bp);
struct xfs_buf *libxfs_getbufr(struct xfs_buftarg *btp, xfs_daddr_t blkno,
int bblen);
+/*
+ * Hash cache implementation
+ */
+/*
+ * xfs_db always writes changes immediately, and so we need to purge buffers
+ * when we get a buffer lookup mismatch due to reading the same block with a
+ * different buffer configuration.
+ *
+ * XXX: probably need to re-implement this
+ */
+#define CACHE_MISCOMPARE_PURGE (1 << 0)
+
+#define HASH_CACHE_RATIO 8
+
+struct btcache_hash {
+ struct list_head chain;
+ unsigned int count;
+ pthread_mutex_t lock;
+};
+
+struct btcache {
+ int flags; /* behavioural flags */
+ unsigned int maxcount; /* max cache nodes */
+ atomic_t count; /* count of nodes */
+ pthread_mutex_t lock; /* node count mutex */
+ unsigned int hashsize; /* hash bucket count */
+ unsigned int hashshift; /* hash key shift */
+ struct btcache_hash *hash; /* hash table buckets */
+ unsigned long long misses; /* cache misses */
+ unsigned long long hits; /* cache hits */
+ unsigned int max; /* max nodes ever used */
+};
+
+struct btcache *btc_init(unsigned int hashsize);
+void btc_destroy(struct btcache *cache);
+
+struct xfs_buf *btc_node_find(struct btcache *cache, struct xfs_buf_map *map);
+void btc_node_insert(struct btcache *cache, struct xfs_buf *bp);
+void btc_node_remove(struct btcache *cache, struct xfs_buf *bp);
+void btc_report(FILE *fp, const char *name, struct xfs_mount *mp);
+
#endif /* __XFS_BUFTARG_H */