From patchwork Wed Oct 10 10:07:26 2012 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Zhiyong Wu X-Patchwork-Id: 1572771 Return-Path: X-Original-To: patchwork-linux-btrfs@patchwork.kernel.org Delivered-To: patchwork-process-083081@patchwork2.kernel.org Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by patchwork2.kernel.org (Postfix) with ESMTP id 656DCDFB34 for ; Wed, 10 Oct 2012 10:12:04 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755600Ab2JJKLv (ORCPT ); Wed, 10 Oct 2012 06:11:51 -0400 Received: from e3.ny.us.ibm.com ([32.97.182.143]:34482 "EHLO e3.ny.us.ibm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755141Ab2JJKIw (ORCPT ); Wed, 10 Oct 2012 06:08:52 -0400 Received: from /spool/local by e3.ny.us.ibm.com with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted for from ; Wed, 10 Oct 2012 06:08:51 -0400 Received: from d01relay01.pok.ibm.com (9.56.227.233) by e3.ny.us.ibm.com (192.168.1.103) with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted; Wed, 10 Oct 2012 06:08:33 -0400 Received: from d01av01.pok.ibm.com (d01av01.pok.ibm.com [9.56.224.215]) by d01relay01.pok.ibm.com (8.13.8/8.13.8/NCO v10.0) with ESMTP id q9AA8XeZ155096; Wed, 10 Oct 2012 06:08:33 -0400 Received: from d01av01.pok.ibm.com (loopback [127.0.0.1]) by d01av01.pok.ibm.com (8.14.4/8.13.1/NCO v10.0 AVout) with ESMTP id q9AA8WiM026563; Wed, 10 Oct 2012 06:08:33 -0400 Received: from us.ibm.com (f15.cn.ibm.com [9.115.122.193]) by d01av01.pok.ibm.com (8.14.4/8.13.1/NCO v10.0 AVin) with SMTP id q9AA8RMs026187; Wed, 10 Oct 2012 06:08:28 -0400 Received: by us.ibm.com (sSMTP sendmail emulation); Wed, 10 Oct 2012 18:08:26 +0800 From: zwu.kernel@gmail.com To: linux-fsdevel@vger.kernel.org Cc: linux-ext4@vger.kernel.org, linux-btrfs@vger.kernel.org, linux-kernel@vger.kernel.org, linuxram@linux.vnet.ibm.com, viro@zeniv.linux.org.uk, david@fromorbit.com, dave@jikos.cz, tytso@mit.edu, cmm@us.ibm.com, Zhi Yong Wu Subject: [RFC v3 04/13] vfs: add function for collecting raw access info Date: Wed, 10 Oct 2012 18:07:26 +0800 Message-Id: <1349863655-29320-5-git-send-email-zwu.kernel@gmail.com> X-Mailer: git-send-email 1.7.6.5 In-Reply-To: <1349863655-29320-1-git-send-email-zwu.kernel@gmail.com> References: <1349863655-29320-1-git-send-email-zwu.kernel@gmail.com> x-cbid: 12101010-8974-0000-0000-00000F4E4071 X-IBM-ISS-SpamDetectors: X-IBM-ISS-DetailInfo: BY=3.00000294; HX=3.00000196; KW=3.00000007; PH=3.00000001; SC=3.00000008; SDB=6.00181223; UDB=6.00041045; UTC=2012-10-10 10:08:49 Sender: linux-btrfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-btrfs@vger.kernel.org From: Zhi Yong Wu Add some utils helpers to update access frequencies for one file or its range. Signed-off-by: Zhi Yong Wu --- fs/hot_tracking.c | 190 ++++++++++++++++++++++++++++++++++++++++++ fs/hot_tracking.h | 12 +++ include/linux/hot_tracking.h | 4 + 3 files changed, 206 insertions(+), 0 deletions(-) diff --git a/fs/hot_tracking.c b/fs/hot_tracking.c index 5fd993e..86c87c7 100644 --- a/fs/hot_tracking.c +++ b/fs/hot_tracking.c @@ -174,6 +174,196 @@ static void hot_inode_tree_exit(struct hot_info *root) } } +struct hot_inode_item +*hot_inode_item_find(struct hot_info *root, u64 ino) +{ + struct hot_inode_item *he; + int ret; + +again: + spin_lock(&root->lock); + he = radix_tree_lookup(&root->hot_inode_tree, ino); + if (he) { + kref_get(&he->hot_inode.refs); + spin_unlock(&root->lock); + return he; + } + spin_unlock(&root->lock); + + he = kmem_cache_zalloc(hot_inode_item_cachep, + GFP_KERNEL | GFP_NOFS); + if (!he) + return ERR_PTR(-ENOMEM); + + hot_inode_item_init(he, ino, &root->hot_inode_tree); + + ret = radix_tree_preload(GFP_NOFS & ~__GFP_HIGHMEM); + if (ret) { + kmem_cache_free(hot_inode_item_cachep, he); + return ERR_PTR(ret); + } + + spin_lock(&root->lock); + ret = radix_tree_insert(&root->hot_inode_tree, ino, he); + if (ret == -EEXIST) { + kmem_cache_free(hot_inode_item_cachep, he); + spin_unlock(&root->lock); + radix_tree_preload_end(); + goto again; + } + spin_unlock(&root->lock); + radix_tree_preload_end(); + + kref_get(&he->hot_inode.refs); + return he; +} + +static struct hot_range_item +*hot_range_item_find(struct hot_inode_item *he, + u32 start) +{ + struct hot_range_item *hr; + int ret; + +again: + spin_lock(&he->lock); + hr = radix_tree_lookup(&he->hot_range_tree, start); + if (hr) { + kref_get(&hr->hot_range.refs); + spin_unlock(&he->lock); + return hr; + } + spin_unlock(&he->lock); + + hr = kmem_cache_zalloc(hot_range_item_cachep, + GFP_KERNEL | GFP_NOFS); + if (!hr) + return ERR_PTR(-ENOMEM); + + hot_range_item_init(hr, start, he); + + ret = radix_tree_preload(GFP_NOFS & ~__GFP_HIGHMEM); + if (ret) { + kmem_cache_free(hot_range_item_cachep, hr); + return ERR_PTR(ret); + } + + spin_lock(&he->lock); + ret = radix_tree_insert(&he->hot_range_tree, start, hr); + if (ret == -EEXIST) { + kmem_cache_free(hot_range_item_cachep, hr); + spin_unlock(&he->lock); + radix_tree_preload_end(); + goto again; + } + spin_unlock(&he->lock); + radix_tree_preload_end(); + + kref_get(&hr->hot_range.refs); + return hr; +} + +/* + * This function does the actual work of updating the frequency numbers, + * whatever they turn out to be. FREQ_POWER determines how many atime + * deltas we keep track of (as a power of 2). So, setting it to anything above + * 16ish is probably overkill. Also, the higher the power, the more bits get + * right shifted out of the timestamp, reducing precision, so take note of that + * as well. + * + * The caller should have already locked freq_data's parent's spinlock. + * + * FREQ_POWER, defined immediately below, determines how heavily to weight + * the current frequency numbers against the newest access. For example, a value + * of 4 means that the new access information will be weighted 1/16th (ie 2^-4) + * as heavily as the existing frequency info. In essence, this is a kludged- + * together version of a weighted average, since we can't afford to keep all of + * the information that it would take to get a _real_ weighted average. + */ +static u64 hot_average_update(struct timespec old_atime, + struct timespec cur_time, u64 old_avg) +{ + struct timespec delta_ts; + u64 new_avg; + u64 new_delta; + + delta_ts = timespec_sub(cur_time, old_atime); + new_delta = timespec_to_ns(&delta_ts) >> FREQ_POWER; + + new_avg = (old_avg << FREQ_POWER) - old_avg + new_delta; + new_avg = new_avg >> FREQ_POWER; + + return new_avg; +} + +static void hot_freq_data_update(struct hot_freq_data *freq_data, bool write) +{ + struct timespec cur_time = current_kernel_time(); + + if (write) { + freq_data->nr_writes += 1; + freq_data->avg_delta_writes = hot_average_update( + freq_data->last_write_time, + cur_time, + freq_data->avg_delta_writes); + freq_data->last_write_time = cur_time; + } else { + freq_data->nr_reads += 1; + freq_data->avg_delta_reads = hot_average_update( + freq_data->last_read_time, + cur_time, + freq_data->avg_delta_reads); + freq_data->last_read_time = cur_time; + } +} + +/* + * Main function to update access frequency from read/writepage(s) hooks + */ +inline void hot_update_freqs(struct hot_info *root, + struct inode *inode, u64 start, + u64 len, int rw) +{ + struct hot_inode_item *he; + struct hot_range_item *hr; + u32 cur, end; + + if (!TRACK_THIS_INODE(inode) || (len == 0)) + return; + + he = hot_inode_item_find(root, inode->i_ino); + if (IS_ERR(he)) { + WARN_ON(1); + return; + } + + spin_lock(&he->hot_inode.lock); + hot_freq_data_update(&he->hot_inode.hot_freq_data, rw); + spin_unlock(&he->hot_inode.lock); + + /* + * Align ranges on RANGE_SIZE boundary + * to prevent proliferation of range structs + */ + end = (start + len + RANGE_SIZE - 1) >> RANGE_BITS; + for (cur = (start >> RANGE_BITS); cur < end; cur++) { + hr = hot_range_item_find(he, cur); + if (IS_ERR(hr)) { + WARN_ON(1); + hot_inode_item_put(he); + return; + } + + spin_lock(&hr->hot_range.lock); + hot_freq_data_update(&hr->hot_range.hot_freq_data, rw); + spin_unlock(&hr->hot_range.lock); + + hot_range_item_put(hr); + } + + hot_inode_item_put(he); +} + /* * Initialize kmem cache for hot_inode_item and hot_range_item. */ diff --git a/fs/hot_tracking.h b/fs/hot_tracking.h index 4e8aa77..37f69ee 100644 --- a/fs/hot_tracking.h +++ b/fs/hot_tracking.h @@ -19,6 +19,18 @@ /* values for hot_freq_data flags */ #define FREQ_DATA_TYPE_INODE (1 << 0) #define FREQ_DATA_TYPE_RANGE (1 << 1) +/* size of sub-file ranges */ +#define RANGE_BITS 20 +#define RANGE_SIZE (1 << RANGE_BITS) + +#define FREQ_POWER 4 + +struct hot_inode_item +*hot_inode_item_find(struct hot_info *root, u64 ino); +void hot_inode_item_put(struct hot_inode_item *he); +inline void hot_update_freqs(struct hot_info *root, + struct inode *inode, u64 start, + u64 len, int rw); void hot_track_init(struct super_block *sb); void hot_track_exit(struct super_block *sb); diff --git a/include/linux/hot_tracking.h b/include/linux/hot_tracking.h index 13aa54b..1e0aed5 100644 --- a/include/linux/hot_tracking.h +++ b/include/linux/hot_tracking.h @@ -75,4 +75,8 @@ extern struct hot_info *global_hot_tracking_info; extern void hot_track_init(struct super_block *sb); extern void hot_track_exit(struct super_block *sb); +extern inline void hot_update_freqs(struct hot_info *root, + struct inode *inode, u64 start, + u64 len, int rw); + #endif /* _LINUX_HOTTRACK_H */