@@ -2527,6 +2527,8 @@ xlog_write(
xlog_write_full(lv, ticket, iclog, &log_offset,
&len, &record_cnt, &data_cnt);
}
+ if (lv->lv_flags & XFS_LOG_VEC_DYNAMIC)
+ kvfree(lv->lv_iovecp);
}
ASSERT(len == 0);
@@ -6,6 +6,8 @@
#ifndef __XFS_LOG_H__
#define __XFS_LOG_H__
+#define XFS_LOG_VEC_DYNAMIC (1 << 0)
+
struct xfs_cil_ctx;
struct xfs_log_vec {
@@ -17,7 +19,8 @@ struct xfs_log_vec {
char *lv_buf; /* formatted buffer */
int lv_bytes; /* accounted space in buffer */
int lv_buf_len; /* aligned size of buffer */
- int lv_size; /* size of allocated lv */
+ int lv_size; /* size of allocated iovecp + buf */
+ int lv_flags; /* lv flags */
};
extern struct kmem_cache *xfs_log_vec_cache;
@@ -42,6 +45,7 @@ static inline void
xlog_finish_iovec(struct xfs_log_vec *lv, struct xfs_log_iovec *vec,
int data_len)
{
+ struct xfs_log_iovec *lvec = lv->lv_iovecp;
struct xlog_op_header *oph = vec->i_addr;
int len;
@@ -71,7 +75,7 @@ xlog_finish_iovec(struct xfs_log_vec *lv, struct xfs_log_iovec *vec,
vec->i_len = len;
/* Catch buffer overruns */
- ASSERT((void *)lv->lv_buf + lv->lv_bytes <= (void *)lv + lv->lv_size);
+ ASSERT((void *)lv->lv_buf + lv->lv_bytes <= (void *)lvec + lv->lv_size);
}
/*
@@ -219,8 +219,7 @@ static inline int
xlog_cil_iovec_space(
uint niovecs)
{
- return round_up((sizeof(struct xfs_log_vec) +
- niovecs * sizeof(struct xfs_log_iovec)),
+ return round_up(niovecs * sizeof(struct xfs_log_iovec),
sizeof(uint64_t));
}
@@ -279,6 +278,7 @@ xlog_cil_alloc_shadow_bufs(
list_for_each_entry(lip, &tp->t_items, li_trans) {
struct xfs_log_vec *lv;
+ struct xfs_log_iovec *lvec;
int niovecs = 0;
int nbytes = 0;
int buf_size;
@@ -330,8 +330,8 @@ xlog_cil_alloc_shadow_bufs(
* if we have no shadow buffer, or it is too small, we need to
* reallocate it.
*/
- if (!lip->li_lv_shadow ||
- buf_size > lip->li_lv_shadow->lv_size) {
+ lv = lip->li_lv_shadow;
+ if (!lv || buf_size > lv->lv_size) {
/*
* We free and allocate here as a realloc would copy
* unnecessary data. We don't use kvzalloc() for the
@@ -339,18 +339,24 @@ xlog_cil_alloc_shadow_bufs(
* the buffer, only the log vector header and the iovec
* storage.
*/
- kvfree(lip->li_lv_shadow);
- lv = xlog_kvmalloc(buf_size);
+ if (lv)
+ kvfree(lv->lv_iovecp);
+ else
+ lv = kmem_cache_alloc(xfs_log_vec_cache,
+ GFP_KERNEL | __GFP_NOFAIL);
- memset(lv, 0, xlog_cil_iovec_space(niovecs));
+ memset(lv, 0, sizeof(struct xfs_log_vec));
+ lvec = xlog_kvmalloc(buf_size);
+ memset(lvec, 0, xlog_cil_iovec_space(niovecs));
+ lv->lv_flags |= XFS_LOG_VEC_DYNAMIC;
INIT_LIST_HEAD(&lv->lv_list);
lv->lv_item = lip;
lv->lv_size = buf_size;
if (ordered)
lv->lv_buf_len = XFS_LOG_VEC_ORDERED;
else
- lv->lv_iovecp = (struct xfs_log_iovec *)&lv[1];
+ lv->lv_iovecp = lvec;
lip->li_lv_shadow = lv;
} else {
/* same or smaller, optimise common overwrite case */
@@ -366,9 +372,9 @@ xlog_cil_alloc_shadow_bufs(
lv->lv_niovecs = niovecs;
/* The allocated data region lies beyond the iovec region */
- lv->lv_buf = (char *)lv + xlog_cil_iovec_space(niovecs);
+ lv->lv_buf = (char *)lv->lv_iovecp +
+ xlog_cil_iovec_space(niovecs);
}
-
}
/*
@@ -502,7 +508,7 @@ xlog_cil_insert_format_items(
/* reset the lv buffer information for new formatting */
lv->lv_buf_len = 0;
lv->lv_bytes = 0;
- lv->lv_buf = (char *)lv +
+ lv->lv_buf = (char *)lv->lv_iovecp +
xlog_cil_iovec_space(lv->lv_niovecs);
} else {
/* switch to shadow buffer! */
@@ -703,7 +709,7 @@ xlog_cil_free_logvec(
while (!list_empty(lv_chain)) {
lv = list_first_entry(lv_chain, struct xfs_log_vec, lv_list);
list_del_init(&lv->lv_list);
- kvfree(lv);
+ kmem_cache_free(xfs_log_vec_cache, lv);
}
}
@@ -1544,7 +1550,8 @@ xlog_cil_process_intents(
set_bit(XFS_LI_WHITEOUT, &ilip->li_flags);
trace_xfs_cil_whiteout_mark(ilip);
len += ilip->li_lv->lv_bytes;
- kvfree(ilip->li_lv);
+ kvfree(ilip->li_lv->lv_iovecp);
+ kmem_cache_free(xfs_log_vec_cache, ilip->li_lv);
ilip->li_lv = NULL;
xfs_trans_del_item(lip);