Message ID | 1457412306-18940-13-git-send-email-famz@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On 08.03.2016 05:45, Fam Zheng wrote: > From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> > > Functions to serialize / deserialize(restore) HBitmap. HBitmap should be > saved to linear sequence of bits independently of endianness and bitmap > array element (unsigned long) size. Therefore Little Endian is chosen. > > These functions are appropriate for dirty bitmap migration, restoring > the bitmap in several steps is available. To save performance, every > step writes only the last level of the bitmap. All other levels are > restored by hbitmap_deserialize_finish() as a last step of restoring. > So, HBitmap is inconsistent while restoring. > > Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> > [Fix left shift operand to 1UL; add "finish" parameter. - Fam] > Signed-off-by: Fam Zheng <famz@redhat.com> > --- > include/qemu/hbitmap.h | 79 ++++++++++++++++++++++++++++ > util/hbitmap.c | 137 +++++++++++++++++++++++++++++++++++++++++++++++++ > 2 files changed, 216 insertions(+) [...] > diff --git a/util/hbitmap.c b/util/hbitmap.c > index 2d3d04c..5f02c17 100644 > --- a/util/hbitmap.c > +++ b/util/hbitmap.c > @@ -395,6 +395,143 @@ bool hbitmap_get(const HBitmap *hb, uint64_t item) > return (hb->levels[HBITMAP_LEVELS - 1][pos >> BITS_PER_LEVEL] & bit) != 0; > } > > +uint64_t hbitmap_serialization_granularity(const HBitmap *hb) > +{ > + /* Require at least 64 bit granularity to be safe on both 64 bit and 32 bit > + * hosts. */ > + return 64 << hb->granularity; > +} > + > +/* Start should be aligned to serialization granularity, chunk size should be > + * aligned to serialization granularity too, except for last chunk. > + */ > +static void serialization_chunk(const HBitmap *hb, > + uint64_t start, uint64_t count, > + unsigned long **first_el, size_t *el_count) > +{ > + uint64_t last = start + count - 1; > + uint64_t gran = hbitmap_serialization_granularity(hb); > + > + assert((start & (gran - 1)) == 0); > + assert((last >> hb->granularity) < hb->size); > + if ((last >> hb->granularity) != hb->size - 1) { > + assert((count & (gran - 1)) == 0); > + } > + > + start = (start >> hb->granularity) >> BITS_PER_LEVEL; > + last = (last >> hb->granularity) >> BITS_PER_LEVEL; > + > + *first_el = &hb->levels[HBITMAP_LEVELS - 1][start]; > + *el_count = last - start + 1; > +} > + > +uint64_t hbitmap_serialization_size(const HBitmap *hb, > + uint64_t start, uint64_t count) > +{ > + uint64_t el_count; > + unsigned long *cur; > + > + if (!count) { > + return 0; > + } > + serialization_chunk(hb, start, count, &cur, &el_count); > + > + return el_count * sizeof(unsigned long); > +} > + > +void hbitmap_serialize_part(const HBitmap *hb, uint8_t *buf, > + uint64_t start, uint64_t count) > +{ > + uint64_t el_count; > + unsigned long *cur, *end; > + > + if (!count) { > + return; > + } > + serialization_chunk(hb, start, count, &cur, &el_count); > + end = cur + el_count; > + > + while (cur != end) { > + unsigned long el = > + (BITS_PER_LONG == 32 ? cpu_to_le32(*cur) : cpu_to_le64(*cur)); Looks a bit fishy, but I can't come up with anything better. (Other than adding cpu_to_leul(); we already do have leul_to_cpu(), so that wouldn't be too far off.) > + > + memcpy(buf, &el, sizeof(el)); One could have used cpu_to_le32/64w((uint32/64_t *)buf, *cur); instead. Maybe I'd like the following better: #if BITS_PER_LONG == 32 cpu_to_le32w((uint32_t *)buf, *cur); #elif BITS_PER_LONG == 64 cpu_to_le64w((uint64_t *)buf, *cur); #else #error Unknown long size #endif Or just #else /* BITS_PER_LONG == 64 */ instead of the #elif. I think that's safe to assume. > + buf += sizeof(el); > + cur++; > + } > +} > + > +void hbitmap_deserialize_part(HBitmap *hb, uint8_t *buf, > + uint64_t start, uint64_t count, > + bool finish) > +{ > + uint64_t el_count; > + unsigned long *cur, *end; > + > + if (!count) { > + return; > + } > + serialization_chunk(hb, start, count, &cur, &el_count); > + end = cur + el_count; > + > + while (cur != end) { > + memcpy(cur, buf, sizeof(*cur)); > + > + if (BITS_PER_LONG == 32) { > + le32_to_cpus((uint32_t *)cur); > + } else { > + le64_to_cpus((uint64_t *)cur); > + } Here, I'd definitely like that variant better, i.e. #if BITS_PER_LONG == 32 le32_to_cpuw(cur, *(uint32_t *)buf); #else /* BITS_PER_LONG == 64 */ le64_to_cpuw(cur, *(uint64_t *)buf); #endif Unless a language lawyer NACKs this because the pointer cast violates strict aliasing. If so, I still strongly recommend replacing the if by an #if, not least because this saves us the pointer cast on cur. (Or does it? Maybe one still needs to explicitly cast an unsigned long * to uint32_t * even if both have the same size. However, in that case we can still do *cur = le32/64_to_cpu(*cur). The above would then become: #if BITS_PER_LONG == 32 *cur = le32_to_cpu(*(uint32_t *)buf); #else /* BITS_PER_LONG == 64 */ *cur = le64_to_cpu(*(uint64_t *)buf); #endif ) > + > + buf += sizeof(unsigned long); > + cur++; > + } > + if (finish) { > + hbitmap_deserialize_finish(hb); > + } > +} > + > +void hbitmap_deserialize_zeroes(HBitmap *hb, uint64_t start, uint64_t count, > + bool finish) > +{ > + uint64_t el_count; > + unsigned long *first; > + > + if (!count) { > + return; > + } > + serialization_chunk(hb, start, count, &first, &el_count); > + > + memset(first, 0, el_count * sizeof(unsigned long)); > + if (finish) { > + hbitmap_deserialize_finish(hb); > + } > +} > + > +void hbitmap_deserialize_finish(HBitmap *bitmap) > +{ > + int64_t i, size, prev_size; > + int lev; > + > + /* restore levels starting from penultimate to zero level, assuming > + * that the last level is ok */ > + size = MAX((bitmap->size + BITS_PER_LONG - 1) >> BITS_PER_LEVEL, 1); > + for (lev = HBITMAP_LEVELS - 1; lev-- > 0; ) { I'm always amazed of how much this notation (x-- > 0, or x --> 0, even worse) confuses me. Or rather I'm always amazed of how much it confuses me and how much other people love it (as opposed to me, who's rather unhappy every time he doesn't use all of the three "parameters" for a for loop). This time I wondered why we start at HBITMAP_LEVELS - 1 -- that's the ultimate and not the penultimate level after all. But of course, the loop condition is evaluated before the loop is executed for the first time, so we will indeed start at HBITMAP_LEVELS - 2. So, yeah, this is correct. > + prev_size = size; > + size = MAX((size + BITS_PER_LONG - 1) >> BITS_PER_LEVEL, 1); > + memset(bitmap->levels[lev], 0, size * sizeof(unsigned long)); > + > + for (i = 0; i < prev_size; ++i) { I hate using stand-alone pre-increment for scalars with a passion. Unfortunately, checkpatch does not. Reviewed-by: Max Reitz <mreitz@redhat.com> (All of the above are me just nagging and giving optional suggestions.) > + if (bitmap->levels[lev + 1][i]) { > + bitmap->levels[lev][i >> BITS_PER_LEVEL] |= > + 1UL << (i & (BITS_PER_LONG - 1)); > + } > + } > + } > + > + bitmap->levels[0][0] |= 1UL << (BITS_PER_LONG - 1); > +} > + > void hbitmap_free(HBitmap *hb) > { > unsigned i; >
On 11.03.2016 19:27, Max Reitz wrote: > On 08.03.2016 05:45, Fam Zheng wrote: >> From: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> >> >> Functions to serialize / deserialize(restore) HBitmap. HBitmap should be >> saved to linear sequence of bits independently of endianness and bitmap >> array element (unsigned long) size. Therefore Little Endian is chosen. >> >> These functions are appropriate for dirty bitmap migration, restoring >> the bitmap in several steps is available. To save performance, every >> step writes only the last level of the bitmap. All other levels are >> restored by hbitmap_deserialize_finish() as a last step of restoring. >> So, HBitmap is inconsistent while restoring. >> >> Signed-off-by: Vladimir Sementsov-Ogievskiy <vsementsov@virtuozzo.com> >> [Fix left shift operand to 1UL; add "finish" parameter. - Fam] >> Signed-off-by: Fam Zheng <famz@redhat.com> >> --- >> include/qemu/hbitmap.h | 79 ++++++++++++++++++++++++++++ >> util/hbitmap.c | 137 +++++++++++++++++++++++++++++++++++++++++++++++++ >> 2 files changed, 216 insertions(+) > [...] > >> diff --git a/util/hbitmap.c b/util/hbitmap.c >> index 2d3d04c..5f02c17 100644 >> --- a/util/hbitmap.c >> +++ b/util/hbitmap.c >> @@ -395,6 +395,143 @@ bool hbitmap_get(const HBitmap *hb, uint64_t item) >> return (hb->levels[HBITMAP_LEVELS - 1][pos >> BITS_PER_LEVEL] & bit) != 0; >> } >> >> +uint64_t hbitmap_serialization_granularity(const HBitmap *hb) >> +{ >> + /* Require at least 64 bit granularity to be safe on both 64 bit and 32 bit >> + * hosts. */ >> + return 64 << hb->granularity; >> +} >> + >> +/* Start should be aligned to serialization granularity, chunk size should be >> + * aligned to serialization granularity too, except for last chunk. >> + */ >> +static void serialization_chunk(const HBitmap *hb, >> + uint64_t start, uint64_t count, >> + unsigned long **first_el, size_t *el_count) >> +{ >> + uint64_t last = start + count - 1; >> + uint64_t gran = hbitmap_serialization_granularity(hb); >> + >> + assert((start & (gran - 1)) == 0); >> + assert((last >> hb->granularity) < hb->size); >> + if ((last >> hb->granularity) != hb->size - 1) { >> + assert((count & (gran - 1)) == 0); >> + } >> + >> + start = (start >> hb->granularity) >> BITS_PER_LEVEL; >> + last = (last >> hb->granularity) >> BITS_PER_LEVEL; >> + >> + *first_el = &hb->levels[HBITMAP_LEVELS - 1][start]; >> + *el_count = last - start + 1; >> +} >> + >> +uint64_t hbitmap_serialization_size(const HBitmap *hb, >> + uint64_t start, uint64_t count) >> +{ >> + uint64_t el_count; >> + unsigned long *cur; >> + >> + if (!count) { >> + return 0; >> + } >> + serialization_chunk(hb, start, count, &cur, &el_count); >> + >> + return el_count * sizeof(unsigned long); >> +} >> + >> +void hbitmap_serialize_part(const HBitmap *hb, uint8_t *buf, >> + uint64_t start, uint64_t count) >> +{ >> + uint64_t el_count; >> + unsigned long *cur, *end; >> + >> + if (!count) { >> + return; >> + } >> + serialization_chunk(hb, start, count, &cur, &el_count); >> + end = cur + el_count; >> + >> + while (cur != end) { >> + unsigned long el = >> + (BITS_PER_LONG == 32 ? cpu_to_le32(*cur) : cpu_to_le64(*cur)); > Looks a bit fishy, but I can't come up with anything better. > > (Other than adding cpu_to_leul(); we already do have leul_to_cpu(), so > that wouldn't be too far off.) > >> + >> + memcpy(buf, &el, sizeof(el)); > One could have used cpu_to_le32/64w((uint32/64_t *)buf, *cur); instead. > > Maybe I'd like the following better: > > #if BITS_PER_LONG == 32 > cpu_to_le32w((uint32_t *)buf, *cur); > #elif BITS_PER_LONG == 64 > cpu_to_le64w((uint64_t *)buf, *cur); > #else > #error Unknown long size > #endif > > Or just > > #else /* BITS_PER_LONG == 64 */ > > instead of the #elif. I think that's safe to assume. I think (BITS_PER_LONG == 32 ? .. : ..) should be optimised away by compiler in any case.. However adding cpu_to_leul should be better. > >> + buf += sizeof(el); >> + cur++; >> + } >> +} >> + >> +void hbitmap_deserialize_part(HBitmap *hb, uint8_t *buf, >> + uint64_t start, uint64_t count, >> + bool finish) >> +{ >> + uint64_t el_count; >> + unsigned long *cur, *end; >> + >> + if (!count) { >> + return; >> + } >> + serialization_chunk(hb, start, count, &cur, &el_count); >> + end = cur + el_count; >> + >> + while (cur != end) { >> + memcpy(cur, buf, sizeof(*cur)); >> + >> + if (BITS_PER_LONG == 32) { >> + le32_to_cpus((uint32_t *)cur); >> + } else { >> + le64_to_cpus((uint64_t *)cur); >> + } > Here, I'd definitely like that variant better, i.e. > > #if BITS_PER_LONG == 32 > le32_to_cpuw(cur, *(uint32_t *)buf); > #else /* BITS_PER_LONG == 64 */ > le64_to_cpuw(cur, *(uint64_t *)buf); > #endif > > Unless a language lawyer NACKs this because the pointer cast violates > strict aliasing. > > If so, I still strongly recommend replacing the if by an #if, not least > because this saves us the pointer cast on cur. > > (Or does it? Maybe one still needs to explicitly cast an unsigned long * > to uint32_t * even if both have the same size. However, in that case we > can still do *cur = le32/64_to_cpu(*cur). The above would then become: > > #if BITS_PER_LONG == 32 > *cur = le32_to_cpu(*(uint32_t *)buf); > #else /* BITS_PER_LONG == 64 */ > *cur = le64_to_cpu(*(uint64_t *)buf); > #endif > > ) > >> + >> + buf += sizeof(unsigned long); >> + cur++; >> + } >> + if (finish) { >> + hbitmap_deserialize_finish(hb); >> + } >> +} >> + >> +void hbitmap_deserialize_zeroes(HBitmap *hb, uint64_t start, uint64_t count, >> + bool finish) >> +{ >> + uint64_t el_count; >> + unsigned long *first; >> + >> + if (!count) { >> + return; >> + } >> + serialization_chunk(hb, start, count, &first, &el_count); >> + >> + memset(first, 0, el_count * sizeof(unsigned long)); >> + if (finish) { >> + hbitmap_deserialize_finish(hb); >> + } >> +} >> + >> +void hbitmap_deserialize_finish(HBitmap *bitmap) >> +{ >> + int64_t i, size, prev_size; >> + int lev; >> + >> + /* restore levels starting from penultimate to zero level, assuming >> + * that the last level is ok */ >> + size = MAX((bitmap->size + BITS_PER_LONG - 1) >> BITS_PER_LEVEL, 1); >> + for (lev = HBITMAP_LEVELS - 1; lev-- > 0; ) { > I'm always amazed of how much this notation (x-- > 0, or x --> 0, even > worse) confuses me. Or rather I'm always amazed of how much it confuses > me and how much other people love it (as opposed to me, who's rather > unhappy every time he doesn't use all of the three "parameters" for a > for loop). > > This time I wondered why we start at HBITMAP_LEVELS - 1 -- that's the > ultimate and not the penultimate level after all. But of course, the > loop condition is evaluated before the loop is executed for the first > time, so we will indeed start at HBITMAP_LEVELS - 2. > > So, yeah, this is correct. This strange notation is used in several places in hbitmap.c, so, I've reused it) I don't like it, but to fix it it should be fixed everywhere in the file as a separate patch I think.. > >> + prev_size = size; >> + size = MAX((size + BITS_PER_LONG - 1) >> BITS_PER_LEVEL, 1); >> + memset(bitmap->levels[lev], 0, size * sizeof(unsigned long)); >> + >> + for (i = 0; i < prev_size; ++i) { > I hate using stand-alone pre-increment for scalars with a passion. I just prefer to write it in the same way for scalars and iterators, for c and c++. > > Unfortunately, checkpatch does not. > > Reviewed-by: Max Reitz <mreitz@redhat.com> > > (All of the above are me just nagging and giving optional suggestions.) > >> + if (bitmap->levels[lev + 1][i]) { >> + bitmap->levels[lev][i >> BITS_PER_LEVEL] |= >> + 1UL << (i & (BITS_PER_LONG - 1)); >> + } >> + } >> + } >> + >> + bitmap->levels[0][0] |= 1UL << (BITS_PER_LONG - 1); >> +} >> + >> void hbitmap_free(HBitmap *hb) >> { >> unsigned i; >>
diff --git a/include/qemu/hbitmap.h b/include/qemu/hbitmap.h index f8ed058..26cac7d 100644 --- a/include/qemu/hbitmap.h +++ b/include/qemu/hbitmap.h @@ -146,6 +146,85 @@ void hbitmap_reset_all(HBitmap *hb); bool hbitmap_get(const HBitmap *hb, uint64_t item); /** + * hbitmap_serialization_granularity: + * @hb: HBitmap to operate on. + * + * Granularity of serialization chunks, used by other serialization functions. + * For every chunk: + * 1. Chunk start should be aligned to this granularity. + * 2. Chunk size should be aligned too, except for last chunk (for which + * start + count == hb->size) + */ +uint64_t hbitmap_serialization_granularity(const HBitmap *hb); + +/** + * hbitmap_serialization_size: + * @hb: HBitmap to operate on. + * @start: Starting bit + * @count: Number of bits + * + * Return number of bytes hbitmap_(de)serialize_part needs + */ +uint64_t hbitmap_serialization_size(const HBitmap *hb, + uint64_t start, uint64_t count); + +/** + * hbitmap_serialize_part + * @hb: HBitmap to operate on. + * @buf: Buffer to store serialized bitmap. + * @start: First bit to store. + * @count: Number of bits to store. + * + * Stores HBitmap data corresponding to given region. The format of saved data + * is linear sequence of bits, so it can be used by hbitmap_deserialize_part + * independently of endianness and size of HBitmap level array elements + */ +void hbitmap_serialize_part(const HBitmap *hb, uint8_t *buf, + uint64_t start, uint64_t count); + +/** + * hbitmap_deserialize_part + * @hb: HBitmap to operate on. + * @buf: Buffer to restore bitmap data from. + * @start: First bit to restore. + * @count: Number of bits to restore. + * @finish: Whether to call hbitmap_deserialize_finish automatically. + * + * Restores HBitmap data corresponding to given region. The format is the same + * as for hbitmap_serialize_part. + * + * If @finish is false, caller must call hbitmap_serialize_finish before using + * the bitmap. + */ +void hbitmap_deserialize_part(HBitmap *hb, uint8_t *buf, + uint64_t start, uint64_t count, + bool finish); + +/** + * hbitmap_deserialize_zeroes + * @hb: HBitmap to operate on. + * @start: First bit to restore. + * @count: Number of bits to restore. + * @finish: Whether to call hbitmap_deserialize_finish automatically. + * + * Fills the bitmap with zeroes. + * + * If @finish is false, caller must call hbitmap_serialize_finish before using + * the bitmap. + */ +void hbitmap_deserialize_zeroes(HBitmap *hb, uint64_t start, uint64_t count, + bool finish); + +/** + * hbitmap_deserialize_finish + * @hb: HBitmap to operate on. + * + * Repair HBitmap after calling hbitmap_deserialize_data. Actually, all HBitmap + * layers are restored here. + */ +void hbitmap_deserialize_finish(HBitmap *hb); + +/** * hbitmap_free: * @hb: HBitmap to operate on. * diff --git a/util/hbitmap.c b/util/hbitmap.c index 2d3d04c..5f02c17 100644 --- a/util/hbitmap.c +++ b/util/hbitmap.c @@ -395,6 +395,143 @@ bool hbitmap_get(const HBitmap *hb, uint64_t item) return (hb->levels[HBITMAP_LEVELS - 1][pos >> BITS_PER_LEVEL] & bit) != 0; } +uint64_t hbitmap_serialization_granularity(const HBitmap *hb) +{ + /* Require at least 64 bit granularity to be safe on both 64 bit and 32 bit + * hosts. */ + return 64 << hb->granularity; +} + +/* Start should be aligned to serialization granularity, chunk size should be + * aligned to serialization granularity too, except for last chunk. + */ +static void serialization_chunk(const HBitmap *hb, + uint64_t start, uint64_t count, + unsigned long **first_el, size_t *el_count) +{ + uint64_t last = start + count - 1; + uint64_t gran = hbitmap_serialization_granularity(hb); + + assert((start & (gran - 1)) == 0); + assert((last >> hb->granularity) < hb->size); + if ((last >> hb->granularity) != hb->size - 1) { + assert((count & (gran - 1)) == 0); + } + + start = (start >> hb->granularity) >> BITS_PER_LEVEL; + last = (last >> hb->granularity) >> BITS_PER_LEVEL; + + *first_el = &hb->levels[HBITMAP_LEVELS - 1][start]; + *el_count = last - start + 1; +} + +uint64_t hbitmap_serialization_size(const HBitmap *hb, + uint64_t start, uint64_t count) +{ + uint64_t el_count; + unsigned long *cur; + + if (!count) { + return 0; + } + serialization_chunk(hb, start, count, &cur, &el_count); + + return el_count * sizeof(unsigned long); +} + +void hbitmap_serialize_part(const HBitmap *hb, uint8_t *buf, + uint64_t start, uint64_t count) +{ + uint64_t el_count; + unsigned long *cur, *end; + + if (!count) { + return; + } + serialization_chunk(hb, start, count, &cur, &el_count); + end = cur + el_count; + + while (cur != end) { + unsigned long el = + (BITS_PER_LONG == 32 ? cpu_to_le32(*cur) : cpu_to_le64(*cur)); + + memcpy(buf, &el, sizeof(el)); + buf += sizeof(el); + cur++; + } +} + +void hbitmap_deserialize_part(HBitmap *hb, uint8_t *buf, + uint64_t start, uint64_t count, + bool finish) +{ + uint64_t el_count; + unsigned long *cur, *end; + + if (!count) { + return; + } + serialization_chunk(hb, start, count, &cur, &el_count); + end = cur + el_count; + + while (cur != end) { + memcpy(cur, buf, sizeof(*cur)); + + if (BITS_PER_LONG == 32) { + le32_to_cpus((uint32_t *)cur); + } else { + le64_to_cpus((uint64_t *)cur); + } + + buf += sizeof(unsigned long); + cur++; + } + if (finish) { + hbitmap_deserialize_finish(hb); + } +} + +void hbitmap_deserialize_zeroes(HBitmap *hb, uint64_t start, uint64_t count, + bool finish) +{ + uint64_t el_count; + unsigned long *first; + + if (!count) { + return; + } + serialization_chunk(hb, start, count, &first, &el_count); + + memset(first, 0, el_count * sizeof(unsigned long)); + if (finish) { + hbitmap_deserialize_finish(hb); + } +} + +void hbitmap_deserialize_finish(HBitmap *bitmap) +{ + int64_t i, size, prev_size; + int lev; + + /* restore levels starting from penultimate to zero level, assuming + * that the last level is ok */ + size = MAX((bitmap->size + BITS_PER_LONG - 1) >> BITS_PER_LEVEL, 1); + for (lev = HBITMAP_LEVELS - 1; lev-- > 0; ) { + prev_size = size; + size = MAX((size + BITS_PER_LONG - 1) >> BITS_PER_LEVEL, 1); + memset(bitmap->levels[lev], 0, size * sizeof(unsigned long)); + + for (i = 0; i < prev_size; ++i) { + if (bitmap->levels[lev + 1][i]) { + bitmap->levels[lev][i >> BITS_PER_LEVEL] |= + 1UL << (i & (BITS_PER_LONG - 1)); + } + } + } + + bitmap->levels[0][0] |= 1UL << (BITS_PER_LONG - 1); +} + void hbitmap_free(HBitmap *hb) { unsigned i;