Commit 5ff7fb77 authored by Juan Quintela's avatar Juan Quintela
Browse files

memory: move bitmap synchronization to its own function



We want to have all the functions that handle directly the dirty
bitmap near.  We will change it later.

Signed-off-by: default avatarJuan Quintela <quintela@redhat.com>
Reviewed-by: default avatarOrit Wasserman <owasserm@redhat.com>
parent c9dd46fc
Loading
Loading
Loading
Loading
+31 −0
Original line number Diff line number Diff line
@@ -79,6 +79,37 @@ static inline void cpu_physical_memory_set_dirty_range(ram_addr_t start,
    xen_modified_memory(start, length);
}

static inline void cpu_physical_memory_set_dirty_lebitmap(unsigned long *bitmap,
                                                          ram_addr_t start,
                                                          ram_addr_t pages)
{
    unsigned int i, j;
    unsigned long page_number, c;
    hwaddr addr;
    ram_addr_t ram_addr;
    unsigned int len = (pages + HOST_LONG_BITS - 1) / HOST_LONG_BITS;
    unsigned long hpratio = getpagesize() / TARGET_PAGE_SIZE;

    /*
     * bitmap-traveling is faster than memory-traveling (for addr...)
     * especially when most of the memory is not dirty.
     */
    for (i = 0; i < len; i++) {
        if (bitmap[i] != 0) {
            c = leul_to_cpu(bitmap[i]);
            do {
                j = ffsl(c) - 1;
                c &= ~(1ul << j);
                page_number = (i * HOST_LONG_BITS + j) * hpratio;
                addr = page_number * TARGET_PAGE_SIZE;
                ram_addr = start + addr;
                cpu_physical_memory_set_dirty_range(ram_addr,
                                                    TARGET_PAGE_SIZE * hpratio);
            } while (c != 0);
        }
    }
}

static inline void cpu_physical_memory_clear_dirty_range(ram_addr_t start,
                                                         ram_addr_t length,
                                                         unsigned client)
+2 −25
Original line number Diff line number Diff line
@@ -380,33 +380,10 @@ static int kvm_set_migration_log(int enable)
static int kvm_get_dirty_pages_log_range(MemoryRegionSection *section,
                                         unsigned long *bitmap)
{
    unsigned int i, j;
    unsigned long page_number, c;
    hwaddr addr;
    ram_addr_t start = section->offset_within_region + section->mr->ram_addr;
    ram_addr_t ram_addr;
    unsigned int pages = int128_get64(section->size) / getpagesize();
    unsigned int len = (pages + HOST_LONG_BITS - 1) / HOST_LONG_BITS;
    unsigned long hpratio = getpagesize() / TARGET_PAGE_SIZE;
    ram_addr_t pages = int128_get64(section->size) / getpagesize();

    /*
     * bitmap-traveling is faster than memory-traveling (for addr...)
     * especially when most of the memory is not dirty.
     */
    for (i = 0; i < len; i++) {
        if (bitmap[i] != 0) {
            c = leul_to_cpu(bitmap[i]);
            do {
                j = ffsl(c) - 1;
                c &= ~(1ul << j);
                page_number = (i * HOST_LONG_BITS + j) * hpratio;
                addr = page_number * TARGET_PAGE_SIZE;
                ram_addr = start + addr;
                cpu_physical_memory_set_dirty_range(ram_addr,
                                                    TARGET_PAGE_SIZE * hpratio);
            } while (c != 0);
        }
    }
    cpu_physical_memory_set_dirty_lebitmap(bitmap, start, pages);
    return 0;
}