/*
 * QEMU Hyper-V Dynamic Memory Protocol driver
 *
 * Copyright (C) 2020-2023 Oracle and/or its affiliates.
 *
 * This work is licensed under the terms of the GNU GPL, version 2 or later.
 * See the COPYING file in the top-level directory.
 */

#include "qemu/osdep.h"
#include "hv-balloon-internal.h"
#include "hv-balloon-our_range_memslots.h"
#include "trace.h"

/* OurRange */
static void our_range_init(OurRange *our_range, uint64_t start, uint64_t count)
{
    assert(count <= UINT64_MAX - start);
    our_range->range.start = start;
    our_range->range.count = count;

    hvb_page_range_tree_init(&our_range->removed_guest);
    hvb_page_range_tree_init(&our_range->removed_both);

    /* mark the whole range as unused but for potential use */
    our_range->added = 0;
    our_range->unusable_tail = 0;
}

static void our_range_destroy(OurRange *our_range)
{
    hvb_page_range_tree_destroy(&our_range->removed_guest);
    hvb_page_range_tree_destroy(&our_range->removed_both);
}

void hvb_our_range_clear_removed_trees(OurRange *our_range)
{
    hvb_page_range_tree_destroy(&our_range->removed_guest);
    hvb_page_range_tree_destroy(&our_range->removed_both);
    hvb_page_range_tree_init(&our_range->removed_guest);
    hvb_page_range_tree_init(&our_range->removed_both);
}

void hvb_our_range_mark_added(OurRange *our_range, uint64_t additional_size)
{
    assert(additional_size <= UINT64_MAX - our_range->added);

    our_range->added += additional_size;

    assert(our_range->added <= UINT64_MAX - our_range->unusable_tail);
    assert(our_range->added + our_range->unusable_tail <=
           our_range->range.count);
}

/* OurRangeMemslots */
static void our_range_memslots_init_slots(OurRangeMemslots *our_range,
                                          MemoryRegion *backing_mr,
                                          Object *memslot_owner)
{
    OurRangeMemslotsSlots *memslots = &our_range->slots;
    unsigned int idx;
    uint64_t memslot_offset;

    assert(memslots->count > 0);
    memslots->slots = g_new0(MemoryRegion, memslots->count);

    /* Initialize our memslots, but don't map them yet. */
    assert(memslots->size_each > 0);
    for (idx = 0, memslot_offset = 0; idx < memslots->count;
         idx++, memslot_offset += memslots->size_each) {
        uint64_t memslot_size;
        g_autofree char *name = NULL;

        /* The size of the last memslot might be smaller. */
        if (idx == memslots->count - 1) {
            uint64_t region_size;

            assert(our_range->mr);
            region_size = memory_region_size(our_range->mr);
            memslot_size = region_size - memslot_offset;
        } else {
            memslot_size = memslots->size_each;
        }

        name = g_strdup_printf("memslot-%u", idx);
        memory_region_init_alias(&memslots->slots[idx], memslot_owner, name,
                                 backing_mr, memslot_offset, memslot_size);
        /*
         * We want to be able to atomically and efficiently activate/deactivate
         * individual memslots without affecting adjacent memslots in memory
         * notifiers.
         */
        memory_region_set_unmergeable(&memslots->slots[idx], true);
    }

    memslots->mapped_count = 0;
}

OurRangeMemslots *hvb_our_range_memslots_new(uint64_t addr,
                                             MemoryRegion *parent_mr,
                                             MemoryRegion *backing_mr,
                                             Object *memslot_owner,
                                             unsigned int memslot_count,
                                             uint64_t memslot_size)
{
    OurRangeMemslots *our_range;

    our_range = g_malloc(sizeof(*our_range));
    our_range_init(&our_range->range,
                   addr / HV_BALLOON_PAGE_SIZE,
                   memory_region_size(parent_mr) / HV_BALLOON_PAGE_SIZE);
    our_range->slots.size_each = memslot_size;
    our_range->slots.count = memslot_count;
    our_range->mr = parent_mr;
    our_range_memslots_init_slots(our_range, backing_mr, memslot_owner);

    return our_range;
}

static void our_range_memslots_free_memslots(OurRangeMemslots *our_range)
{
    OurRangeMemslotsSlots *memslots = &our_range->slots;
    unsigned int idx;
    uint64_t offset;

    memory_region_transaction_begin();
    for (idx = 0, offset = 0; idx < memslots->mapped_count;
         idx++, offset += memslots->size_each) {
        trace_hv_balloon_unmap_slot(idx, memslots->count, offset);
        assert(memory_region_is_mapped(&memslots->slots[idx]));
        memory_region_del_subregion(our_range->mr, &memslots->slots[idx]);
    }
    memory_region_transaction_commit();

    for (idx = 0; idx < memslots->count; idx++) {
        object_unparent(OBJECT(&memslots->slots[idx]));
    }

    g_clear_pointer(&our_range->slots.slots, g_free);
}

void hvb_our_range_memslots_free(OurRangeMemslots *our_range)
{
    OurRangeMemslotsSlots *memslots = &our_range->slots;
    MemoryRegion *hostmem_mr;
    RAMBlock *rb;

    assert(our_range->slots.count > 0);
    assert(our_range->slots.slots);

    hostmem_mr = memslots->slots[0].alias;
    rb = hostmem_mr->ram_block;
    ram_block_discard_range(rb, 0, qemu_ram_get_used_length(rb));

    our_range_memslots_free_memslots(our_range);
    our_range_destroy(&our_range->range);
    g_free(our_range);
}

void hvb_our_range_memslots_ensure_mapped_additional(OurRangeMemslots *our_range,
                                                     uint64_t additional_map_size)
{
    OurRangeMemslotsSlots *memslots = &our_range->slots;
    uint64_t total_map_size;
    unsigned int idx;
    uint64_t offset;

    total_map_size = (our_range->range.added + additional_map_size) *
        HV_BALLOON_PAGE_SIZE;
    idx = memslots->mapped_count;
    assert(memslots->size_each > 0);
    offset = idx * memslots->size_each;

    /*
     * Activate all memslots covered by the newly added region in a single
     * transaction.
     */
    memory_region_transaction_begin();
    for ( ; idx < memslots->count;
          idx++, offset += memslots->size_each) {
        /*
         * If this memslot starts beyond or at the end of the range to map so
         * does every next one.
         */
        if (offset >= total_map_size) {
            break;
        }

        /*
         * Instead of enabling/disabling memslot, we add/remove them. This
         * should make address space updates faster, because we don't have to
         * loop over many disabled subregions.
         */
        trace_hv_balloon_map_slot(idx, memslots->count, offset);
        assert(!memory_region_is_mapped(&memslots->slots[idx]));
        memory_region_add_subregion(our_range->mr, offset,
                                    &memslots->slots[idx]);

        memslots->mapped_count++;
    }
    memory_region_transaction_commit();
}