Message ID | 1386764361-15260-15-git-send-email-pbonzini@redhat.com |
---|---|
State | New |
Headers | show |
On Wed, 11 Dec 2013 13:19:21 +0100 Paolo Bonzini <pbonzini@redhat.com> wrote: > This option provides the infrastructure for binding guest NUMA nodes > to host NUMA nodes. For example: > > -object memory-ram,size=1024M,policy=membind,host-nodes=0,id=ram-node0 \ > -numa node,nodeid=0,cpus=0,memdev=ram-node0 \ > -object memory-ram,size=1024M,policy=interleave,host-nodes=1-3,id=ram-node1 \ > -numa node,nodeid=1,cpus=1,memdev=ram-node1 I was thinking about a bit more radical change: -object memory-ram,size=1024M,policy=membind,host-nodes=0,id=ram-node0 -device dimm,memdev=ram-node0,node=0 -object memory-ram,size=1024M,policy=membind,host-nodes=1,id=ram-node1 -device dimm,memdev=ram-node1,node=1 that would allow to avoid synthetic -numa option but would require conversion of initial RAM to dimms. That would be more flexible for example allowing bind several backends to one node (like: 1Gb_hugepage + 2Mb_hugepage ones) > > The option replaces "-numa mem". > > Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> > --- > include/sysemu/sysemu.h | 2 ++ > numa.c | 64 +++++++++++++++++++++++++++++++++++++++++++++++-- > qapi-schema.json | 6 ++++- > 3 files changed, 69 insertions(+), 3 deletions(-) > > diff --git a/include/sysemu/sysemu.h b/include/sysemu/sysemu.h > index e9da760..acfc0c7 100644 > --- a/include/sysemu/sysemu.h > +++ b/include/sysemu/sysemu.h > @@ -12,6 +12,7 @@ > #include "qemu/bitmap.h" > #include "qom/object.h" > #include "hw/boards.h" > +#include "sysemu/hostmem.h" > > /* vl.c */ > > @@ -140,6 +141,7 @@ extern int nb_numa_nodes; > typedef struct node_info { > uint64_t node_mem; > DECLARE_BITMAP(node_cpu, MAX_CPUMASK_BITS); > + HostMemoryBackend *node_memdev; > } NodeInfo; > extern NodeInfo numa_info[MAX_NODES]; > void set_numa_nodes(void); > diff --git a/numa.c b/numa.c > index f903b9e..686dbfa 100644 > --- a/numa.c > +++ b/numa.c > @@ -27,6 +27,8 @@ > #include "qapi-visit.h" > #include "qapi/opts-visitor.h" > #include "qapi/dealloc-visitor.h" > +#include "qapi/qmp/qerror.h" > + > QemuOptsList qemu_numa_opts = { > .name = "numa", > .implied_opt_name = "type", > @@ -34,10 +36,13 @@ QemuOptsList qemu_numa_opts = { > .desc = { { 0 } } /* validated with OptsVisitor */ > }; > > +static int have_memdevs = -1; > + > static int numa_node_parse(NumaNodeOptions *opts) > { > uint16_t nodenr; > uint16List *cpus = NULL; > + Error *local_err = NULL; > > if (opts->has_nodeid) { > nodenr = opts->nodeid; > @@ -60,6 +65,19 @@ static int numa_node_parse(NumaNodeOptions *opts) > bitmap_set(numa_info[nodenr].node_cpu, cpus->value, 1); > } > > + if (opts->has_mem && opts->has_memdev) { > + fprintf(stderr, "qemu: cannot specify both mem= and memdev=\n"); > + return -1; > + } > + > + if (have_memdevs == -1) { > + have_memdevs = opts->has_memdev; > + } > + if (opts->has_memdev != have_memdevs) { > + fprintf(stderr, "qemu: memdev option must be specified for either " > + "all or no nodes\n"); > + } > + > if (opts->has_mem) { > int64_t mem_size; > char *endptr; > @@ -70,7 +88,19 @@ static int numa_node_parse(NumaNodeOptions *opts) > } > numa_info[nodenr].node_mem = mem_size; > } > + if (opts->has_memdev) { > + Object *o; > + o = object_resolve_path_type(opts->memdev, TYPE_MEMORY_BACKEND, NULL); > + if (!o) { > + error_setg(&local_err, "memdev=%s is ambiguous", opts->memdev); > + qerror_report_err(local_err); > + return -1; > + } > > + object_ref(o); > + numa_info[nodenr].node_mem = object_property_get_int(o, "size", NULL); > + numa_info[nodenr].node_memdev = MEMORY_BACKEND(o); > + } there wouldn't be any need for keeping numa mappings in separate structure, if we use node property of dimm device. acpi_build then could build SRAT directly enumerating present dimm devices. > return 0; > } > > @@ -188,12 +218,42 @@ void set_numa_modes(void) > } > } > > +static void allocate_system_memory_nonnuma(MemoryRegion *mr, Object *owner, > + const char *name, > + QEMUMachineInitArgs *args) > +{ > + uint64_t ram_size = args->ram_size; > + > + memory_region_init_ram(mr, owner, name, ram_size); > + vmstate_register_ram_global(mr); > +} > + > void memory_region_allocate_system_memory(MemoryRegion *mr, Object *owner, > const char *name, > QEMUMachineInitArgs *args) > { > uint64_t ram_size = args->ram_size; > + uint64_t addr = 0; > + int i; > > - memory_region_init_ram(mr, owner, name, ram_size); > - vmstate_register_ram_global(mr); > + if (nb_numa_nodes == 0 || !have_memdevs) { > + allocate_system_memory_nonnuma(mr, owner, name, args); > + return; > + } > + > + memory_region_init(mr, owner, name, ram_size); container might clip subregions due to presence of hole, it's size should be ram_size + hole_size. > + for (i = 0; i < nb_numa_nodes; i++) { > + Error *local_err = NULL; > + uint64_t size = numa_info[i].node_mem; > + HostMemoryBackend *backend = numa_info[i].node_memdev; > + MemoryRegion *seg = host_memory_backend_get_memory(backend, &local_err); > + if (local_err) { > + qerror_report_err(local_err); > + exit(1); > + } > + > + memory_region_add_subregion(mr, addr, seg); > + vmstate_register_ram_global(seg); > + addr += size; > + } > } > diff --git a/qapi-schema.json b/qapi-schema.json > index d99e39d..e449316 100644 > --- a/qapi-schema.json > +++ b/qapi-schema.json > @@ -4256,7 +4256,10 @@ > # > # @cpus: #optional VCPUs belong to this node > # > -# @mem: #optional memory size of this node > +# @memdev: #optional memory backend object. If specified for one node, > +# it must be specified for all nodes. > +# > +# @mem: #optional memory size of this node; mutually exclusive with @memdev. > # > # Since: 2.0 > ## > @@ -4264,4 +4267,5 @@ > 'data': { > '*nodeid': 'uint16', > '*cpus': ['uint16'], > + '*memdev': 'str', > '*mem': 'str' }} > -- > 1.8.4.2 >
Il 11/12/2013 16:36, Igor Mammedov ha scritto: >> > -object memory-ram,size=1024M,policy=membind,host-nodes=0,id=ram-node0 \ >> > -numa node,nodeid=0,cpus=0,memdev=ram-node0 \ >> > -object memory-ram,size=1024M,policy=interleave,host-nodes=1-3,id=ram-node1 \ >> > -numa node,nodeid=1,cpus=1,memdev=ram-node1 > > I was thinking about a bit more radical change: > -object memory-ram,size=1024M,policy=membind,host-nodes=0,id=ram-node0 > -device dimm,memdev=ram-node0,node=0 > -object memory-ram,size=1024M,policy=membind,host-nodes=1,id=ram-node1 > -device dimm,memdev=ram-node1,node=1 > > that would allow to avoid synthetic -numa option You still need -numa for cpus, at least in the short/medium-term. > but would require conversion > of initial RAM to dimms. That would be more flexible for example allowing > bind several backends to one node (like: 1Gb_hugepage + 2Mb_hugepage ones) Yes, that's another possibility, perhaps even cleaner. With "-numa node,memdev=", your board only needs to use memory_region_allocate_system_memory in order to apply a NUMA policy to guest RAM; it also works with guests or boards that are not themselves NUMA-aware. Doing the same with "-device dimm" requires changing the board to support RAM-as-dimms; do you have any idea how easy/hard that would be? Paolo
diff --git a/include/sysemu/sysemu.h b/include/sysemu/sysemu.h index e9da760..acfc0c7 100644 --- a/include/sysemu/sysemu.h +++ b/include/sysemu/sysemu.h @@ -12,6 +12,7 @@ #include "qemu/bitmap.h" #include "qom/object.h" #include "hw/boards.h" +#include "sysemu/hostmem.h" /* vl.c */ @@ -140,6 +141,7 @@ extern int nb_numa_nodes; typedef struct node_info { uint64_t node_mem; DECLARE_BITMAP(node_cpu, MAX_CPUMASK_BITS); + HostMemoryBackend *node_memdev; } NodeInfo; extern NodeInfo numa_info[MAX_NODES]; void set_numa_nodes(void); diff --git a/numa.c b/numa.c index f903b9e..686dbfa 100644 --- a/numa.c +++ b/numa.c @@ -27,6 +27,8 @@ #include "qapi-visit.h" #include "qapi/opts-visitor.h" #include "qapi/dealloc-visitor.h" +#include "qapi/qmp/qerror.h" + QemuOptsList qemu_numa_opts = { .name = "numa", .implied_opt_name = "type", @@ -34,10 +36,13 @@ QemuOptsList qemu_numa_opts = { .desc = { { 0 } } /* validated with OptsVisitor */ }; +static int have_memdevs = -1; + static int numa_node_parse(NumaNodeOptions *opts) { uint16_t nodenr; uint16List *cpus = NULL; + Error *local_err = NULL; if (opts->has_nodeid) { nodenr = opts->nodeid; @@ -60,6 +65,19 @@ static int numa_node_parse(NumaNodeOptions *opts) bitmap_set(numa_info[nodenr].node_cpu, cpus->value, 1); } + if (opts->has_mem && opts->has_memdev) { + fprintf(stderr, "qemu: cannot specify both mem= and memdev=\n"); + return -1; + } + + if (have_memdevs == -1) { + have_memdevs = opts->has_memdev; + } + if (opts->has_memdev != have_memdevs) { + fprintf(stderr, "qemu: memdev option must be specified for either " + "all or no nodes\n"); + } + if (opts->has_mem) { int64_t mem_size; char *endptr; @@ -70,7 +88,19 @@ static int numa_node_parse(NumaNodeOptions *opts) } numa_info[nodenr].node_mem = mem_size; } + if (opts->has_memdev) { + Object *o; + o = object_resolve_path_type(opts->memdev, TYPE_MEMORY_BACKEND, NULL); + if (!o) { + error_setg(&local_err, "memdev=%s is ambiguous", opts->memdev); + qerror_report_err(local_err); + return -1; + } + object_ref(o); + numa_info[nodenr].node_mem = object_property_get_int(o, "size", NULL); + numa_info[nodenr].node_memdev = MEMORY_BACKEND(o); + } return 0; } @@ -188,12 +218,42 @@ void set_numa_modes(void) } } +static void allocate_system_memory_nonnuma(MemoryRegion *mr, Object *owner, + const char *name, + QEMUMachineInitArgs *args) +{ + uint64_t ram_size = args->ram_size; + + memory_region_init_ram(mr, owner, name, ram_size); + vmstate_register_ram_global(mr); +} + void memory_region_allocate_system_memory(MemoryRegion *mr, Object *owner, const char *name, QEMUMachineInitArgs *args) { uint64_t ram_size = args->ram_size; + uint64_t addr = 0; + int i; - memory_region_init_ram(mr, owner, name, ram_size); - vmstate_register_ram_global(mr); + if (nb_numa_nodes == 0 || !have_memdevs) { + allocate_system_memory_nonnuma(mr, owner, name, args); + return; + } + + memory_region_init(mr, owner, name, ram_size); + for (i = 0; i < nb_numa_nodes; i++) { + Error *local_err = NULL; + uint64_t size = numa_info[i].node_mem; + HostMemoryBackend *backend = numa_info[i].node_memdev; + MemoryRegion *seg = host_memory_backend_get_memory(backend, &local_err); + if (local_err) { + qerror_report_err(local_err); + exit(1); + } + + memory_region_add_subregion(mr, addr, seg); + vmstate_register_ram_global(seg); + addr += size; + } } diff --git a/qapi-schema.json b/qapi-schema.json index d99e39d..e449316 100644 --- a/qapi-schema.json +++ b/qapi-schema.json @@ -4256,7 +4256,10 @@ # # @cpus: #optional VCPUs belong to this node # -# @mem: #optional memory size of this node +# @memdev: #optional memory backend object. If specified for one node, +# it must be specified for all nodes. +# +# @mem: #optional memory size of this node; mutually exclusive with @memdev. # # Since: 2.0 ## @@ -4264,4 +4267,5 @@ 'data': { '*nodeid': 'uint16', '*cpus': ['uint16'], + '*memdev': 'str', '*mem': 'str' }}
This option provides the infrastructure for binding guest NUMA nodes to host NUMA nodes. For example: -object memory-ram,size=1024M,policy=membind,host-nodes=0,id=ram-node0 \ -numa node,nodeid=0,cpus=0,memdev=ram-node0 \ -object memory-ram,size=1024M,policy=interleave,host-nodes=1-3,id=ram-node1 \ -numa node,nodeid=1,cpus=1,memdev=ram-node1 The option replaces "-numa mem". Signed-off-by: Paolo Bonzini <pbonzini@redhat.com> --- include/sysemu/sysemu.h | 2 ++ numa.c | 64 +++++++++++++++++++++++++++++++++++++++++++++++-- qapi-schema.json | 6 ++++- 3 files changed, 69 insertions(+), 3 deletions(-)