Hi Avi,
I have addressed your comments and re-attached the fixed up patch. Most of
the things you suggested I implemented, but a few I didnt so I will comment
inline...
>>> On Thu, Apr 5, 2007 at 3:07 AM, in message <[EMAIL PROTECTED]>,
Avi Kivity <[EMAIL PROTECTED]> wrote:
> Gregory Haskins wrote:
>> The MMIO registration code has been broken out as a new patch from the
> in- kernel APIC work with the following changes per Avi's request:
>>
>> 1) Supports dynamic registration
>> 2) Uses gpa_t addresses
>> 3) Explicit per- cpu mappings
>>
>> In addition, I have added the concept of distinct VCPU and VM level
> registrations (where VCPU devices will eclipse competing VM registrations (if
> any). This will be key down the road where LAPICs should use VCPU
> registration, but IOAPICs should use VM level.
>>
>> Signed- off- by: Gregory Haskins <[EMAIL PROTECTED]>
>>
>> ---
>> drivers/kvm/kvm.h | 50 +++++++++++++++++++++++++++++++++++++++++++++
>> drivers/kvm/kvm_main.c | 53
>> +++++++++++++++++++++++++++++++++++++++--------
>> 2 files changed, 94 insertions(+), 9 deletions(- )
>>
>> diff -- git a/drivers/kvm/kvm.h b/drivers/kvm/kvm.h
>> index fceeb84..3334730 100644
>> --- a/drivers/kvm/kvm.h
>> +++ b/drivers/kvm/kvm.h
>> @@ - 236,6 +236,54 @@ struct kvm_pio_request {
>> int rep;
>> };
>>
>> +struct kvm_io_device {
>> + unsigned long (*read)(struct kvm_io_device *this,
>> + gpa_t addr,
>> + unsigned long length);
>> + void (*write)(struct kvm_io_device *this,
>> + gpa_t addr,
>> + unsigned long length,
>> + unsigned long val);
>>
>
> length could be just an int.
Done
>
>> + int (*in_range)(struct kvm_io_device *this, gpa_t addr);
>>
>
> Do you see any reason to have this as a callback and not a pair of gpas?
I believe Dor replied earlier stating the reason of being able to support
holes. Another reason that I can think of that I particularly like about this
design (which I am not claiming as my own) is that the device can relocate
(e.g. LAPIC base addr) without worrying about reprogramming the bus.
>
>> +
>> + void *private;
>> + struct list_head link;
>>
>
> Having these in an array would be much more efficient. A fixed size
> array of moderate size should suffice.
Done. Maximum # devices is currently 6, because anything beyond that and I
think we need to revisit the linear alg ;)
>
>> +};
>> +
>> +/* It would be nice to use something smarter than a linear search, TBD...
>> + Thankfully we dont expect many devices to register (famous last words
> :),
>> + so until then it will suffice. At least its abstracted so we can change
>> + in one place.
>> + */
>>
>
> /*
> * kernel comments look
> * like this
> */
Done
>
>> +struct kvm_io_bus {
>> + struct list_head list;
>> +};
>> +
>> +static inline void
>> +kvm_io_bus_init(struct kvm_io_bus *bus)
>>
>
> function declarations on one line please.
Done (though I hate lines that runneth over 80 ;)
>
>> +{
>> + INIT_LIST_HEAD(&bus- >list);
>> +}
>> +
>> +static inline struct kvm_io_device*
>>
>
> C style pointers:
> struct blah *somthing();
Done.
>
>> +kvm_io_bus_find_dev(struct kvm_io_bus *bus, gpa_t addr)
>> +{
>> + struct kvm_io_device *pos = NULL;
>> +
>> + list_for_each_entry(pos, &bus- >list, link) {
>> + if(pos- >in_range(pos, addr))
>> + return pos;
>> + }
>>
>
> space after if. avoid redundant {}. Have Documentaion/CodingStyle
> tattooed somewhere easily accessible.
Done
>
>> +
>> + return NULL;
>> +}
>> +
>> +static inline void
>> +kvm_io_bus_register_dev(struct kvm_io_bus *bus, struct kvm_io_device *dev)
>> +{
>> + list_add_tail(&dev- >link, &bus- >list);
>> +}
>> +
>> struct kvm_vcpu {
>> struct kvm *kvm;
>> union {
>> @@ - 294,6 +342,7 @@ struct kvm_vcpu {
>> gpa_t mmio_phys_addr;
>> struct kvm_pio_request pio;
>> void *pio_data;
>> + struct kvm_io_bus mmio_bus;
>>
>> int sigset_active;
>> sigset_t sigset;
>> @@ - 345,6 +394,7 @@ struct kvm {
>> unsigned long rmap_overflow;
>> struct list_head vm_list;
>> struct file *filp;
>> + struct kvm_io_bus mmio_bus;
>>
>
> The per- vcpu I/O bus is special in that it has exactly one component,
> and one which can change its address. I think we can special case it
> and just check for apic addresses explicitly when searching the bus.
I am loath to make special cases if they can be avoided. I think performance
wise a kvm_io_bus with one device wont be much different than having a special
case check against apicbase. And the advantage that this buys us is future
platforms (e.g. IA64?) may have more than one per-cpu MMIO address. I also
realize that future platforms may be divergent from the entire in-kernel code
base altogether, but I think the general and flexible way is better if there
are no compromising tradeoffs, even if its only for example/reference. In this
case I dont think there are any tradeoffs, so I left it. If you insist, I will
pull it ;)
>
>> };
>>
>> struct kvm_stat {
>> diff -- git a/drivers/kvm/kvm_main.c b/drivers/kvm/kvm_main.c
>> index 4473174..da119c0 100644
>> --- a/drivers/kvm/kvm_main.c
>> +++ b/drivers/kvm/kvm_main.c
>> @@ - 294,6 +294,7 @@ static struct kvm *kvm_create_vm(void)
>>
>> spin_lock_init(&kvm- >lock);
>> INIT_LIST_HEAD(&kvm- >active_mmu_pages);
>> + kvm_io_bus_init(&kvm- >mmio_bus);
>> for (i = 0; i < KVM_MAX_VCPUS; ++i) {
>> struct kvm_vcpu *vcpu = &kvm- >vcpus[i];
>>
>> @@ - 302,6 +303,7 @@ static struct kvm *kvm_create_vm(void)
>> vcpu- >kvm = kvm;
>> vcpu- >mmu.root_hpa = INVALID_PAGE;
>> INIT_LIST_HEAD(&vcpu- >free_pages);
>> + kvm_io_bus_init(&vcpu- >mmio_bus);
>> spin_lock(&kvm_lock);
>> list_add(&kvm- >vm_list, &vm_list);
>> spin_unlock(&kvm_lock);
>> @@ - 1015,12 +1017,30 @@ static int emulator_write_std(unsigned long addr,
>> return X86EMUL_UNHANDLEABLE;
>> }
>>
>> +static struct kvm_io_device* vcpu_find_mmio_dev(struct kvm_vcpu *vcpu,
>> + gpa_t addr)
>> +{
>> + struct kvm_io_device *mmio_dev;
>> +
>> + /* First check the local CPU addresses */
>> + mmio_dev = kvm_io_bus_find_dev(&vcpu- >mmio_bus, addr);
>> + if(!mmio_dev) {
>> + /* Then check the entire VM */
>> + mmio_dev = kvm_io_bus_find_dev(&vcpu- >kvm- >mmio_bus, addr);
>> + }
>>
>
> space, comment, braces
I believe I fixed this, but I am a little confused about what you were pointing
out. The space is obvious. I believe you were pointing out that the braces
weren't needed because its technically a single-line, and that the comment is
fine. If I needed to change the comment too, let me know.
>
>> +
>> + return mmio_dev;
>> +}
>> +
>> static int emulator_read_emulated(unsigned long addr,
>> unsigned long *val,
>> unsigned int bytes,
>> struct x86_emulate_ctxt *ctxt)
>> {
>> struct kvm_vcpu *vcpu = ctxt- >vcpu;
>> + gpa_t gpa;
>> + int i;
>> + struct kvm_io_device *mmio_dev;
>>
>> if (vcpu- >mmio_read_completed) {
>> memcpy(val, vcpu- >mmio_data, bytes);
>> @@ - 1029,18 +1049,24 @@ static int emulator_read_emulated(unsigned long
>> addr,
>> } else if (emulator_read_std(addr, val, bytes, ctxt)
>> == X86EMUL_CONTINUE)
>> return X86EMUL_CONTINUE;
>> - else {
>> - gpa_t gpa = vcpu- >mmu.gva_to_gpa(vcpu, addr);
>>
>> - if (gpa == UNMAPPED_GVA)
>> - return X86EMUL_PROPAGATE_FAULT;
>> - vcpu- >mmio_needed = 1;
>> - vcpu- >mmio_phys_addr = gpa;
>> - vcpu- >mmio_size = bytes;
>> - vcpu- >mmio_is_write = 0;
>> + gpa = vcpu- >mmu.gva_to_gpa(vcpu, addr);
>> + if (gpa == UNMAPPED_GVA)
>> + return vcpu_printf(vcpu, "not present\n"),
>> X86EMUL_PROPAGATE_FAULT;
>>
>
>
> The vcpu_printf() snuck in somehow.
Opps. Carry over from the apic branch merge. Fixed.
>
>>
>> - return X86EMUL_UNHANDLEABLE;
>> + /* Is thie MMIO handled locally? */
>> + mmio_dev = vcpu_find_mmio_dev(vcpu, gpa);
>> + if(mmio_dev) {
>> + *val = mmio_dev- >read(mmio_dev, gpa, bytes);
>> + return X86EMUL_CONTINUE;
>> }
>> +
>> + vcpu- >mmio_needed = 1;
>> + vcpu- >mmio_phys_addr = gpa;
>> + vcpu- >mmio_size = bytes;
>> + vcpu- >mmio_is_write = 0;
>> +
>> + return X86EMUL_UNHANDLEABLE;
>> }
>>
>> static int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
>> @@ - 1070,6 +1096,8 @@ static int emulator_write_emulated(unsigned long addr,
>> {
>> struct kvm_vcpu *vcpu = ctxt- >vcpu;
>> gpa_t gpa = vcpu- >mmu.gva_to_gpa(vcpu, addr);
>> + int i;
>> + struct kvm_io_device *mmio_dev;
>>
>> if (gpa == UNMAPPED_GVA)
>> return X86EMUL_PROPAGATE_FAULT;
>> @@ - 1077,6 +1105,13 @@ static int emulator_write_emulated(unsigned long
>> addr,
>> if (emulator_write_phys(vcpu, gpa, val, bytes))
>> return X86EMUL_CONTINUE;
>>
>> + /* Is thie MMIO handled locally? */
>>
>
> spelling
Done
>
>> + mmio_dev = vcpu_find_mmio_dev(vcpu, gpa);
>> + if(mmio_dev) {
>> + mmio_dev- >write(mmio_dev, gpa, bytes, val);
>> + return X86EMUL_CONTINUE;
>> + }
>> +
>> vcpu- >mmio_needed = 1;
>> vcpu- >mmio_phys_addr = gpa;
>> vcpu- >mmio_size = bytes;
>>
>>
>>
>
> Please fix and *test*. Boot at least 32- bit Windows with ACPI HAL and
> 64- bit Linux, the more the better of course.
I have confirmed that my 64 bit linux guest boots fine. I don't currently have
any other guests. Careful review of the code leads me to believe this should
be an inert change, so I wont go through the effort of finding an XP CD to
install unless you insist ;)
Regards,
-Greg
diff --git a/drivers/kvm/kvm.h b/drivers/kvm/kvm.h
index fceeb84..c1923df 100644
--- a/drivers/kvm/kvm.h
+++ b/drivers/kvm/kvm.h
@@ -236,6 +236,56 @@ struct kvm_pio_request {
int rep;
};
+struct kvm_io_device {
+ unsigned long (*read)(struct kvm_io_device *this,
+ gpa_t addr,
+ int length);
+ void (*write)(struct kvm_io_device *this,
+ gpa_t addr,
+ int length,
+ unsigned long val);
+ int (*in_range)(struct kvm_io_device *this, gpa_t addr);
+
+ void *private;
+};
+
+/* It would be nice to use something smarter than a linear search, TBD...
+ * Thankfully we dont expect many devices to register (famous last words :),
+ * so until then it will suffice. At least its abstracted so we can change
+ * in one place.
+ */
+struct kvm_io_bus {
+ int dev_count;
+#define NR_IOBUS_DEVS 6
+ struct kvm_io_device *devs[NR_IOBUS_DEVS];
+};
+
+static inline void kvm_io_bus_init(struct kvm_io_bus *bus)
+{
+ memset(bus, 0, sizeof(*bus));
+}
+
+static inline struct kvm_io_device *kvm_io_bus_find_dev(struct kvm_io_bus
*bus, gpa_t addr)
+{
+ int i;
+
+ for(i=0; i<bus->dev_count; i++) {
+ struct kvm_io_device *pos = bus->devs[i];
+
+ if (pos->in_range(pos, addr))
+ return pos;
+ }
+
+ return NULL;
+}
+
+static inline void kvm_io_bus_register_dev(struct kvm_io_bus *bus, struct
kvm_io_device *dev)
+{
+ BUG_ON(bus->dev_count >= (NR_IOBUS_DEVS-1));
+
+ bus->devs[bus->dev_count++] = dev;
+}
+
struct kvm_vcpu {
struct kvm *kvm;
union {
@@ -294,6 +344,7 @@ struct kvm_vcpu {
gpa_t mmio_phys_addr;
struct kvm_pio_request pio;
void *pio_data;
+ struct kvm_io_bus mmio_bus;
int sigset_active;
sigset_t sigset;
@@ -345,6 +396,7 @@ struct kvm {
unsigned long rmap_overflow;
struct list_head vm_list;
struct file *filp;
+ struct kvm_io_bus mmio_bus;
};
struct kvm_stat {
diff --git a/drivers/kvm/kvm_main.c b/drivers/kvm/kvm_main.c
index 4473174..9ca0ad3 100644
--- a/drivers/kvm/kvm_main.c
+++ b/drivers/kvm/kvm_main.c
@@ -294,6 +294,7 @@ static struct kvm *kvm_create_vm(void)
spin_lock_init(&kvm->lock);
INIT_LIST_HEAD(&kvm->active_mmu_pages);
+ kvm_io_bus_init(&kvm->mmio_bus);
for (i = 0; i < KVM_MAX_VCPUS; ++i) {
struct kvm_vcpu *vcpu = &kvm->vcpus[i];
@@ -302,6 +303,7 @@ static struct kvm *kvm_create_vm(void)
vcpu->kvm = kvm;
vcpu->mmu.root_hpa = INVALID_PAGE;
INIT_LIST_HEAD(&vcpu->free_pages);
+ kvm_io_bus_init(&vcpu->mmio_bus);
spin_lock(&kvm_lock);
list_add(&kvm->vm_list, &vm_list);
spin_unlock(&kvm_lock);
@@ -1015,12 +1017,28 @@ static int emulator_write_std(unsigned long addr,
return X86EMUL_UNHANDLEABLE;
}
+static struct kvm_io_device* vcpu_find_mmio_dev(struct kvm_vcpu *vcpu,
+ gpa_t addr)
+{
+ struct kvm_io_device *mmio_dev;
+
+ /* First check the local CPU addresses */
+ mmio_dev = kvm_io_bus_find_dev(&vcpu->mmio_bus, addr);
+ if (!mmio_dev)
+ /* Then check the entire VM */
+ mmio_dev = kvm_io_bus_find_dev(&vcpu->kvm->mmio_bus, addr);
+
+ return mmio_dev;
+}
+
static int emulator_read_emulated(unsigned long addr,
unsigned long *val,
unsigned int bytes,
struct x86_emulate_ctxt *ctxt)
{
- struct kvm_vcpu *vcpu = ctxt->vcpu;
+ struct kvm_vcpu *vcpu = ctxt->vcpu;
+ struct kvm_io_device *mmio_dev;
+ gpa_t gpa;
if (vcpu->mmio_read_completed) {
memcpy(val, vcpu->mmio_data, bytes);
@@ -1029,18 +1047,24 @@ static int emulator_read_emulated(unsigned long addr,
} else if (emulator_read_std(addr, val, bytes, ctxt)
== X86EMUL_CONTINUE)
return X86EMUL_CONTINUE;
- else {
- gpa_t gpa = vcpu->mmu.gva_to_gpa(vcpu, addr);
- if (gpa == UNMAPPED_GVA)
- return X86EMUL_PROPAGATE_FAULT;
- vcpu->mmio_needed = 1;
- vcpu->mmio_phys_addr = gpa;
- vcpu->mmio_size = bytes;
- vcpu->mmio_is_write = 0;
+ gpa = vcpu->mmu.gva_to_gpa(vcpu, addr);
+ if (gpa == UNMAPPED_GVA)
+ return X86EMUL_PROPAGATE_FAULT;
- return X86EMUL_UNHANDLEABLE;
+ /* Is this MMIO handled locally? */
+ mmio_dev = vcpu_find_mmio_dev(vcpu, gpa);
+ if (mmio_dev) {
+ *val = mmio_dev->read(mmio_dev, gpa, bytes);
+ return X86EMUL_CONTINUE;
}
+
+ vcpu->mmio_needed = 1;
+ vcpu->mmio_phys_addr = gpa;
+ vcpu->mmio_size = bytes;
+ vcpu->mmio_is_write = 0;
+
+ return X86EMUL_UNHANDLEABLE;
}
static int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
@@ -1068,8 +1092,9 @@ static int emulator_write_emulated(unsigned long addr,
unsigned int bytes,
struct x86_emulate_ctxt *ctxt)
{
- struct kvm_vcpu *vcpu = ctxt->vcpu;
- gpa_t gpa = vcpu->mmu.gva_to_gpa(vcpu, addr);
+ struct kvm_vcpu *vcpu = ctxt->vcpu;
+ struct kvm_io_device *mmio_dev;
+ gpa_t gpa = vcpu->mmu.gva_to_gpa(vcpu, addr);
if (gpa == UNMAPPED_GVA)
return X86EMUL_PROPAGATE_FAULT;
@@ -1077,6 +1102,13 @@ static int emulator_write_emulated(unsigned long addr,
if (emulator_write_phys(vcpu, gpa, val, bytes))
return X86EMUL_CONTINUE;
+ /* Is this MMIO handled locally? */
+ mmio_dev = vcpu_find_mmio_dev(vcpu, gpa);
+ if (mmio_dev) {
+ mmio_dev->write(mmio_dev, gpa, bytes, val);
+ return X86EMUL_CONTINUE;
+ }
+
vcpu->mmio_needed = 1;
vcpu->mmio_phys_addr = gpa;
vcpu->mmio_size = bytes;
-------------------------------------------------------------------------
Take Surveys. Earn Cash. Influence the Future of IT
Join SourceForge.net's Techsay panel and you'll get the chance to share your
opinions on IT & business topics through brief surveys-and earn cash
http://www.techsay.com/default.php?page=join.php&p=sourceforge&CID=DEVDEV
_______________________________________________
kvm-devel mailing list
kvm-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/kvm-devel