From mboxrd@z Thu Jan 1 00:00:00 1970 From: Paul Durrant Subject: [PATCH v2 07/17] x86/hvm: only call hvm_io_assist() from hvm_wait_for_io() Date: Thu, 11 Jun 2015 16:42:51 +0100 Message-ID: <1434037381-10917-8-git-send-email-paul.durrant@citrix.com> References: <1434037381-10917-1-git-send-email-paul.durrant@citrix.com> Mime-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Return-path: Received: from mail6.bemta14.messagelabs.com ([193.109.254.103]) by lists.xen.org with esmtp (Exim 4.72) (envelope-from ) id 1Z34dY-0006TQ-W3 for xen-devel@lists.xenproject.org; Thu, 11 Jun 2015 15:43:25 +0000 In-Reply-To: <1434037381-10917-1-git-send-email-paul.durrant@citrix.com> List-Unsubscribe: , List-Post: List-Help: List-Subscribe: , Sender: xen-devel-bounces@lists.xen.org Errors-To: xen-devel-bounces@lists.xen.org To: xen-devel@lists.xenproject.org Cc: Andrew Cooper , Paul Durrant , Keir Fraser , Jan Beulich List-Id: xen-devel@lists.xenproject.org By removing the calls in hvmemul_do_io() (which is replaced by a single assignment) and hvm_complete_assist_request() (which is replaced by a call to process_portio_intercept() with a suitable set of ops) then hvm_io_assist() can be moved into hvm.c and made static (and hence be a candidate for inlining). This patch also fixes the I/O state test at the end of hvm_io_assist() to check the correct value. Since the ioreq server patch series was integrated the current ioreq state is no longer an indicator of in-flight I/O state, since an I/O sheduled by re-emulation may be targetted at a different ioreq server. Signed-off-by: Paul Durrant Cc: Keir Fraser Cc: Jan Beulich Cc: Andrew Cooper --- xen/arch/x86/hvm/emulate.c | 36 +++++++++++++++++--- xen/arch/x86/hvm/hvm.c | 74 ++++++++++++++++++++++++------------------ xen/arch/x86/hvm/intercept.c | 4 +-- xen/arch/x86/hvm/io.c | 39 ---------------------- xen/include/asm-x86/hvm/io.h | 3 +- 5 files changed, 79 insertions(+), 77 deletions(-) diff --git a/xen/arch/x86/hvm/emulate.c b/xen/arch/x86/hvm/emulate.c index 38b8956..4b3581a 100644 --- a/xen/arch/x86/hvm/emulate.c +++ b/xen/arch/x86/hvm/emulate.c @@ -50,6 +50,34 @@ static void hvmtrace_io_assist(int is_mmio, ioreq_t *p) trace_var(event, 0/*!cycles*/, size, buffer); } +static int null_read(struct hvm_io_handler *io_handler, + struct vcpu *v, + uint64_t addr, + uint32_t size, + uint64_t *data) +{ + *data = ~0ul; + return X86EMUL_OKAY; +} + +static int null_write(struct hvm_io_handler *io_handler, + struct vcpu *v, + uint64_t addr, + uint32_t size, + uint64_t data) +{ + return X86EMUL_OKAY; +} + +static const struct hvm_io_ops null_ops = { + .read = null_read, + .write = null_write +}; + +static struct hvm_io_handler null_handler = { + .ops = &null_ops +}; + static int hvmemul_do_io( int is_mmio, paddr_t addr, unsigned long *reps, unsigned int size, uint8_t dir, bool_t df, bool_t data_is_addr, uint64_t data) @@ -145,8 +173,7 @@ static int hvmemul_do_io( switch ( rc ) { case X86EMUL_OKAY: - p.state = STATE_IORESP_READY; - hvm_io_assist(&p); + vio->io_data = p.data; vio->io_state = HVMIO_none; break; case X86EMUL_UNHANDLEABLE: @@ -157,8 +184,9 @@ static int hvmemul_do_io( /* If there is no suitable backing DM, just ignore accesses */ if ( !s ) { - hvm_complete_assist_req(&p); - rc = X86EMUL_OKAY; + rc = process_io_intercept(curr, &p, &null_handler); + if ( rc == X86EMUL_OKAY ) + vio->io_data = p.data; vio->io_state = HVMIO_none; } else diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index dbba6e3..799f0e7 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -411,6 +411,49 @@ bool_t hvm_io_pending(struct vcpu *v) return 0; } +static void hvm_io_assist(ioreq_t *p) +{ + struct vcpu *curr = current; + struct hvm_vcpu_io *vio = &curr->arch.hvm_vcpu.hvm_io; + enum hvm_io_state io_state; + + p->state = STATE_IOREQ_NONE; + + io_state = vio->io_state; + vio->io_state = HVMIO_none; + + switch ( io_state ) + { + case HVMIO_awaiting_completion: + vio->io_state = HVMIO_completed; + vio->io_data = p->data; + break; + case HVMIO_handle_mmio_awaiting_completion: + vio->io_state = HVMIO_completed; + vio->io_data = p->data; + (void)handle_mmio(); + break; + case HVMIO_handle_pio_awaiting_completion: + if ( vio->io_size == 4 ) /* Needs zero extension. */ + guest_cpu_user_regs()->rax = (uint32_t)p->data; + else + memcpy(&guest_cpu_user_regs()->rax, &p->data, vio->io_size); + break; + default: + break; + } + + /* + * Re-emulation may have scheduled another I/O so io_state set + * at the top of the function may have changed. + */ + if ( vio->io_state == HVMIO_none ) + { + msix_write_completion(curr); + vcpu_end_shutdown_deferral(curr); + } +} + static bool_t hvm_wait_for_io(struct hvm_ioreq_vcpu *sv, ioreq_t *p) { /* NB. Optimised for common case (p->state == STATE_IOREQ_NONE). */ @@ -2667,37 +2710,6 @@ bool_t hvm_send_assist_req(struct hvm_ioreq_server *s, ioreq_t *proto_p) return 0; } -void hvm_complete_assist_req(ioreq_t *p) -{ - switch ( p->type ) - { - case IOREQ_TYPE_PCI_CONFIG: - ASSERT_UNREACHABLE(); - break; - case IOREQ_TYPE_COPY: - case IOREQ_TYPE_PIO: - if ( p->dir == IOREQ_READ ) - { - if ( !p->data_is_ptr ) - p->data = ~0ul; - else - { - int i, step = p->df ? -p->size : p->size; - uint32_t data = ~0; - - for ( i = 0; i < p->count; i++ ) - hvm_copy_to_guest_phys(p->data + step * i, &data, - p->size); - } - } - /* FALLTHRU */ - default: - p->state = STATE_IORESP_READY; - hvm_io_assist(p); - break; - } -} - void hvm_broadcast_assist_req(ioreq_t *p) { struct domain *d = current->domain; diff --git a/xen/arch/x86/hvm/intercept.c b/xen/arch/x86/hvm/intercept.c index c0c9254..0537d39 100644 --- a/xen/arch/x86/hvm/intercept.c +++ b/xen/arch/x86/hvm/intercept.c @@ -152,8 +152,8 @@ static const struct hvm_io_ops portio_ops = { .write = hvm_portio_write }; -static int process_io_intercept(struct vcpu *v, ioreq_t *p, - struct hvm_io_handler *handler) +int process_io_intercept(struct vcpu *v, ioreq_t *p, + struct hvm_io_handler *handler) { const struct hvm_io_ops *ops = handler->ops; int rc = X86EMUL_OKAY, i, step = p->df ? -p->size : p->size; diff --git a/xen/arch/x86/hvm/io.c b/xen/arch/x86/hvm/io.c index 87625ed..ee0ed82 100644 --- a/xen/arch/x86/hvm/io.c +++ b/xen/arch/x86/hvm/io.c @@ -169,45 +169,6 @@ int handle_pio(uint16_t port, unsigned int size, int dir) return 1; } -void hvm_io_assist(ioreq_t *p) -{ - struct vcpu *curr = current; - struct hvm_vcpu_io *vio = &curr->arch.hvm_vcpu.hvm_io; - enum hvm_io_state io_state; - - p->state = STATE_IOREQ_NONE; - - io_state = vio->io_state; - vio->io_state = HVMIO_none; - - switch ( io_state ) - { - case HVMIO_awaiting_completion: - vio->io_state = HVMIO_completed; - vio->io_data = p->data; - break; - case HVMIO_handle_mmio_awaiting_completion: - vio->io_state = HVMIO_completed; - vio->io_data = p->data; - (void)handle_mmio(); - break; - case HVMIO_handle_pio_awaiting_completion: - if ( vio->io_size == 4 ) /* Needs zero extension. */ - guest_cpu_user_regs()->rax = (uint32_t)p->data; - else - memcpy(&guest_cpu_user_regs()->rax, &p->data, vio->io_size); - break; - default: - break; - } - - if ( p->state == STATE_IOREQ_NONE ) - { - msix_write_completion(curr); - vcpu_end_shutdown_deferral(curr); - } -} - static bool_t dpci_portio_accept(struct hvm_io_handler *io_handler, struct vcpu *v, uint64_t addr, diff --git a/xen/include/asm-x86/hvm/io.h b/xen/include/asm-x86/hvm/io.h index 50965f1..7aa1f80 100644 --- a/xen/include/asm-x86/hvm/io.h +++ b/xen/include/asm-x86/hvm/io.h @@ -94,6 +94,8 @@ void hvm_register_io_handler(struct domain *d, struct hvm_io_handler *hvm_find_io_handler(struct vcpu *v, ioreq_t *p); +int process_io_intercept(struct vcpu *v, ioreq_t *p, + struct hvm_io_handler *handler); int hvm_io_intercept(ioreq_t *p); void register_mmio_handler(struct domain *d, @@ -113,7 +115,6 @@ int handle_mmio_with_translation(unsigned long gva, unsigned long gpfn, struct npfec); int handle_pio(uint16_t port, unsigned int size, int dir); void hvm_interrupt_post(struct vcpu *v, int vector, int type); -void hvm_io_assist(ioreq_t *p); void hvm_dpci_eoi(struct domain *d, unsigned int guest_irq, const union vioapic_redir_entry *ent); void msix_write_completion(struct vcpu *); -- 1.7.10.4