123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387 |
- /*
- * access.h - access guest memory
- *
- * Copyright IBM Corp. 2008,2009
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License (version 2 only)
- * as published by the Free Software Foundation.
- *
- * Author(s): Carsten Otte <cotte@de.ibm.com>
- */
- #ifndef __KVM_S390_GACCESS_H
- #define __KVM_S390_GACCESS_H
- #include <linux/compiler.h>
- #include <linux/kvm_host.h>
- #include <asm/uaccess.h>
- #include "kvm-s390.h"
- static inline void __user *__guestaddr_to_user(struct kvm_vcpu *vcpu,
- unsigned long guestaddr)
- {
- unsigned long prefix = vcpu->arch.sie_block->prefix;
- if (guestaddr < 2 * PAGE_SIZE)
- guestaddr += prefix;
- else if ((guestaddr >= prefix) && (guestaddr < prefix + 2 * PAGE_SIZE))
- guestaddr -= prefix;
- return (void __user *) gmap_fault(guestaddr, vcpu->arch.gmap);
- }
- static inline int get_guest_u64(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u64 *result)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- BUG_ON(guestaddr & 7);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return get_user(*result, (unsigned long __user *) uptr);
- }
- static inline int get_guest_u32(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u32 *result)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- BUG_ON(guestaddr & 3);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return get_user(*result, (u32 __user *) uptr);
- }
- static inline int get_guest_u16(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u16 *result)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- BUG_ON(guestaddr & 1);
- if (IS_ERR(uptr))
- return PTR_ERR(uptr);
- return get_user(*result, (u16 __user *) uptr);
- }
- static inline int get_guest_u8(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u8 *result)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return get_user(*result, (u8 __user *) uptr);
- }
- static inline int put_guest_u64(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u64 value)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- BUG_ON(guestaddr & 7);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return put_user(value, (u64 __user *) uptr);
- }
- static inline int put_guest_u32(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u32 value)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- BUG_ON(guestaddr & 3);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return put_user(value, (u32 __user *) uptr);
- }
- static inline int put_guest_u16(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u16 value)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- BUG_ON(guestaddr & 1);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return put_user(value, (u16 __user *) uptr);
- }
- static inline int put_guest_u8(struct kvm_vcpu *vcpu, unsigned long guestaddr,
- u8 value)
- {
- void __user *uptr = __guestaddr_to_user(vcpu, guestaddr);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- return put_user(value, (u8 __user *) uptr);
- }
- static inline int __copy_to_guest_slow(struct kvm_vcpu *vcpu,
- unsigned long guestdest,
- void *from, unsigned long n)
- {
- int rc;
- unsigned long i;
- u8 *data = from;
- for (i = 0; i < n; i++) {
- rc = put_guest_u8(vcpu, guestdest++, *(data++));
- if (rc < 0)
- return rc;
- }
- return 0;
- }
- static inline int __copy_to_guest_fast(struct kvm_vcpu *vcpu,
- unsigned long guestdest,
- void *from, unsigned long n)
- {
- int r;
- void __user *uptr;
- unsigned long size;
- if (guestdest + n < guestdest)
- return -EFAULT;
- /* simple case: all within one segment table entry? */
- if ((guestdest & PMD_MASK) == ((guestdest+n) & PMD_MASK)) {
- uptr = (void __user *) gmap_fault(guestdest, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- r = copy_to_user(uptr, from, n);
- if (r)
- r = -EFAULT;
- goto out;
- }
- /* copy first segment */
- uptr = (void __user *)gmap_fault(guestdest, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- size = PMD_SIZE - (guestdest & ~PMD_MASK);
- r = copy_to_user(uptr, from, size);
- if (r) {
- r = -EFAULT;
- goto out;
- }
- from += size;
- n -= size;
- guestdest += size;
- /* copy full segments */
- while (n >= PMD_SIZE) {
- uptr = (void __user *)gmap_fault(guestdest, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- r = copy_to_user(uptr, from, PMD_SIZE);
- if (r) {
- r = -EFAULT;
- goto out;
- }
- from += PMD_SIZE;
- n -= PMD_SIZE;
- guestdest += PMD_SIZE;
- }
- /* copy the tail segment */
- if (n) {
- uptr = (void __user *)gmap_fault(guestdest, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- r = copy_to_user(uptr, from, n);
- if (r)
- r = -EFAULT;
- }
- out:
- return r;
- }
- static inline int copy_to_guest_absolute(struct kvm_vcpu *vcpu,
- unsigned long guestdest,
- void *from, unsigned long n)
- {
- return __copy_to_guest_fast(vcpu, guestdest, from, n);
- }
- static inline int copy_to_guest(struct kvm_vcpu *vcpu, unsigned long guestdest,
- void *from, unsigned long n)
- {
- unsigned long prefix = vcpu->arch.sie_block->prefix;
- if ((guestdest < 2 * PAGE_SIZE) && (guestdest + n > 2 * PAGE_SIZE))
- goto slowpath;
- if ((guestdest < prefix) && (guestdest + n > prefix))
- goto slowpath;
- if ((guestdest < prefix + 2 * PAGE_SIZE)
- && (guestdest + n > prefix + 2 * PAGE_SIZE))
- goto slowpath;
- if (guestdest < 2 * PAGE_SIZE)
- guestdest += prefix;
- else if ((guestdest >= prefix) && (guestdest < prefix + 2 * PAGE_SIZE))
- guestdest -= prefix;
- return __copy_to_guest_fast(vcpu, guestdest, from, n);
- slowpath:
- return __copy_to_guest_slow(vcpu, guestdest, from, n);
- }
- static inline int __copy_from_guest_slow(struct kvm_vcpu *vcpu, void *to,
- unsigned long guestsrc,
- unsigned long n)
- {
- int rc;
- unsigned long i;
- u8 *data = to;
- for (i = 0; i < n; i++) {
- rc = get_guest_u8(vcpu, guestsrc++, data++);
- if (rc < 0)
- return rc;
- }
- return 0;
- }
- static inline int __copy_from_guest_fast(struct kvm_vcpu *vcpu, void *to,
- unsigned long guestsrc,
- unsigned long n)
- {
- int r;
- void __user *uptr;
- unsigned long size;
- if (guestsrc + n < guestsrc)
- return -EFAULT;
- /* simple case: all within one segment table entry? */
- if ((guestsrc & PMD_MASK) == ((guestsrc+n) & PMD_MASK)) {
- uptr = (void __user *) gmap_fault(guestsrc, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- r = copy_from_user(to, uptr, n);
- if (r)
- r = -EFAULT;
- goto out;
- }
- /* copy first segment */
- uptr = (void __user *)gmap_fault(guestsrc, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- size = PMD_SIZE - (guestsrc & ~PMD_MASK);
- r = copy_from_user(to, uptr, size);
- if (r) {
- r = -EFAULT;
- goto out;
- }
- to += size;
- n -= size;
- guestsrc += size;
- /* copy full segments */
- while (n >= PMD_SIZE) {
- uptr = (void __user *)gmap_fault(guestsrc, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- r = copy_from_user(to, uptr, PMD_SIZE);
- if (r) {
- r = -EFAULT;
- goto out;
- }
- to += PMD_SIZE;
- n -= PMD_SIZE;
- guestsrc += PMD_SIZE;
- }
- /* copy the tail segment */
- if (n) {
- uptr = (void __user *)gmap_fault(guestsrc, vcpu->arch.gmap);
- if (IS_ERR((void __force *) uptr))
- return PTR_ERR((void __force *) uptr);
- r = copy_from_user(to, uptr, n);
- if (r)
- r = -EFAULT;
- }
- out:
- return r;
- }
- static inline int copy_from_guest_absolute(struct kvm_vcpu *vcpu, void *to,
- unsigned long guestsrc,
- unsigned long n)
- {
- return __copy_from_guest_fast(vcpu, to, guestsrc, n);
- }
- static inline int copy_from_guest(struct kvm_vcpu *vcpu, void *to,
- unsigned long guestsrc, unsigned long n)
- {
- unsigned long prefix = vcpu->arch.sie_block->prefix;
- if ((guestsrc < 2 * PAGE_SIZE) && (guestsrc + n > 2 * PAGE_SIZE))
- goto slowpath;
- if ((guestsrc < prefix) && (guestsrc + n > prefix))
- goto slowpath;
- if ((guestsrc < prefix + 2 * PAGE_SIZE)
- && (guestsrc + n > prefix + 2 * PAGE_SIZE))
- goto slowpath;
- if (guestsrc < 2 * PAGE_SIZE)
- guestsrc += prefix;
- else if ((guestsrc >= prefix) && (guestsrc < prefix + 2 * PAGE_SIZE))
- guestsrc -= prefix;
- return __copy_from_guest_fast(vcpu, to, guestsrc, n);
- slowpath:
- return __copy_from_guest_slow(vcpu, to, guestsrc, n);
- }
- #endif
|