123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455 |
- /*
- * User Space Access Routines
- *
- * Copyright (C) 2000-2002 Hewlett-Packard (John Marvin)
- * Copyright (C) 2000 Richard Hirst <rhirst with parisc-linux.org>
- * Copyright (C) 2001 Matthieu Delahaye <delahaym at esiee.fr>
- * Copyright (C) 2003 Randolph Chung <tausq with parisc-linux.org>
- * Copyright (C) 2017 Helge Deller <deller@gmx.de>
- * Copyright (C) 2017 John David Anglin <dave.anglin@bell.net>
- *
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License as published by
- * the Free Software Foundation; either version 2, or (at your option)
- * any later version.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License
- * along with this program; if not, write to the Free Software
- * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
- */
- /*
- * These routines still have plenty of room for optimization
- * (word & doubleword load/store, dual issue, store hints, etc.).
- */
- /*
- * The following routines assume that space register 3 (sr3) contains
- * the space id associated with the current users address space.
- */
- .text
-
- #include <asm/assembly.h>
- #include <asm/errno.h>
- #include <linux/linkage.h>
- /*
- * get_sr gets the appropriate space value into
- * sr1 for kernel/user space access, depending
- * on the flag stored in the task structure.
- */
- .macro get_sr
- mfctl %cr30,%r1
- ldw TI_SEGMENT(%r1),%r22
- mfsp %sr3,%r1
- or,<> %r22,%r0,%r0
- copy %r0,%r1
- mtsp %r1,%sr1
- .endm
- .macro fixup_branch lbl
- ldil L%\lbl, %r1
- ldo R%\lbl(%r1), %r1
- bv %r0(%r1)
- .endm
- /*
- * unsigned long lclear_user(void *to, unsigned long n)
- *
- * Returns 0 for success.
- * otherwise, returns number of bytes not transferred.
- */
- ENTRY_CFI(lclear_user)
- .proc
- .callinfo NO_CALLS
- .entry
- comib,=,n 0,%r25,$lclu_done
- get_sr
- $lclu_loop:
- addib,<> -1,%r25,$lclu_loop
- 1: stbs,ma %r0,1(%sr1,%r26)
- $lclu_done:
- bv %r0(%r2)
- copy %r25,%r28
- .exit
- ENDPROC_CFI(lclear_user)
- .section .fixup,"ax"
- 2: fixup_branch $lclu_done
- ldo 1(%r25),%r25
- .previous
- ASM_EXCEPTIONTABLE_ENTRY(1b,2b)
- .procend
- /*
- * long lstrnlen_user(char *s, long n)
- *
- * Returns 0 if exception before zero byte or reaching N,
- * N+1 if N would be exceeded,
- * else strlen + 1 (i.e. includes zero byte).
- */
- ENTRY_CFI(lstrnlen_user)
- .proc
- .callinfo NO_CALLS
- .entry
- comib,= 0,%r25,$lslen_nzero
- copy %r26,%r24
- get_sr
- 1: ldbs,ma 1(%sr1,%r26),%r1
- $lslen_loop:
- comib,=,n 0,%r1,$lslen_done
- addib,<> -1,%r25,$lslen_loop
- 2: ldbs,ma 1(%sr1,%r26),%r1
- $lslen_done:
- bv %r0(%r2)
- sub %r26,%r24,%r28
- .exit
- $lslen_nzero:
- b $lslen_done
- ldo 1(%r26),%r26 /* special case for N == 0 */
- ENDPROC_CFI(lstrnlen_user)
- .section .fixup,"ax"
- 3: fixup_branch $lslen_done
- copy %r24,%r26 /* reset r26 so 0 is returned on fault */
- .previous
- ASM_EXCEPTIONTABLE_ENTRY(1b,3b)
- ASM_EXCEPTIONTABLE_ENTRY(2b,3b)
- .procend
- /*
- * unsigned long pa_memcpy(void *dstp, const void *srcp, unsigned long len)
- *
- * Inputs:
- * - sr1 already contains space of source region
- * - sr2 already contains space of destination region
- *
- * Returns:
- * - number of bytes that could not be copied.
- * On success, this will be zero.
- *
- * This code is based on a C-implementation of a copy routine written by
- * Randolph Chung, which in turn was derived from the glibc.
- *
- * Several strategies are tried to try to get the best performance for various
- * conditions. In the optimal case, we copy by loops that copy 32- or 16-bytes
- * at a time using general registers. Unaligned copies are handled either by
- * aligning the destination and then using shift-and-write method, or in a few
- * cases by falling back to a byte-at-a-time copy.
- *
- * Testing with various alignments and buffer sizes shows that this code is
- * often >10x faster than a simple byte-at-a-time copy, even for strangely
- * aligned operands. It is interesting to note that the glibc version of memcpy
- * (written in C) is actually quite fast already. This routine is able to beat
- * it by 30-40% for aligned copies because of the loop unrolling, but in some
- * cases the glibc version is still slightly faster. This lends more
- * credibility that gcc can generate very good code as long as we are careful.
- *
- * Possible optimizations:
- * - add cache prefetching
- * - try not to use the post-increment address modifiers; they may create
- * additional interlocks. Assumption is that those were only efficient on old
- * machines (pre PA8000 processors)
- */
- dst = arg0
- src = arg1
- len = arg2
- end = arg3
- t1 = r19
- t2 = r20
- t3 = r21
- t4 = r22
- srcspc = sr1
- dstspc = sr2
- t0 = r1
- a1 = t1
- a2 = t2
- a3 = t3
- a0 = t4
- save_src = ret0
- save_dst = ret1
- save_len = r31
- ENTRY_CFI(pa_memcpy)
- .proc
- .callinfo NO_CALLS
- .entry
- /* Last destination address */
- add dst,len,end
- /* short copy with less than 16 bytes? */
- cmpib,COND(>>=),n 15,len,.Lbyte_loop
- /* same alignment? */
- xor src,dst,t0
- extru t0,31,2,t1
- cmpib,<>,n 0,t1,.Lunaligned_copy
- #ifdef CONFIG_64BIT
- /* only do 64-bit copies if we can get aligned. */
- extru t0,31,3,t1
- cmpib,<>,n 0,t1,.Lalign_loop32
- /* loop until we are 64-bit aligned */
- .Lalign_loop64:
- extru dst,31,3,t1
- cmpib,=,n 0,t1,.Lcopy_loop_16_start
- 20: ldb,ma 1(srcspc,src),t1
- 21: stb,ma t1,1(dstspc,dst)
- b .Lalign_loop64
- ldo -1(len),len
- ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done)
- .Lcopy_loop_16_start:
- ldi 31,t0
- .Lcopy_loop_16:
- cmpb,COND(>>=),n t0,len,.Lword_loop
- 10: ldd 0(srcspc,src),t1
- 11: ldd 8(srcspc,src),t2
- ldo 16(src),src
- 12: std,ma t1,8(dstspc,dst)
- 13: std,ma t2,8(dstspc,dst)
- 14: ldd 0(srcspc,src),t1
- 15: ldd 8(srcspc,src),t2
- ldo 16(src),src
- 16: std,ma t1,8(dstspc,dst)
- 17: std,ma t2,8(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(11b,.Lcopy16_fault)
- ASM_EXCEPTIONTABLE_ENTRY(12b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(13b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(14b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(15b,.Lcopy16_fault)
- ASM_EXCEPTIONTABLE_ENTRY(16b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(17b,.Lcopy_done)
- b .Lcopy_loop_16
- ldo -32(len),len
- .Lword_loop:
- cmpib,COND(>>=),n 3,len,.Lbyte_loop
- 20: ldw,ma 4(srcspc,src),t1
- 21: stw,ma t1,4(dstspc,dst)
- b .Lword_loop
- ldo -4(len),len
- ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done)
- #endif /* CONFIG_64BIT */
- /* loop until we are 32-bit aligned */
- .Lalign_loop32:
- extru dst,31,2,t1
- cmpib,=,n 0,t1,.Lcopy_loop_8
- 20: ldb,ma 1(srcspc,src),t1
- 21: stb,ma t1,1(dstspc,dst)
- b .Lalign_loop32
- ldo -1(len),len
- ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done)
- .Lcopy_loop_8:
- cmpib,COND(>>=),n 15,len,.Lbyte_loop
- 10: ldw 0(srcspc,src),t1
- 11: ldw 4(srcspc,src),t2
- 12: stw,ma t1,4(dstspc,dst)
- 13: stw,ma t2,4(dstspc,dst)
- 14: ldw 8(srcspc,src),t1
- 15: ldw 12(srcspc,src),t2
- ldo 16(src),src
- 16: stw,ma t1,4(dstspc,dst)
- 17: stw,ma t2,4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(11b,.Lcopy8_fault)
- ASM_EXCEPTIONTABLE_ENTRY(12b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(13b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(14b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(15b,.Lcopy8_fault)
- ASM_EXCEPTIONTABLE_ENTRY(16b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(17b,.Lcopy_done)
- b .Lcopy_loop_8
- ldo -16(len),len
- .Lbyte_loop:
- cmpclr,COND(<>) len,%r0,%r0
- b,n .Lcopy_done
- 20: ldb 0(srcspc,src),t1
- ldo 1(src),src
- 21: stb,ma t1,1(dstspc,dst)
- b .Lbyte_loop
- ldo -1(len),len
- ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done)
- .Lcopy_done:
- bv %r0(%r2)
- sub end,dst,ret0
- /* src and dst are not aligned the same way. */
- /* need to go the hard way */
- .Lunaligned_copy:
- /* align until dst is 32bit-word-aligned */
- extru dst,31,2,t1
- cmpib,=,n 0,t1,.Lcopy_dstaligned
- 20: ldb 0(srcspc,src),t1
- ldo 1(src),src
- 21: stb,ma t1,1(dstspc,dst)
- b .Lunaligned_copy
- ldo -1(len),len
- ASM_EXCEPTIONTABLE_ENTRY(20b,.Lcopy_done)
- ASM_EXCEPTIONTABLE_ENTRY(21b,.Lcopy_done)
- .Lcopy_dstaligned:
- /* store src, dst and len in safe place */
- copy src,save_src
- copy dst,save_dst
- copy len,save_len
- /* len now needs give number of words to copy */
- SHRREG len,2,len
- /*
- * Copy from a not-aligned src to an aligned dst using shifts.
- * Handles 4 words per loop.
- */
- depw,z src,28,2,t0
- subi 32,t0,t0
- mtsar t0
- extru len,31,2,t0
- cmpib,= 2,t0,.Lcase2
- /* Make src aligned by rounding it down. */
- depi 0,31,2,src
- cmpiclr,<> 3,t0,%r0
- b,n .Lcase3
- cmpiclr,<> 1,t0,%r0
- b,n .Lcase1
- .Lcase0:
- cmpb,COND(=) %r0,len,.Lcda_finish
- nop
- 1: ldw,ma 4(srcspc,src), a3
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- 1: ldw,ma 4(srcspc,src), a0
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- b,n .Ldo3
- .Lcase1:
- 1: ldw,ma 4(srcspc,src), a2
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- 1: ldw,ma 4(srcspc,src), a3
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- ldo -1(len),len
- cmpb,COND(=),n %r0,len,.Ldo0
- .Ldo4:
- 1: ldw,ma 4(srcspc,src), a0
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- shrpw a2, a3, %sar, t0
- 1: stw,ma t0, 4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done)
- .Ldo3:
- 1: ldw,ma 4(srcspc,src), a1
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- shrpw a3, a0, %sar, t0
- 1: stw,ma t0, 4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done)
- .Ldo2:
- 1: ldw,ma 4(srcspc,src), a2
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- shrpw a0, a1, %sar, t0
- 1: stw,ma t0, 4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done)
- .Ldo1:
- 1: ldw,ma 4(srcspc,src), a3
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- shrpw a1, a2, %sar, t0
- 1: stw,ma t0, 4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done)
- ldo -4(len),len
- cmpb,COND(<>) %r0,len,.Ldo4
- nop
- .Ldo0:
- shrpw a2, a3, %sar, t0
- 1: stw,ma t0, 4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcopy_done)
- .Lcda_rdfault:
- .Lcda_finish:
- /* calculate new src, dst and len and jump to byte-copy loop */
- sub dst,save_dst,t0
- add save_src,t0,src
- b .Lbyte_loop
- sub save_len,t0,len
- .Lcase3:
- 1: ldw,ma 4(srcspc,src), a0
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- 1: ldw,ma 4(srcspc,src), a1
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- b .Ldo2
- ldo 1(len),len
- .Lcase2:
- 1: ldw,ma 4(srcspc,src), a1
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- 1: ldw,ma 4(srcspc,src), a2
- ASM_EXCEPTIONTABLE_ENTRY(1b,.Lcda_rdfault)
- b .Ldo1
- ldo 2(len),len
- /* fault exception fixup handlers: */
- #ifdef CONFIG_64BIT
- .Lcopy16_fault:
- b .Lcopy_done
- 10: std,ma t1,8(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done)
- #endif
- .Lcopy8_fault:
- b .Lcopy_done
- 10: stw,ma t1,4(dstspc,dst)
- ASM_EXCEPTIONTABLE_ENTRY(10b,.Lcopy_done)
- .exit
- ENDPROC_CFI(pa_memcpy)
- .procend
- .end
|