1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175
|
/* Vector optimized 32/64 bit S/390 version of wcscat.
Copyright (C) 2015-2016 Free Software Foundation, Inc.
This file is part of the GNU C Library.
The GNU C Library is free software; you can redistribute it and/or
modify it under the terms of the GNU Lesser General Public
License as published by the Free Software Foundation; either
version 2.1 of the License, or (at your option) any later version.
The GNU C Library is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
License along with the GNU C Library; if not, see
<http://www.gnu.org/licenses/>. */
#if defined HAVE_S390_VX_ASM_SUPPORT && IS_IN (libc)
# include "sysdep.h"
# include "asm-syntax.h"
.text
/* wchar_t * wcscat (wchar_t *dest, const wchar_t *src)
Concatenate two strings.
Register usage:
-r0=saved dest pointer for return
-r1=tmp
-r2=dest
-r3=src
-r4=tmp
-r5=current_len
-v16=part of src
-v17=index of zero
-v18=part of src
*/
ENTRY(__wcscat_vx)
.machine "z13"
.machinemode "zarch_nohighgprs"
vlbb %v16,0(%r2),6 /* Load s until next 4k-byte boundary. */
lcbb %r1,0(%r2),6 /* Get bytes to 4k-byte boundary or 16. */
/* __wcslen_c can handle non 4byte aligned pointers,
but __wcscpy_c not. Thus if either src or dest is
not 4byte aligned, use __wcscat_c. */
tmll %r2,3 /* Test if s is 4-byte aligned? */
jne .Lfallback /* And use common-code variant if not. */
tmll %r3,3 /* Test if src is 4-byte aligned? */
jne .Lfallback /* And use common-code variant if not. */
lgr %r0,%r2 /* Save destination pointer for return. */
/* WCSLEN
r1 = loaded bytes (tmp)
r4 = zero byte index (tmp)
r2 = dst
*/
vfenezf %v16,%v16,%v16 /* Find element not equal with zero search. */
vlgvb %r5,%v16,7 /* Load zero index or 16 if not found. */
clrjl %r5,%r1,.Llen_end /* Found zero within loaded bytes, end. */
/* Align s to 16 byte. */
risbgn %r1,%r2,60,128+63,0 /* %r3 = bits 60-63 of %r2 'and' 15. */
lghi %r5,16 /* current_len = 16. */
slr %r5,%r1 /* Compute bytes to 16bytes boundary. */
/* Find zero in 16byte aligned loop. */
.Llen_loop:
vl %v16,0(%r5,%r2) /* Load s. */
vfenezfs %v16,%v16,%v16 /* Find element not equal with zero search. */
je .Llen_found /* Jump away if zero was found. */
vl %v16,16(%r5,%r2)
vfenezfs %v16,%v16,%v16
je .Llen_found16
vl %v16,32(%r5,%r2)
vfenezfs %v16,%v16,%v16
je .Llen_found32
vl %v16,48(%r5,%r2)
vfenezfs %v16,%v16,%v16
je .Llen_found48
aghi %r5,64
j .Llen_loop /* No zero -> loop. */
.Llen_found48:
aghi %r5,16
.Llen_found32:
aghi %r5,16
.Llen_found16:
aghi %r5,16
.Llen_found:
vlgvb %r4,%v16,7 /* Load byte index of zero. */
algr %r5,%r4
.Llen_end:
/* WCSCPY
%r1 = loaded bytes (tmp)
%r4 = zero byte index (tmp)
%r3 = curr src pointer
%r2 = curr dst pointer
*/
la %r2,0(%r5,%r2) /* strcpy at end of dst-string. */
vlbb %v16,0(%r3),6 /* Load s until next 4k-byte boundary. */
lcbb %r1,0(%r3),6 /* Get bytes to 4k-byte boundary or 16. */
vfenezf %v17,%v16,%v16 /* Find element not equal with zero search. */
vlgvb %r5,%v17,7 /* Load zero index or 16 if not found. */
clrjl %r5,%r1,.Lcpy_found_align /* If found zero within loaded bytes,
copy bytes before and return. */
/* Align s to 16 byte. */
risbgn %r4,%r3,60,128+63,0 /* %r3 = bits 60-63 of %r2 'and' 15. */
lghi %r5,15 /* current_len = 15. */
slr %r5,%r4 /* Compute highest index to 16byte boundary. */
vstl %v16,%r5,0(%r2) /* Copy loaded characters - no zero. */
ahi %r5,1 /* Start loop at next character. */
/* Find zero in 16byte aligned loop. */
.Lcpy_loop:
vl %v16,0(%r5,%r3) /* Load s. */
vfenezfs %v17,%v16,%v16 /* Find element not equal with zero search. */
je .Lcpy_found_v16_0 /* Jump away if zero was found. */
vl %v18,16(%r5,%r3) /* Load next part of s. */
vst %v16,0(%r5,%r2) /* Save previous part without zero to dst. */
vfenezfs %v17,%v18,%v18
je .Lcpy_found_v18_16
vl %v16,32(%r5,%r3)
vst %v18,16(%r5,%r2)
vfenezfs %v17,%v16,%v16
je .Lcpy_found_v16_32
vl %v18,48(%r5,%r3)
vst %v16,32(%r5,%r2)
vfenezfs %v17,%v18,%v18
je .Lcpy_found_v18_48
vst %v18,48(%r5,%r2)
aghi %r5,64
j .Lcpy_loop /* No zero -> loop. */
.Lcpy_found_v16_32:
aghi %r5,32
.Lcpy_found_v16_0:
la %r4,0(%r5,%r2)
vlgvb %r1,%v17,7 /* Load byte index of zero. */
aghi %r1,3 /* Also copy remaining bytes of zero. */
vstl %v16,%r1,0(%r4) /* Copy characters including zero. */
lgr %r2,%r0 /* Load saved dest-ptr. */
br %r14
.Lcpy_found_v18_48:
aghi %r5,32
.Lcpy_found_v18_16:
la %r4,16(%r5,%r2)
vlgvb %r1,%v17,7 /* Load byte index of zero. */
aghi %r1,3 /* Also copy remaining bytes of zero. */
vstl %v18,%r1,0(%r4) /* Copy characters including zero. */
lgr %r2,%r0 /* Load saved dest-ptr. */
br %r14
.Lcpy_found_align:
aghi %r5,3 /* Also copy remaining bytes of found zero. */
vstl %v16,%r5,0(%r2) /* Copy characters including zero. */
lgr %r2,%r0 /* Load saved dest-ptr. */
br %r14
.Lfallback:
jg __wcscat_c
END(__wcscat_vx)
#endif /* HAVE_S390_VX_ASM_SUPPORT && IS_IN (libc) */
|