| /* |
| * Copyright (C) 2004, 2007-2010, 2011-2014 Synopsys, Inc. All rights reserved. |
| * |
| * SPDX-License-Identifier: GPL-2.0+ |
| */ |
| |
| /* |
| * If dst and src are 4 byte aligned, copy 8 bytes at a time. |
| * If the src is 4, but not 8 byte aligned, we first read 4 bytes to get |
| * it 8 byte aligned. Thus, we can do a little read-ahead, without |
| * dereferencing a cache line that we should not touch. |
| * Note that short and long instructions have been scheduled to avoid |
| * branch stalls. |
| * The beq_s to r3z could be made unaligned & long to avoid a stall |
| * there, but it is not likely to be taken often, and it would also be likely |
| * to cost an unaligned mispredict at the next call. |
| */ |
| |
| .global strcpy |
| .align 4 |
| strcpy: |
| or %r2, %r0, %r1 |
| bmsk_s %r2, %r2, 1 |
| brne.d %r2, 0, charloop |
| mov_s %r10, %r0 |
| ld_s %r3, [%r1, 0] |
| mov %r8, 0x01010101 |
| bbit0.d %r1, 2, loop_start |
| ror %r12, %r8 |
| sub %r2, %r3, %r8 |
| bic_s %r2, %r2, %r3 |
| tst_s %r2,%r12 |
| bne r3z |
| mov_s %r4,%r3 |
| .balign 4 |
| loop: |
| ld.a %r3, [%r1, 4] |
| st.ab %r4, [%r10, 4] |
| loop_start: |
| ld.a %r4, [%r1, 4] |
| sub %r2, %r3, %r8 |
| bic_s %r2, %r2, %r3 |
| tst_s %r2, %r12 |
| bne_s r3z |
| st.ab %r3, [%r10, 4] |
| sub %r2, %r4, %r8 |
| bic %r2, %r2, %r4 |
| tst %r2, %r12 |
| beq loop |
| mov_s %r3, %r4 |
| #ifdef __LITTLE_ENDIAN__ |
| r3z: bmsk.f %r1, %r3, 7 |
| lsr_s %r3, %r3, 8 |
| #else /* __BIG_ENDIAN__ */ |
| r3z: lsr.f %r1, %r3, 24 |
| asl_s %r3, %r3, 8 |
| #endif /* _ENDIAN__ */ |
| bne.d r3z |
| stb.ab %r1, [%r10, 1] |
| j_s [%blink] |
| |
| .balign 4 |
| charloop: |
| ldb.ab %r3, [%r1, 1] |
| brne.d %r3, 0, charloop |
| stb.ab %r3, [%r10, 1] |
| j [%blink] |