| /* Copyright (C) 2010-2021 Free Software Foundation, Inc. |
| Contributed by Bernd Schmidt <bernds@codesourcery.com>. |
| |
| This file is free software; you can redistribute it and/or modify it |
| under the terms of the GNU General Public License as published by the |
| Free Software Foundation; either version 3, or (at your option) any |
| later version. |
| |
| This file is distributed in the hope that it will be useful, but |
| WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| General Public License for more details. |
| |
| Under Section 7 of GPL version 3, you are granted additional |
| permissions described in the GCC Runtime Library Exception, version |
| 3.1, as published by the Free Software Foundation. |
| |
| You should have received a copy of the GNU General Public License and |
| a copy of the GCC Runtime Library Exception along with this program; |
| see the files COPYING3 and COPYING.RUNTIME respectively. If not, see |
| <http://www.gnu.org/licenses/>. */ |
| |
| ;; ABI considerations for the divide functions |
| ;; The following registers are call-used: |
| ;; __c6xabi_divi A0,A1,A2,A4,A6,B0,B1,B2,B4,B5 |
| ;; __c6xabi_divu A0,A1,A2,A4,A6,B0,B1,B2,B4 |
| ;; __c6xabi_remi A1,A2,A4,A5,A6,B0,B1,B2,B4 |
| ;; __c6xabi_remu A1,A4,A5,A7,B0,B1,B2,B4 |
| ;; |
| ;; In our implementation, divu and remu are leaf functions, |
| ;; while both divi and remi call into divu. |
| ;; A0 is not clobbered by any of the functions. |
| ;; divu does not clobber B2 either, which is taken advantage of |
| ;; in remi. |
| ;; divi uses B5 to hold the original return address during |
| ;; the call to divu. |
| ;; remi uses B2 and A5 to hold the input values during the |
| ;; call to divu. It stores B3 in on the stack. |
| |
| #ifdef L_divsi3 |
| .text |
| .align 2 |
| .global __c6xabi_divi |
| .hidden __c6xabi_divi |
| .type __c6xabi_divi, STT_FUNC |
| |
| __c6xabi_divi: |
| call .s2 __c6xabi_divu |
| || mv .d2 B3, B5 |
| || cmpgt .l1 0, A4, A1 |
| || cmpgt .l2 0, B4, B1 |
| |
| [A1] neg .l1 A4, A4 |
| || [B1] neg .l2 B4, B4 |
| || xor .s1x A1, B1, A1 |
| |
| #ifdef _TMS320C6400 |
| [A1] addkpc .s2 1f, B3, 4 |
| #else |
| [A1] mvkl .s2 1f, B3 |
| [A1] mvkh .s2 1f, B3 |
| nop 2 |
| #endif |
| 1: |
| neg .l1 A4, A4 |
| || mv .l2 B3,B5 |
| || ret .s2 B5 |
| nop 5 |
| #endif |
| |
| #if defined L_modsi3 || defined L_divmodsi4 |
| .align 2 |
| #ifdef L_modsi3 |
| #define MOD_OUTPUT_REG A4 |
| .global __c6xabi_remi |
| .hidden __c6xabi_remi |
| .type __c6xabi_remi, STT_FUNC |
| #else |
| #define MOD_OUTPUT_REG A5 |
| .global __c6xabi_divremi |
| .hidden __c6xabi_divremi |
| .type __c6xabi_divremi, STT_FUNC |
| __c6xabi_divremi: |
| #endif |
| |
| __c6xabi_remi: |
| stw .d2t2 B3, *B15--[2] |
| || cmpgt .l1 0, A4, A1 |
| || cmpgt .l2 0, B4, B2 |
| || mv .s1 A4, A5 |
| || call .s2 __c6xabi_divu |
| |
| [A1] neg .l1 A4, A4 |
| || [B2] neg .l2 B4, B4 |
| || xor .s2x B2, A1, B0 |
| || mv .d2 B4, B2 |
| |
| #ifdef _TMS320C6400 |
| [B0] addkpc .s2 1f, B3, 1 |
| [!B0] addkpc .s2 2f, B3, 1 |
| nop 2 |
| #else |
| [B0] mvkl .s2 1f,B3 |
| [!B0] mvkl .s2 2f,B3 |
| |
| [B0] mvkh .s2 1f,B3 |
| [!B0] mvkh .s2 2f,B3 |
| #endif |
| 1: |
| neg .l1 A4, A4 |
| 2: |
| ldw .d2t2 *++B15[2], B3 |
| |
| #ifdef _TMS320C6400_PLUS |
| mpy32 .m1x A4, B2, A6 |
| nop 3 |
| ret .s2 B3 |
| sub .l1 A5, A6, MOD_OUTPUT_REG |
| nop 4 |
| #else |
| mpyu .m1x A4, B2, A1 |
| nop 1 |
| mpylhu .m1x A4, B2, A6 |
| || mpylhu .m2x B2, A4, B2 |
| nop 1 |
| add .l1x A6, B2, A6 |
| || ret .s2 B3 |
| shl .s1 A6, 16, A6 |
| add .d1 A6, A1, A6 |
| sub .l1 A5, A6, MOD_OUTPUT_REG |
| nop 2 |
| #endif |
| |
| #endif |
| |
| #if defined L_udivsi3 || defined L_udivmodsi4 |
| .align 2 |
| #ifdef L_udivsi3 |
| .global __c6xabi_divu |
| .hidden __c6xabi_divu |
| .type __c6xabi_divu, STT_FUNC |
| __c6xabi_divu: |
| #else |
| .global __c6xabi_divremu |
| .hidden __c6xabi_divremu |
| .type __c6xabi_divremu, STT_FUNC |
| __c6xabi_divremu: |
| #endif |
| ;; We use a series of up to 31 subc instructions. First, we find |
| ;; out how many leading zero bits there are in the divisor. This |
| ;; gives us both a shift count for aligning (shifting) the divisor |
| ;; to the, and the number of times we have to execute subc. |
| |
| ;; At the end, we have both the remainder and most of the quotient |
| ;; in A4. The top bit of the quotient is computed first and is |
| ;; placed in A2. |
| |
| ;; Return immediately if the dividend is zero. Setting B4 to 1 |
| ;; is a trick to allow us to leave the following insns in the jump |
| ;; delay slot without affecting the result. |
| mv .s2x A4, B1 |
| |
| #ifndef _TMS320C6400 |
| [!b1] mvk .s2 1, B4 |
| #endif |
| [b1] lmbd .l2 1, B4, B1 |
| ||[!b1] b .s2 B3 ; RETURN A |
| #ifdef _TMS320C6400 |
| ||[!b1] mvk .d2 1, B4 |
| #endif |
| #ifdef L_udivmodsi4 |
| ||[!b1] zero .s1 A5 |
| #endif |
| mv .l1x B1, A6 |
| || shl .s2 B4, B1, B4 |
| |
| ;; The loop performs a maximum of 28 steps, so we do the |
| ;; first 3 here. |
| cmpltu .l1x A4, B4, A2 |
| [!A2] sub .l1x A4, B4, A4 |
| || shru .s2 B4, 1, B4 |
| || xor .s1 1, A2, A2 |
| |
| shl .s1 A2, 31, A2 |
| || [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| |
| ;; RETURN A may happen here (note: must happen before the next branch) |
| 0: |
| cmpgt .l2 B1, 7, B0 |
| || [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| || [b0] b .s1 0b |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| ;; loop backwards branch happens here |
| |
| ret .s2 B3 |
| || mvk .s1 32, A1 |
| sub .l1 A1, A6, A6 |
| #ifdef L_udivmodsi4 |
| || extu .s1 A4, A6, A5 |
| #endif |
| shl .s1 A4, A6, A4 |
| shru .s1 A4, 1, A4 |
| || sub .l1 A6, 1, A6 |
| or .l1 A2, A4, A4 |
| shru .s1 A4, A6, A4 |
| nop |
| |
| #endif |
| |
| #ifdef L_umodsi3 |
| .align 2 |
| .global __c6xabi_remu |
| .hidden __c6xabi_remu |
| .type __c6xabi_remu, STT_FUNC |
| __c6xabi_remu: |
| ;; The ABI seems designed to prevent these functions calling each other, |
| ;; so we duplicate most of the divsi3 code here. |
| mv .s2x A4, B1 |
| #ifndef _TMS320C6400 |
| [!b1] mvk .s2 1, B4 |
| #endif |
| lmbd .l2 1, B4, B1 |
| ||[!b1] b .s2 B3 ; RETURN A |
| #ifdef _TMS320C6400 |
| ||[!b1] mvk .d2 1, B4 |
| #endif |
| |
| mv .l1x B1, A7 |
| || shl .s2 B4, B1, B4 |
| |
| cmpltu .l1x A4, B4, A1 |
| [!a1] sub .l1x A4, B4, A4 |
| shru .s2 B4, 1, B4 |
| |
| 0: |
| cmpgt .l2 B1, 7, B0 |
| || [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| ;; RETURN A may happen here (note: must happen before the next branch) |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| || [b0] b .s1 0b |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| ;; loop backwards branch happens here |
| |
| ret .s2 B3 |
| [b1] subc .l1x A4,B4,A4 |
| || [b1] add .s2 -1, B1, B1 |
| [b1] subc .l1x A4,B4,A4 |
| |
| extu .s1 A4, A7, A4 |
| nop 2 |
| #endif |
| |
| #if defined L_strasgi_64plus && defined _TMS320C6400_PLUS |
| |
| .align 2 |
| .global __c6xabi_strasgi_64plus |
| .hidden __c6xabi_strasgi_64plus |
| .type __c6xabi_strasgi_64plus, STT_FUNC |
| __c6xabi_strasgi_64plus: |
| shru .s2x a6, 2, b31 |
| || mv .s1 a4, a30 |
| || mv .d2 b4, b30 |
| |
| add .s2 -4, b31, b31 |
| |
| sploopd 1 |
| || mvc .s2 b31, ilc |
| ldw .d2t2 *b30++, b31 |
| nop 4 |
| mv .s1x b31,a31 |
| spkernel 6, 0 |
| || stw .d1t1 a31, *a30++ |
| |
| ret .s2 b3 |
| nop 5 |
| #endif |
| |
| #ifdef L_strasgi |
| .global __c6xabi_strasgi |
| .type __c6xabi_strasgi, STT_FUNC |
| __c6xabi_strasgi: |
| ;; This is essentially memcpy, with alignment known to be at least |
| ;; 4, and the size a multiple of 4 greater than or equal to 28. |
| ldw .d2t1 *B4++, A0 |
| || mvk .s2 16, B1 |
| ldw .d2t1 *B4++, A1 |
| || mvk .s2 20, B2 |
| || sub .d1 A6, 24, A6 |
| ldw .d2t1 *B4++, A5 |
| ldw .d2t1 *B4++, A7 |
| || mv .l2x A6, B7 |
| ldw .d2t1 *B4++, A8 |
| ldw .d2t1 *B4++, A9 |
| || mv .s2x A0, B5 |
| || cmpltu .l2 B2, B7, B0 |
| |
| 0: |
| stw .d1t2 B5, *A4++ |
| ||[b0] ldw .d2t1 *B4++, A0 |
| || mv .s2x A1, B5 |
| || mv .l2 B7, B6 |
| |
| [b0] sub .d2 B6, 24, B7 |
| ||[b0] b .s2 0b |
| || cmpltu .l2 B1, B6, B0 |
| |
| [b0] ldw .d2t1 *B4++, A1 |
| || stw .d1t2 B5, *A4++ |
| || mv .s2x A5, B5 |
| || cmpltu .l2 12, B6, B0 |
| |
| [b0] ldw .d2t1 *B4++, A5 |
| || stw .d1t2 B5, *A4++ |
| || mv .s2x A7, B5 |
| || cmpltu .l2 8, B6, B0 |
| |
| [b0] ldw .d2t1 *B4++, A7 |
| || stw .d1t2 B5, *A4++ |
| || mv .s2x A8, B5 |
| || cmpltu .l2 4, B6, B0 |
| |
| [b0] ldw .d2t1 *B4++, A8 |
| || stw .d1t2 B5, *A4++ |
| || mv .s2x A9, B5 |
| || cmpltu .l2 0, B6, B0 |
| |
| [b0] ldw .d2t1 *B4++, A9 |
| || stw .d1t2 B5, *A4++ |
| || mv .s2x A0, B5 |
| || cmpltu .l2 B2, B7, B0 |
| |
| ;; loop back branch happens here |
| |
| cmpltu .l2 B1, B6, B0 |
| || ret .s2 b3 |
| |
| [b0] stw .d1t1 A1, *A4++ |
| || cmpltu .l2 12, B6, B0 |
| [b0] stw .d1t1 A5, *A4++ |
| || cmpltu .l2 8, B6, B0 |
| [b0] stw .d1t1 A7, *A4++ |
| || cmpltu .l2 4, B6, B0 |
| [b0] stw .d1t1 A8, *A4++ |
| || cmpltu .l2 0, B6, B0 |
| [b0] stw .d1t1 A9, *A4++ |
| |
| ;; return happens here |
| |
| #endif |
| |
| #ifdef _TMS320C6400_PLUS |
| #ifdef L_push_rts |
| .align 2 |
| .global __c6xabi_push_rts |
| .hidden __c6xabi_push_rts |
| .type __c6xabi_push_rts, STT_FUNC |
| __c6xabi_push_rts: |
| stw .d2t2 B14, *B15--[2] |
| stdw .d2t1 A15:A14, *B15-- |
| || b .s2x A3 |
| stdw .d2t2 B13:B12, *B15-- |
| stdw .d2t1 A13:A12, *B15-- |
| stdw .d2t2 B11:B10, *B15-- |
| stdw .d2t1 A11:A10, *B15-- |
| stdw .d2t2 B3:B2, *B15-- |
| #endif |
| |
| #ifdef L_pop_rts |
| .align 2 |
| .global __c6xabi_pop_rts |
| .hidden __c6xabi_pop_rts |
| .type __c6xabi_pop_rts, STT_FUNC |
| __c6xabi_pop_rts: |
| lddw .d2t2 *++B15, B3:B2 |
| lddw .d2t1 *++B15, A11:A10 |
| lddw .d2t2 *++B15, B11:B10 |
| lddw .d2t1 *++B15, A13:A12 |
| lddw .d2t2 *++B15, B13:B12 |
| lddw .d2t1 *++B15, A15:A14 |
| || b .s2 B3 |
| ldw .d2t2 *++B15[2], B14 |
| nop 4 |
| #endif |
| |
| #ifdef L_call_stub |
| .align 2 |
| .global __c6xabi_call_stub |
| .type __c6xabi_call_stub, STT_FUNC |
| __c6xabi_call_stub: |
| stw .d2t1 A2, *B15--[2] |
| stdw .d2t1 A7:A6, *B15-- |
| || call .s2 B31 |
| stdw .d2t1 A1:A0, *B15-- |
| stdw .d2t2 B7:B6, *B15-- |
| stdw .d2t2 B5:B4, *B15-- |
| stdw .d2t2 B1:B0, *B15-- |
| stdw .d2t2 B3:B2, *B15-- |
| || addkpc .s2 1f, B3, 0 |
| 1: |
| lddw .d2t2 *++B15, B3:B2 |
| lddw .d2t2 *++B15, B1:B0 |
| lddw .d2t2 *++B15, B5:B4 |
| lddw .d2t2 *++B15, B7:B6 |
| lddw .d2t1 *++B15, A1:A0 |
| lddw .d2t1 *++B15, A7:A6 |
| || b .s2 B3 |
| ldw .d2t1 *++B15[2], A2 |
| nop 4 |
| #endif |
| |
| #endif |
| |