summaryrefslogtreecommitdiff
path: root/gcc/config/c6x/lib1funcs.asm
diff options
context:
space:
mode:
Diffstat (limited to 'gcc/config/c6x/lib1funcs.asm')
-rw-r--r--gcc/config/c6x/lib1funcs.asm438
1 files changed, 0 insertions, 438 deletions
diff --git a/gcc/config/c6x/lib1funcs.asm b/gcc/config/c6x/lib1funcs.asm
deleted file mode 100644
index 5bf34474bbd..00000000000
--- a/gcc/config/c6x/lib1funcs.asm
+++ /dev/null
@@ -1,438 +0,0 @@
-/* Copyright 2010, 2011 Free Software Foundation, Inc.
- Contributed by Bernd Schmidt <bernds@codesourcery.com>.
-
-This file is free software; you can redistribute it and/or modify it
-under the terms of the GNU General Public License as published by the
-Free Software Foundation; either version 3, or (at your option) any
-later version.
-
-This file is distributed in the hope that it will be useful, but
-WITHOUT ANY WARRANTY; without even the implied warranty of
-MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
-General Public License for more details.
-
-Under Section 7 of GPL version 3, you are granted additional
-permissions described in the GCC Runtime Library Exception, version
-3.1, as published by the Free Software Foundation.
-
-You should have received a copy of the GNU General Public License and
-a copy of the GCC Runtime Library Exception along with this program;
-see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
-<http://www.gnu.org/licenses/>. */
-
- ;; ABI considerations for the divide functions
- ;; The following registers are call-used:
- ;; __c6xabi_divi A0,A1,A2,A4,A6,B0,B1,B2,B4,B5
- ;; __c6xabi_divu A0,A1,A2,A4,A6,B0,B1,B2,B4
- ;; __c6xabi_remi A1,A2,A4,A5,A6,B0,B1,B2,B4
- ;; __c6xabi_remu A1,A4,A5,A7,B0,B1,B2,B4
- ;;
- ;; In our implementation, divu and remu are leaf functions,
- ;; while both divi and remi call into divu.
- ;; A0 is not clobbered by any of the functions.
- ;; divu does not clobber B2 either, which is taken advantage of
- ;; in remi.
- ;; divi uses B5 to hold the original return address during
- ;; the call to divu.
- ;; remi uses B2 and A5 to hold the input values during the
- ;; call to divu. It stores B3 in on the stack.
-
-#ifdef L_divsi3
-.text
-.align 2
-.global __c6xabi_divi
-.hidden __c6xabi_divi
-.type __c6xabi_divi, STT_FUNC
-
-__c6xabi_divi:
- call .s2 __c6xabi_divu
-|| mv .d2 B3, B5
-|| cmpgt .l1 0, A4, A1
-|| cmpgt .l2 0, B4, B1
-
- [A1] neg .l1 A4, A4
-|| [B1] neg .l2 B4, B4
-|| xor .s1x A1, B1, A1
-
-#ifdef _TMS320C6400
- [A1] addkpc .s2 1f, B3, 4
-#else
- [A1] mvkl .s2 1f, B3
- [A1] mvkh .s2 1f, B3
- nop 2
-#endif
-1:
- neg .l1 A4, A4
-|| mv .l2 B3,B5
-|| ret .s2 B5
- nop 5
-#endif
-
-#if defined L_modsi3 || defined L_divmodsi4
-.align 2
-#ifdef L_modsi3
-#define MOD_OUTPUT_REG A4
-.global __c6xabi_remi
-.hidden __c6xabi_remi
-.type __c6xabi_remi, STT_FUNC
-#else
-#define MOD_OUTPUT_REG A5
-.global __c6xabi_divremi
-.hidden __c6xabi_divremi
-.type __c6xabi_divremi, STT_FUNC
-__c6xabi_divremi:
-#endif
-
-__c6xabi_remi:
- stw .d2t2 B3, *B15--[2]
-|| cmpgt .l1 0, A4, A1
-|| cmpgt .l2 0, B4, B2
-|| mv .s1 A4, A5
-|| call .s2 __c6xabi_divu
-
- [A1] neg .l1 A4, A4
-|| [B2] neg .l2 B4, B4
-|| xor .s2x B2, A1, B0
-|| mv .d2 B4, B2
-
-#ifdef _TMS320C6400
- [B0] addkpc .s2 1f, B3, 1
- [!B0] addkpc .s2 2f, B3, 1
- nop 2
-#else
- [B0] mvkl .s2 1f,B3
- [!B0] mvkl .s2 2f,B3
-
- [B0] mvkh .s2 1f,B3
- [!B0] mvkh .s2 2f,B3
-#endif
-1:
- neg .l1 A4, A4
-2:
- ldw .d2t2 *++B15[2], B3
-
-#ifdef _TMS320C6400_PLUS
- mpy32 .m1x A4, B2, A6
- nop 3
- ret .s2 B3
- sub .l1 A5, A6, MOD_OUTPUT_REG
- nop 4
-#else
- mpyu .m1x A4, B2, A1
- nop 1
- mpylhu .m1x A4, B2, A6
-|| mpylhu .m2x B2, A4, B2
- nop 1
- add .l1x A6, B2, A6
-|| ret .s2 B3
- shl .s1 A6, 16, A6
- add .d1 A6, A1, A6
- sub .l1 A5, A6, MOD_OUTPUT_REG
- nop 2
-#endif
-
-#endif
-
-#if defined L_udivsi3 || defined L_udivmodsi4
-.align 2
-#ifdef L_udivsi3
-.global __c6xabi_divu
-.hidden __c6xabi_divu
-.type __c6xabi_divu, STT_FUNC
-__c6xabi_divu:
-#else
-.global __c6xabi_divremu
-.hidden __c6xabi_divremu
-.type __c6xabi_divremu, STT_FUNC
-__c6xabi_divremu:
-#endif
- ;; We use a series of up to 31 subc instructions. First, we find
- ;; out how many leading zero bits there are in the divisor. This
- ;; gives us both a shift count for aligning (shifting) the divisor
- ;; to the, and the number of times we have to execute subc.
-
- ;; At the end, we have both the remainder and most of the quotient
- ;; in A4. The top bit of the quotient is computed first and is
- ;; placed in A2.
-
- ;; Return immediately if the dividend is zero. Setting B4 to 1
- ;; is a trick to allow us to leave the following insns in the jump
- ;; delay slot without affecting the result.
- mv .s2x A4, B1
-
-#ifndef _TMS320C6400
-[!b1] mvk .s2 1, B4
-#endif
-[b1] lmbd .l2 1, B4, B1
-||[!b1] b .s2 B3 ; RETURN A
-#ifdef _TMS320C6400
-||[!b1] mvk .d2 1, B4
-#endif
-#ifdef L_udivmodsi4
-||[!b1] zero .s1 A5
-#endif
- mv .l1x B1, A6
-|| shl .s2 B4, B1, B4
-
- ;; The loop performs a maximum of 28 steps, so we do the
- ;; first 3 here.
- cmpltu .l1x A4, B4, A2
-[!A2] sub .l1x A4, B4, A4
-|| shru .s2 B4, 1, B4
-|| xor .s1 1, A2, A2
-
- shl .s1 A2, 31, A2
-|| [b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-
- ;; RETURN A may happen here (note: must happen before the next branch)
-0:
- cmpgt .l2 B1, 7, B0
-|| [b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-|| [b0] b .s1 0b
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
- ;; loop backwards branch happens here
-
- ret .s2 B3
-|| mvk .s1 32, A1
- sub .l1 A1, A6, A6
-#ifdef L_udivmodsi4
-|| extu .s1 A4, A6, A5
-#endif
- shl .s1 A4, A6, A4
- shru .s1 A4, 1, A4
-|| sub .l1 A6, 1, A6
- or .l1 A2, A4, A4
- shru .s1 A4, A6, A4
- nop
-
-#endif
-
-#ifdef L_umodsi3
-.align 2
-.global __c6xabi_remu
-.hidden __c6xabi_remu
-.type __c6xabi_remu, STT_FUNC
-__c6xabi_remu:
- ;; The ABI seems designed to prevent these functions calling each other,
- ;; so we duplicate most of the divsi3 code here.
- mv .s2x A4, B1
-#ifndef _TMS320C6400
-[!b1] mvk .s2 1, B4
-#endif
- lmbd .l2 1, B4, B1
-||[!b1] b .s2 B3 ; RETURN A
-#ifdef _TMS320C6400
-||[!b1] mvk .d2 1, B4
-#endif
-
- mv .l1x B1, A7
-|| shl .s2 B4, B1, B4
-
- cmpltu .l1x A4, B4, A1
-[!a1] sub .l1x A4, B4, A4
- shru .s2 B4, 1, B4
-
-0:
- cmpgt .l2 B1, 7, B0
-|| [b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
- ;; RETURN A may happen here (note: must happen before the next branch)
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-|| [b0] b .s1 0b
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
- ;; loop backwards branch happens here
-
- ret .s2 B3
-[b1] subc .l1x A4,B4,A4
-|| [b1] add .s2 -1, B1, B1
-[b1] subc .l1x A4,B4,A4
-
- extu .s1 A4, A7, A4
- nop 2
-#endif
-
-#if defined L_strasgi_64plus && defined _TMS320C6400_PLUS
-
-.align 2
-.global __c6xabi_strasgi_64plus
-.hidden __c6xabi_strasgi_64plus
-.type __c6xabi_strasgi_64plus, STT_FUNC
-__c6xabi_strasgi_64plus:
- shru .s2x a6, 2, b31
-|| mv .s1 a4, a30
-|| mv .d2 b4, b30
-
- add .s2 -4, b31, b31
-
- sploopd 1
-|| mvc .s2 b31, ilc
- ldw .d2t2 *b30++, b31
- nop 4
- mv .s1x b31,a31
- spkernel 6, 0
-|| stw .d1t1 a31, *a30++
-
- ret .s2 b3
- nop 5
-#endif
-
-#ifdef L_strasgi
-.global __c6xabi_strasgi
-.type __c6xabi_strasgi, STT_FUNC
-__c6xabi_strasgi:
- ;; This is essentially memcpy, with alignment known to be at least
- ;; 4, and the size a multiple of 4 greater than or equal to 28.
- ldw .d2t1 *B4++, A0
-|| mvk .s2 16, B1
- ldw .d2t1 *B4++, A1
-|| mvk .s2 20, B2
-|| sub .d1 A6, 24, A6
- ldw .d2t1 *B4++, A5
- ldw .d2t1 *B4++, A7
-|| mv .l2x A6, B7
- ldw .d2t1 *B4++, A8
- ldw .d2t1 *B4++, A9
-|| mv .s2x A0, B5
-|| cmpltu .l2 B2, B7, B0
-
-0:
- stw .d1t2 B5, *A4++
-||[b0] ldw .d2t1 *B4++, A0
-|| mv .s2x A1, B5
-|| mv .l2 B7, B6
-
-[b0] sub .d2 B6, 24, B7
-||[b0] b .s2 0b
-|| cmpltu .l2 B1, B6, B0
-
-[b0] ldw .d2t1 *B4++, A1
-|| stw .d1t2 B5, *A4++
-|| mv .s2x A5, B5
-|| cmpltu .l2 12, B6, B0
-
-[b0] ldw .d2t1 *B4++, A5
-|| stw .d1t2 B5, *A4++
-|| mv .s2x A7, B5
-|| cmpltu .l2 8, B6, B0
-
-[b0] ldw .d2t1 *B4++, A7
-|| stw .d1t2 B5, *A4++
-|| mv .s2x A8, B5
-|| cmpltu .l2 4, B6, B0
-
-[b0] ldw .d2t1 *B4++, A8
-|| stw .d1t2 B5, *A4++
-|| mv .s2x A9, B5
-|| cmpltu .l2 0, B6, B0
-
-[b0] ldw .d2t1 *B4++, A9
-|| stw .d1t2 B5, *A4++
-|| mv .s2x A0, B5
-|| cmpltu .l2 B2, B7, B0
-
- ;; loop back branch happens here
-
- cmpltu .l2 B1, B6, B0
-|| ret .s2 b3
-
-[b0] stw .d1t1 A1, *A4++
-|| cmpltu .l2 12, B6, B0
-[b0] stw .d1t1 A5, *A4++
-|| cmpltu .l2 8, B6, B0
-[b0] stw .d1t1 A7, *A4++
-|| cmpltu .l2 4, B6, B0
-[b0] stw .d1t1 A8, *A4++
-|| cmpltu .l2 0, B6, B0
-[b0] stw .d1t1 A9, *A4++
-
- ;; return happens here
-
-#endif
-
-#ifdef _TMS320C6400_PLUS
-#ifdef L_push_rts
-.align 2
-.global __c6xabi_push_rts
-.hidden __c6xabi_push_rts
-.type __c6xabi_push_rts, STT_FUNC
-__c6xabi_push_rts:
- stw .d2t2 B14, *B15--[2]
- stdw .d2t1 A15:A14, *B15--
-|| b .s2x A3
- stdw .d2t2 B13:B12, *B15--
- stdw .d2t1 A13:A12, *B15--
- stdw .d2t2 B11:B10, *B15--
- stdw .d2t1 A11:A10, *B15--
- stdw .d2t2 B3:B2, *B15--
-#endif
-
-#ifdef L_pop_rts
-.align 2
-.global __c6xabi_pop_rts
-.hidden __c6xabi_pop_rts
-.type __c6xabi_pop_rts, STT_FUNC
-__c6xabi_pop_rts:
- lddw .d2t2 *++B15, B3:B2
- lddw .d2t1 *++B15, A11:A10
- lddw .d2t2 *++B15, B11:B10
- lddw .d2t1 *++B15, A13:A12
- lddw .d2t2 *++B15, B13:B12
- lddw .d2t1 *++B15, A15:A14
-|| b .s2 B3
- ldw .d2t2 *++B15[2], B14
- nop 4
-#endif
-
-#ifdef L_call_stub
-.align 2
-.global __c6xabi_call_stub
-.type __c6xabi_call_stub, STT_FUNC
-__c6xabi_call_stub:
- stw .d2t1 A2, *B15--[2]
- stdw .d2t1 A7:A6, *B15--
-|| call .s2 B31
- stdw .d2t1 A1:A0, *B15--
- stdw .d2t2 B7:B6, *B15--
- stdw .d2t2 B5:B4, *B15--
- stdw .d2t2 B1:B0, *B15--
- stdw .d2t2 B3:B2, *B15--
-|| addkpc .s2 1f, B3, 0
-1:
- lddw .d2t2 *++B15, B3:B2
- lddw .d2t2 *++B15, B1:B0
- lddw .d2t2 *++B15, B5:B4
- lddw .d2t2 *++B15, B7:B6
- lddw .d2t1 *++B15, A1:A0
- lddw .d2t1 *++B15, A7:A6
-|| b .s2 B3
- ldw .d2t1 *++B15[2], A2
- nop 4
-#endif
-
-#endif
-