summaryrefslogtreecommitdiff
path: root/libc/sysdeps/powerpc
diff options
context:
space:
mode:
Diffstat (limited to 'libc/sysdeps/powerpc')
-rw-r--r--libc/sysdeps/powerpc/bits/link.h2
-rw-r--r--libc/sysdeps/powerpc/dl-procinfo.c12
-rw-r--r--libc/sysdeps/powerpc/dl-procinfo.h36
-rw-r--r--libc/sysdeps/powerpc/fpu/feholdexcpt.c2
-rw-r--r--libc/sysdeps/powerpc/fpu/fenv_const.c8
-rw-r--r--libc/sysdeps/powerpc/fpu/fesetenv.c8
-rw-r--r--libc/sysdeps/powerpc/fpu/feupdateenv.c6
-rw-r--r--libc/sysdeps/powerpc/fpu/s_rintf.c2
-rw-r--r--libc/sysdeps/powerpc/fpu/t_sqrt.c2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/bits/atomic.h2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/dl-machine.c2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_floor.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_floorf.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_isnan.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_lround.S10
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_round.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_roundf.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_trunc.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/fpu/s_truncf.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/lshift.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrint.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrintf.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llround.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrt.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrtf.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/hp-timing.c2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/memcmp.S126
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/memset.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power4/wordcopy.c8
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_llround.S8
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_lround.S8
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrt.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrtf.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrint.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrintf.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llround.S8
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/memcpy.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/wcschr.c6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/wcsrchr.c6
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6/wordcopy.c8
-rw-r--r--libc/sysdeps/powerpc/powerpc32/power6x/fpu/s_lround.S8
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_ceill.S6
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_floor.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_floorf.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_isnan.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_llrint.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_llrintf.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_llround.S8
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_llroundf.S8
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_round.S10
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_roundf.S10
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_roundl.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_trunc.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_truncf.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/fpu/s_truncl.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/hp-timing.c2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrt.c4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrtf.c4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power4/memcmp.S114
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power4/memcpy.S70
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power5+/fpu/s_llround.S10
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power5/fpu/s_isnan.S4
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power6/memcpy.S76
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power6/wordcopy.c8
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_isnan.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llrint.S2
-rw-r--r--libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llround.S10
-rw-r--r--libc/sysdeps/powerpc/powerpc64/strlen.S10
-rw-r--r--libc/sysdeps/powerpc/powerpc64/strncmp.S10
-rw-r--r--libc/sysdeps/powerpc/sysdep.h4
71 files changed, 391 insertions, 367 deletions
diff --git a/libc/sysdeps/powerpc/bits/link.h b/libc/sysdeps/powerpc/bits/link.h
index b907bd608..f06092f10 100644
--- a/libc/sysdeps/powerpc/bits/link.h
+++ b/libc/sysdeps/powerpc/bits/link.h
@@ -83,7 +83,7 @@ typedef struct La_ppc64_retval
uint64_t lrv_r3;
uint64_t lrv_r4;
double lrv_fp[4]; /* f1-f4, float - complex long double. */
- uint32_t lrv_v2[4]; /* v2. */
+ uint32_t lrv_v2[4]; /* v2. */
} La_ppc64_retval;
diff --git a/libc/sysdeps/powerpc/dl-procinfo.c b/libc/sysdeps/powerpc/dl-procinfo.c
index fb5051f1b..0f5c2b3d6 100644
--- a/libc/sysdeps/powerpc/dl-procinfo.c
+++ b/libc/sysdeps/powerpc/dl-procinfo.c
@@ -13,7 +13,7 @@
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
+ License along with the GNU C Library. If not, see
<http://www.gnu.org/licenses/>. */
/* This information must be kept in sync with the _DL_HWCAP_COUNT and
@@ -49,7 +49,7 @@ PROCINFO_CLASS const char _dl_powerpc_cap_flags[25][10]
#endif
#ifndef PROCINFO_DECL
= {
- "vsx",
+ "vsx",
"arch_2_06", "power6x", "dfp", "pa6t",
"arch_2_05", "ic_snoop", "smt", "booke",
"cellbe", "power5+", "power5", "power4",
@@ -67,7 +67,7 @@ PROCINFO_CLASS const char _dl_powerpc_cap_flags[25][10]
#if !defined PROCINFO_DECL && defined SHARED
._dl_powerpc_platforms
#else
-PROCINFO_CLASS const char _dl_powerpc_platforms[9][12]
+PROCINFO_CLASS const char _dl_powerpc_platforms[13][12]
#endif
#ifndef PROCINFO_DECL
= {
@@ -79,7 +79,11 @@ PROCINFO_CLASS const char _dl_powerpc_platforms[9][12]
[PPC_PLATFORM_CELL_BE] = "ppc-cell-be",
[PPC_PLATFORM_POWER6X] = "power6x",
[PPC_PLATFORM_POWER7] = "power7",
- [PPC_PLATFORM_PPCA2] = "ppca2"
+ [PPC_PLATFORM_PPCA2] = "ppca2",
+ [PPC_PLATFORM_PPC405] = "ppc405",
+ [PPC_PLATFORM_PPC440] = "ppc440",
+ [PPC_PLATFORM_PPC464] = "ppc464",
+ [PPC_PLATFORM_PPC476] = "ppc476"
}
#endif
#if !defined SHARED || defined PROCINFO_DECL
diff --git a/libc/sysdeps/powerpc/dl-procinfo.h b/libc/sysdeps/powerpc/dl-procinfo.h
index 19aa93b18..568fe19dd 100644
--- a/libc/sysdeps/powerpc/dl-procinfo.h
+++ b/libc/sysdeps/powerpc/dl-procinfo.h
@@ -13,14 +13,14 @@
Lesser General Public License for more details.
You should have received a copy of the GNU Lesser General Public
- License along with the GNU C Library; if not, see
+ License along with the GNU C Library. If not, see
<http://www.gnu.org/licenses/>. */
#ifndef _DL_PROCINFO_H
-#define _DL_PROCINFO_H 1
+#define _DL_PROCINFO_H 1
#include <ldsodefs.h>
-#include <sysdep.h> /* This defines the PPC_FEATURE_* macros. */
+#include <sysdep.h> /* This defines the PPC_FEATURE_* macros. */
/* There are 25 bits used, but they are bits 7..31. */
#define _DL_HWCAP_FIRST 7
@@ -30,12 +30,12 @@
#define HWCAP_IMPORTANT (PPC_FEATURE_HAS_ALTIVEC \
+ PPC_FEATURE_HAS_DFP)
-#define _DL_PLATFORMS_COUNT 9
+#define _DL_PLATFORMS_COUNT 13
-#define _DL_FIRST_PLATFORM 32
+#define _DL_FIRST_PLATFORM 32
/* Mask to filter out platforms. */
-#define _DL_HWCAP_PLATFORM (((1ULL << _DL_PLATFORMS_COUNT) - 1) \
- << _DL_FIRST_PLATFORM)
+#define _DL_HWCAP_PLATFORM (((1ULL << _DL_PLATFORMS_COUNT) - 1) \
+ << _DL_FIRST_PLATFORM)
/* Platform bits (relative to _DL_FIRST_PLATFORM). */
#define PPC_PLATFORM_POWER4 0
@@ -47,6 +47,10 @@
#define PPC_PLATFORM_POWER6X 6
#define PPC_PLATFORM_POWER7 7
#define PPC_PLATFORM_PPCA2 8
+#define PPC_PLATFORM_PPC405 9
+#define PPC_PLATFORM_PPC440 10
+#define PPC_PLATFORM_PPC464 11
+#define PPC_PLATFORM_PPC476 12
static inline const char *
__attribute__ ((unused))
@@ -111,7 +115,7 @@ _dl_string_platform (const char *str)
return -1;
}
if (str[1] == '\0')
- return ret;
+ return ret;
}
else if (strncmp (str, GLRO(dl_powerpc_platforms)[PPC_PLATFORM_PPC970],
3) == 0)
@@ -127,6 +131,22 @@ _dl_string_platform (const char *str)
GLRO(dl_powerpc_platforms)[PPC_PLATFORM_PPCA2] + 3)
== 0)
return _DL_FIRST_PLATFORM + PPC_PLATFORM_PPCA2;
+ else if (strcmp (str + 3,
+ GLRO(dl_powerpc_platforms)[PPC_PLATFORM_PPC405] + 3)
+ == 0)
+ return _DL_FIRST_PLATFORM + PPC_PLATFORM_PPC405;
+ else if (strcmp (str + 3,
+ GLRO(dl_powerpc_platforms)[PPC_PLATFORM_PPC440] + 3)
+ == 0)
+ return _DL_FIRST_PLATFORM + PPC_PLATFORM_PPC440;
+ else if (strcmp (str + 3,
+ GLRO(dl_powerpc_platforms)[PPC_PLATFORM_PPC464] + 3)
+ == 0)
+ return _DL_FIRST_PLATFORM + PPC_PLATFORM_PPC464;
+ else if (strcmp (str + 3,
+ GLRO(dl_powerpc_platforms)[PPC_PLATFORM_PPC476] + 3)
+ == 0)
+ return _DL_FIRST_PLATFORM + PPC_PLATFORM_PPC476;
}
return -1;
diff --git a/libc/sysdeps/powerpc/fpu/feholdexcpt.c b/libc/sysdeps/powerpc/fpu/feholdexcpt.c
index 671724b28..013d2bfbb 100644
--- a/libc/sysdeps/powerpc/fpu/feholdexcpt.c
+++ b/libc/sysdeps/powerpc/fpu/feholdexcpt.c
@@ -32,7 +32,7 @@ feholdexcept (fenv_t *envp)
flag. */
new.l[1] = old.l[1] & 7;
new.l[0] = old.l[0];
-
+
/* If the old env had any enabled exceptions, then mask SIGFPE in the
MSR FE0/FE1 bits. This may allow the FPU to run faster because it
always takes the default action and can not generate SIGFPE. */
diff --git a/libc/sysdeps/powerpc/fpu/fenv_const.c b/libc/sysdeps/powerpc/fpu/fenv_const.c
index d7bb79c74..47761ebec 100644
--- a/libc/sysdeps/powerpc/fpu/fenv_const.c
+++ b/libc/sysdeps/powerpc/fpu/fenv_const.c
@@ -16,17 +16,17 @@
License along with the GNU C Library; if not, see
<http://www.gnu.org/licenses/>. */
-/* We want to specify the bit pattern of the __fe_*_env constants, so
+/* We want to specify the bit pattern of the __fe_*_env constants, so
pretend they're really `long long' instead of `double'. */
/* If the default argument is used we use this value. */
-const unsigned long long __fe_dfl_env __attribute__ ((aligned (8))) =
+const unsigned long long __fe_dfl_env __attribute__ ((aligned (8))) =
0xfff8000000000000ULL;
/* Floating-point environment where none of the exceptions are masked. */
-const unsigned long long __fe_enabled_env __attribute__ ((aligned (8))) =
+const unsigned long long __fe_enabled_env __attribute__ ((aligned (8))) =
0xfff80000000000f8ULL;
/* Floating-point environment with the NI bit set. */
-const unsigned long long __fe_nonieee_env __attribute__ ((aligned (8))) =
+const unsigned long long __fe_nonieee_env __attribute__ ((aligned (8))) =
0xfff8000000000004ULL;
diff --git a/libc/sysdeps/powerpc/fpu/fesetenv.c b/libc/sysdeps/powerpc/fpu/fesetenv.c
index 953af5ddc..e92adb4c5 100644
--- a/libc/sysdeps/powerpc/fpu/fesetenv.c
+++ b/libc/sysdeps/powerpc/fpu/fesetenv.c
@@ -29,21 +29,21 @@ __fesetenv (const fenv_t *envp)
/* get the currently set exceptions. */
new.fenv = *envp;
old.fenv = fegetenv_register ();
-
+
/* If the old env has no enabled exceptions and the new env has any enabled
exceptions, then unmask SIGFPE in the MSR FE0/FE1 bits. This will put the
hardware into "precise mode" and may cause the FPU to run slower on some
hardware. */
if ((old.l[1] & _FPU_MASK_ALL) == 0 && (new.l[1] & _FPU_MASK_ALL) != 0)
(void)__fe_nomask_env ();
-
+
/* If the old env had any enabled exceptions and the new env has no enabled
exceptions, then mask SIGFPE in the MSR FE0/FE1 bits. This may allow the
- FPU to run faster because it always takes the default action and can not
+ FPU to run faster because it always takes the default action and can not
generate SIGFPE. */
if ((old.l[1] & _FPU_MASK_ALL) != 0 && (new.l[1] & _FPU_MASK_ALL) == 0)
(void)__fe_mask_env ();
-
+
fesetenv_register (*envp);
/* Success. */
diff --git a/libc/sysdeps/powerpc/fpu/feupdateenv.c b/libc/sysdeps/powerpc/fpu/feupdateenv.c
index 9faf930f3..6500ea173 100644
--- a/libc/sysdeps/powerpc/fpu/feupdateenv.c
+++ b/libc/sysdeps/powerpc/fpu/feupdateenv.c
@@ -35,17 +35,17 @@ __feupdateenv (const fenv_t *envp)
exceptions. Leave fraction rounded/inexact and FP result/CC bits
unchanged. */
new.l[1] = (old.l[1] & 0x1FFFFF00) | (new.l[1] & 0x1FF80FFF);
-
+
/* If the old env has no enabled exceptions and the new env has any enabled
exceptions, then unmask SIGFPE in the MSR FE0/FE1 bits. This will put
the hardware into "precise mode" and may cause the FPU to run slower on
some hardware. */
if ((old.l[1] & _FPU_MASK_ALL) == 0 && (new.l[1] & _FPU_MASK_ALL) != 0)
(void)__fe_nomask_env ();
-
+
/* If the old env had any enabled exceptions and the new env has no enabled
exceptions, then mask SIGFPE in the MSR FE0/FE1 bits. This may allow the
- FPU to run faster because it always takes the default action and can not
+ FPU to run faster because it always takes the default action and can not
generate SIGFPE. */
if ((old.l[1] & _FPU_MASK_ALL) != 0 && (new.l[1] & _FPU_MASK_ALL) == 0)
(void)__fe_mask_env ();
diff --git a/libc/sysdeps/powerpc/fpu/s_rintf.c b/libc/sysdeps/powerpc/fpu/s_rintf.c
index 7b1476780..4a32a4343 100644
--- a/libc/sysdeps/powerpc/fpu/s_rintf.c
+++ b/libc/sysdeps/powerpc/fpu/s_rintf.c
@@ -36,7 +36,7 @@ __rintf (float x)
x = -(x - TWO23);
}
}
-
+
return x;
}
weak_alias (__rintf, rintf)
diff --git a/libc/sysdeps/powerpc/fpu/t_sqrt.c b/libc/sysdeps/powerpc/fpu/t_sqrt.c
index c49380c0f..9ed7436ae 100644
--- a/libc/sysdeps/powerpc/fpu/t_sqrt.c
+++ b/libc/sysdeps/powerpc/fpu/t_sqrt.c
@@ -50,7 +50,7 @@ const float __t_sqrt[1024] = {
0.9847,0.5077, 0.9857,0.5072, 0.9867,0.5067, 0.9877,0.5062, 0.9887,0.5057,
0.9897,0.5052, 0.9907,0.5047, 0.9917,0.5042, 0.9926,0.5037, 0.9936,0.5032,
0.9946,0.5027, 0.9956,0.5022, 0.9966,0.5017, 0.9976,0.5012, 0.9985,0.5007,
-0.9995,0.5002,
+0.9995,0.5002,
1.0010,0.4995, 1.0029,0.4985, 1.0049,0.4976, 1.0068,0.4966, 1.0088,0.4957,
1.0107,0.4947, 1.0126,0.4938, 1.0145,0.4928, 1.0165,0.4919, 1.0184,0.4910,
1.0203,0.4901, 1.0222,0.4891, 1.0241,0.4882, 1.0260,0.4873, 1.0279,0.4864,
diff --git a/libc/sysdeps/powerpc/powerpc32/bits/atomic.h b/libc/sysdeps/powerpc/powerpc32/bits/atomic.h
index 3e3a1effe..a09e4d441 100644
--- a/libc/sysdeps/powerpc/powerpc32/bits/atomic.h
+++ b/libc/sysdeps/powerpc/powerpc32/bits/atomic.h
@@ -24,7 +24,7 @@
atomic operation. In that case we don't expect additional updates
adjacent to the lock word after the Store Conditional and the hint
should be false. */
-
+
#if defined _ARCH_PWR6 || defined _ARCH_PWR6X
# define MUTEX_HINT_ACQ ",1"
# define MUTEX_HINT_REL ",0"
diff --git a/libc/sysdeps/powerpc/powerpc32/dl-machine.c b/libc/sysdeps/powerpc/powerpc32/dl-machine.c
index 475d238ba..170764ff7 100644
--- a/libc/sysdeps/powerpc/powerpc32/dl-machine.c
+++ b/libc/sysdeps/powerpc/powerpc32/dl-machine.c
@@ -129,7 +129,7 @@ __elf_preferred_address (struct link_map *loader, size_t maplength,
_dl_loaded does not work for static binaries loading
e.g. libnss_*.so. */
if ((mapend >= high || l->l_type == lt_executable)
- && high >= mapstart)
+ && high >= mapstart)
high = mapstart;
else if (mapend >= low && low >= mapstart)
low = mapend;
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_floor.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_floor.S
index 21154ae20..168bb17b7 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_floor.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_floor.S
@@ -59,7 +59,7 @@ ENTRY (__floor)
fnabs fp1,fp1 /* if (x == 0.0) */
/* x = -0.0; */
.L9:
- mtfsf 0x01,fp11 /* restore previous rounding mode. */
+ mtfsf 0x01,fp11 /* restore previous rounding mode. */
blr
END (__floor)
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_floorf.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_floorf.S
index e5faf2c64..4d6e90c9b 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_floorf.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_floorf.S
@@ -58,7 +58,7 @@ ENTRY (__floorf)
fnabs fp1,fp1 /* if (x == 0.0) */
/* x = -0.0; */
.L9:
- mtfsf 0x01,fp11 /* restore previous rounding mode. */
+ mtfsf 0x01,fp11 /* restore previous rounding mode. */
blr
END (__floorf)
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_isnan.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_isnan.S
index ac8b08856..98d10daf6 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_isnan.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_isnan.S
@@ -21,7 +21,7 @@
/* int __isnan(x) */
.machine power4
-EALIGN (__isnan, 4, 0)
+EALIGN (__isnan, 4, 0)
mffs fp0
mtfsb0 4*cr6+lt /* reset_fpscr_bit (FPSCR_VE) */
fcmpu cr7,fp1,fp1
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_lround.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_lround.S
index d1472ce76..92dc3787d 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_lround.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_lround.S
@@ -24,13 +24,13 @@
.LC0: /* 0.5 */
.long 0x3f000000
.section ".text"
-
+
/* long [r3] lround (float x [fp1])
- IEEE 1003.1 lround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 lround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "round to Nearest" mode. Instead we set
"round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value. It is necessary to detect when x is
@@ -60,7 +60,7 @@ ENTRY (__lround)
blt- cr6,.Lretzero
fadd fp3,fp2,fp10 /* |x|+=0.5 bias to prepare to round. */
bge cr7,.Lconvert /* x is positive so don't negate x. */
- fnabs fp3,fp3 /* -(|x|+=0.5) */
+ fnabs fp3,fp3 /* -(|x|+=0.5) */
.Lconvert:
fctiwz fp4,fp3 /* Convert to Integer word lround toward 0. */
stfd fp4,8(r1)
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_round.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_round.S
index 030204f85..73a100422 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_round.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_round.S
@@ -27,11 +27,11 @@
.long 0x3f000000
/* double [fp1] round (double x [fp1])
- IEEE 1003.1 round function. IEEE specifies "round to the nearest
+ IEEE 1003.1 round function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "Round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "Round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "Round to Nearest" mode. Instead we set
"Round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value. */
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_roundf.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_roundf.S
index adf4d3530..2ed9ca7b4 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_roundf.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_roundf.S
@@ -26,11 +26,11 @@
.long 0x3f000000
/* float [fp1] roundf (float x [fp1])
- IEEE 1003.1 round function. IEEE specifies "round to the nearest
+ IEEE 1003.1 round function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "Round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "Round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "Round to Nearest" mode. Instead we set
"Round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value. */
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_trunc.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_trunc.S
index fe3e2b975..0f9e3600d 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_trunc.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_trunc.S
@@ -26,7 +26,7 @@
/* double [fp1] trunc (double x [fp1])
IEEE 1003.1 trunc function. IEEE specifies "trunc to the integer
- value, in floating format, nearest to but no larger in magnitude
+ value, in floating format, nearest to but no larger in magnitude
then the argument."
We set "round toward Zero" mode and trunc by adding +-2**52 then
subtracting +-2**52. */
diff --git a/libc/sysdeps/powerpc/powerpc32/fpu/s_truncf.S b/libc/sysdeps/powerpc/powerpc32/fpu/s_truncf.S
index a74e3dc38..effbede6d 100644
--- a/libc/sysdeps/powerpc/powerpc32/fpu/s_truncf.S
+++ b/libc/sysdeps/powerpc/powerpc32/fpu/s_truncf.S
@@ -25,7 +25,7 @@
/* float [fp1] truncf (float x [fp1])
IEEE 1003.1 trunc function. IEEE specifies "trunc to the integer
- value, in floating format, nearest to but no larger in magnitude
+ value, in floating format, nearest to but no larger in magnitude
then the argument."
We set "round toward Zero" mode and trunc by adding +-2**23 then
subtracting +-2**23. */
diff --git a/libc/sysdeps/powerpc/powerpc32/lshift.S b/libc/sysdeps/powerpc/powerpc32/lshift.S
index 1d83910b3..9ee58e6a5 100644
--- a/libc/sysdeps/powerpc/powerpc32/lshift.S
+++ b/libc/sysdeps/powerpc/powerpc32/lshift.S
@@ -19,7 +19,7 @@
#include <sysdep.h>
/* mp_limb_t mpn_lshift (mp_ptr wp, mp_srcptr up, mp_size_t usize,
- unsigned int cnt) */
+ unsigned int cnt) */
EALIGN (__mpn_lshift, 3, 0)
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrint.S b/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrint.S
index 2ac986db8..55b2850fd 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrint.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrint.S
@@ -20,7 +20,7 @@
#include <math_ldbl_opt.h>
/* long long int[r3, r4] __llrint (double x[fp1]) */
-ENTRY (__llrint)
+ENTRY (__llrint)
CALL_MCOUNT
stwu r1,-16(r1)
cfi_adjust_cfa_offset (16)
@@ -31,7 +31,7 @@ ENTRY (__llrint)
nop
lwz r3,8(r1)
lwz r4,12(r1)
- addi r1,r1,16
+ addi r1,r1,16
blr
END (__llrint)
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrintf.S b/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrintf.S
index 98e3aafc8..cc80fcb02 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrintf.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llrintf.S
@@ -19,7 +19,7 @@
#include <sysdep.h>
/* long long int[r3, r4] __llrintf (float x[fp1]) */
-ENTRY (__llrintf)
+ENTRY (__llrintf)
CALL_MCOUNT
stwu r1,-16(r1)
cfi_adjust_cfa_offset (16)
@@ -30,7 +30,7 @@ ENTRY (__llrintf)
nop
lwz r3,8(r1)
lwz r4,12(r1)
- addi r1,r1,16
+ addi r1,r1,16
blr
END (__llrintf)
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llround.S b/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llround.S
index 07beb0a56..631180f07 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llround.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/fpu/s_llround.S
@@ -29,11 +29,11 @@
.section ".text"
/* long [r3] lround (float x [fp1])
- IEEE 1003.1 lround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 lround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "round to Nearest" mode. Instead we set
"round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value.
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrt.S b/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrt.S
index 4f1c17680..3648e4a69 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrt.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrt.S
@@ -25,20 +25,20 @@
sets the appropriate floating point exceptions. Extended checking is
only needed to set errno (via __kernel_standard) if the input value
is negative.
-
+
The fsqrt will set FPCC and FU (Floating Point Unordered or NaN
to indicated that the input value was negative or NaN. Use Move to
Condition Register from FPSCR to copy the FPCC field to cr1. The
branch on summary overflow transfers control to w_sqrt to process
any error conditions. Otherwise we can return the result directly.
-
+
This part of the function is a leaf routine, so no need to stack a
frame or execute prologue/epilogue code. This means it is safe to
transfer directly to w_sqrt as long as the input value (f1) is
preserved. Putting the sqrt result into f2 (double parameter 2)
allows passing both the input value and sqrt result into the extended
wrapper so there is no need to recompute.
-
+
This tactic avoids the overhead of stacking a frame for the normal
(non-error) case. Until gcc supports prologue shrink-wrapping
this is the best we can do. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrtf.S b/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrtf.S
index 0da5b7a8e..153843c7c 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrtf.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/fpu/w_sqrtf.S
@@ -25,20 +25,20 @@
sets the appropriate floating point exceptions. Extended checking is
only needed to set errno (via __kernel_standard) if the input value
is negative.
-
+
The fsqrts will set FPCC and FU (Floating Point Unordered or NaN
to indicated that the input value was negative or NaN. Use Move to
Condition Register from FPSCR to copy the FPCC field to cr1. The
branch on summary overflow transfers control to w_sqrt to process
any error conditions. Otherwise we can return the result directly.
-
+
This part of the function is a leaf routine, so no need to stack a
frame or execute prologue/epilogue code. This means it is safe to
transfer directly to w_sqrt as long as the input value (f1) is
preserved. Putting the sqrt result into f2 (float parameter 2)
allows passing both the input value and sqrt result into the extended
wrapper so there is no need to recompute.
-
+
This tactic avoids the overhead of stacking a frame for the normal
(non-error) case. Until gcc supports prologue shrink-wrapping
this is the best we can do. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/hp-timing.c b/libc/sysdeps/powerpc/powerpc32/power4/hp-timing.c
index f54a5f879..5073adb0e 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/hp-timing.c
+++ b/libc/sysdeps/powerpc/powerpc32/power4/hp-timing.c
@@ -1,4 +1,4 @@
-/* Support for high precision, low overhead timing functions.
+/* Support for high precision, low overhead timing functions.
powerpc64 version.
Copyright (C) 2005-2013 Free Software Foundation, Inc.
This file is part of the GNU C Library.
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/memcmp.S b/libc/sysdeps/powerpc/powerpc32/power4/memcmp.S
index edec7ab27..d7050a2f7 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/memcmp.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/memcmp.S
@@ -53,17 +53,17 @@ EALIGN (memcmp, 4, 0)
blt cr1, L(bytealigned)
stwu 1,-64(1)
cfi_adjust_cfa_offset(64)
- stw r31,48(1)
+ stw r31,48(1)
cfi_offset(31,(48-64))
- stw r30,44(1)
+ stw r30,44(1)
cfi_offset(30,(44-64))
bne L(unaligned)
/* At this point we know both strings have the same alignment and the
compare length is at least 8 bytes. rBITDIF contains the low order
2 bits of rSTR1 and cr5 contains the result of the logical compare
- of rBITDIF to 0. If rBITDIF == 0 then we are already word
+ of rBITDIF to 0. If rBITDIF == 0 then we are already word
aligned and can perform the word aligned loop.
-
+
Otherwise we know the two strings have the same alignment (but not
yet word aligned). So we force the string addresses to the next lower
word boundary and special case this first word using shift left to
@@ -143,7 +143,7 @@ L(Waligned):
beq L(dP4)
bgt cr1, L(dP3)
beq cr1, L(dP2)
-
+
/* Remainder is 4 */
.align 4
L(dP1):
@@ -152,7 +152,7 @@ L(dP1):
(8-15 byte compare), we want to use only volatile registers. This
means we can avoid restoring non-volatile registers since we did not
change any on the early exit path. The key here is the non-early
- exit path only cares about the condition code (cr5), not about which
+ exit path only cares about the condition code (cr5), not about which
register pair was used. */
lwz rWORD5, 0(rSTR1)
lwz rWORD6, 0(rSTR2)
@@ -170,7 +170,7 @@ L(dP1e):
cmplw cr6, rWORD5, rWORD6
bne cr5, L(dLcr5)
bne cr0, L(dLcr0)
-
+
lwzu rWORD7, 16(rSTR1)
lwzu rWORD8, 16(rSTR2)
bne cr1, L(dLcr1)
@@ -188,7 +188,7 @@ L(dP1x):
bne L(d00)
li rRTN, 0
blr
-
+
/* Remainder is 8 */
.align 4
L(dP2):
@@ -230,7 +230,7 @@ L(dP2x):
bne L(d00)
li rRTN, 0
blr
-
+
/* Remainder is 12 */
.align 4
L(dP3):
@@ -273,7 +273,7 @@ L(dP3x):
bne L(d00)
li rRTN, 0
blr
-
+
/* Count is a multiple of 16, remainder is 0 */
.align 4
L(dP4):
@@ -316,8 +316,8 @@ L(dLoop3):
lwzu rWORD8, 16(rSTR2)
bne- cr1, L(dLcr1)
cmplw cr0, rWORD1, rWORD2
- bdnz+ L(dLoop)
-
+ bdnz+ L(dLoop)
+
L(dL4):
cmplw cr1, rWORD3, rWORD4
bne cr6, L(dLcr6)
@@ -332,7 +332,7 @@ L(d24):
bne cr6, L(dLcr6)
L(d14):
slwi. r12, rN, 3
- bne cr5, L(dLcr5)
+ bne cr5, L(dLcr5)
L(d04):
lwz r30,44(1)
lwz r31,48(1)
@@ -341,10 +341,10 @@ L(d04):
beq L(zeroLength)
/* At this point we have a remainder of 1 to 3 bytes to compare. Since
we are aligned it is safe to load the whole word, and use
- shift right to eliminate bits beyond the compare length. */
+ shift right to eliminate bits beyond the compare length. */
L(d00):
lwz rWORD1, 4(rSTR1)
- lwz rWORD2, 4(rSTR2)
+ lwz rWORD2, 4(rSTR2)
srw rWORD1, rWORD1, rN
srw rWORD2, rWORD2, rN
cmplw rWORD1,rWORD2
@@ -392,22 +392,22 @@ L(dLcr5x):
bgtlr cr5
li rRTN, -1
blr
-
+
.align 4
L(bytealigned):
cfi_adjust_cfa_offset(-64)
mtctr rN /* Power4 wants mtctr 1st in dispatch group */
/* We need to prime this loop. This loop is swing modulo scheduled
- to avoid pipe delays. The dependent instruction latencies (load to
+ to avoid pipe delays. The dependent instruction latencies (load to
compare to conditional branch) is 2 to 3 cycles. In this loop each
dispatch group ends in a branch and takes 1 cycle. Effectively
- the first iteration of the loop only serves to load operands and
- branches based on compares are delayed until the next loop.
+ the first iteration of the loop only serves to load operands and
+ branches based on compares are delayed until the next loop.
So we must precondition some registers and condition codes so that
we don't exit the loop early on the first iteration. */
-
+
lbz rWORD1, 0(rSTR1)
lbz rWORD2, 0(rSTR2)
bdz- L(b11)
@@ -427,7 +427,7 @@ L(bLoop):
cmplw cr6, rWORD5, rWORD6
bdz- L(b3i)
-
+
lbzu rWORD3, 1(rSTR1)
lbzu rWORD4, 1(rSTR2)
bne- cr1, L(bLcr1)
@@ -441,10 +441,10 @@ L(bLoop):
cmplw cr1, rWORD3, rWORD4
bdnz+ L(bLoop)
-
+
/* We speculatively loading bytes before we have tested the previous
bytes. But we must avoid overrunning the length (in the ctr) to
- prevent these speculative loads from causing a segfault. In this
+ prevent these speculative loads from causing a segfault. In this
case the loop will exit early (before the all pending bytes are
tested. In this case we must complete the pending operations
before returning. */
@@ -488,7 +488,7 @@ L(bx56):
nop
L(b12):
bne- cr0, L(bx12)
-L(bx34):
+L(bx34):
sub rRTN, rWORD3, rWORD4
blr
@@ -497,7 +497,7 @@ L(bx12):
sub rRTN, rWORD1, rWORD2
blr
- .align 4
+ .align 4
L(zeroLengthReturn):
L(zeroLength):
@@ -509,9 +509,9 @@ L(zeroLength):
/* At this point we know the strings have different alignment and the
compare length is at least 8 bytes. rBITDIF contains the low order
2 bits of rSTR1 and cr5 contains the result of the logical compare
- of rBITDIF to 0. If rBITDIF == 0 then rStr1 is word aligned and can
+ of rBITDIF to 0. If rBITDIF == 0 then rStr1 is word aligned and can
perform the Wunaligned loop.
-
+
Otherwise we know that rSTR1 is not aready word aligned yet.
So we can force the string addresses to the next lower word
boundary and special case this first word using shift left to
@@ -531,13 +531,13 @@ L(zeroLength):
#define rE r0 /* Right rotation temp for rWORD6. */
#define rG r12 /* Right rotation temp for rWORD8. */
L(unaligned):
- stw r29,40(r1)
- cfi_offset(r29,(40-64))
+ stw r29,40(r1)
+ cfi_offset(r29,(40-64))
clrlwi rSHL, rSTR2, 30
- stw r28,36(r1)
+ stw r28,36(r1)
cfi_offset(r28,(36-64))
beq cr5, L(Wunaligned)
- stw r27,32(r1)
+ stw r27,32(r1)
cfi_offset(r27,(32-64))
/* Adjust the logical start of rSTR2 to compensate for the extra bits
in the 1st rSTR1 W. */
@@ -545,19 +545,19 @@ L(unaligned):
/* But do not attempt to address the W before that W that contains
the actual start of rSTR2. */
clrrwi rSTR2, rSTR2, 2
- stw r26,28(r1)
+ stw r26,28(r1)
cfi_offset(r26,(28-64))
/* Compute the left/right shift counts for the unalign rSTR2,
- compensating for the logical (W aligned) start of rSTR1. */
+ compensating for the logical (W aligned) start of rSTR1. */
clrlwi rSHL, r27, 30
- clrrwi rSTR1, rSTR1, 2
- stw r25,24(r1)
+ clrrwi rSTR1, rSTR1, 2
+ stw r25,24(r1)
cfi_offset(r25,(24-64))
slwi rSHL, rSHL, 3
cmplw cr5, r27, rSTR2
add rN, rN, rBITDIF
slwi r11, rBITDIF, 3
- stw r24,20(r1)
+ stw r24,20(r1)
cfi_offset(r24,(20-64))
subfic rSHR, rSHL, 32
srwi rTMP, rN, 4 /* Divide by 16 */
@@ -633,16 +633,16 @@ L(duPs4):
compare length is at least 8 bytes. */
.align 4
L(Wunaligned):
- stw r27,32(r1)
+ stw r27,32(r1)
cfi_offset(r27,(32-64))
clrrwi rSTR2, rSTR2, 2
- stw r26,28(r1)
+ stw r26,28(r1)
cfi_offset(r26,(28-64))
srwi rTMP, rN, 4 /* Divide by 16 */
- stw r25,24(r1)
+ stw r25,24(r1)
cfi_offset(r25,(24-64))
andi. rBITDIF, rN, 12 /* Get the W remainder */
- stw r24,20(r1)
+ stw r24,20(r1)
cfi_offset(r24,(20-64))
slwi rSHL, rSHL, 3
lwz rWORD6, 0(rSTR2)
@@ -656,7 +656,7 @@ L(Wunaligned):
mtctr rTMP /* Power4 wants mtctr 1st in dispatch group */
bgt cr1, L(duP3)
beq cr1, L(duP2)
-
+
/* Remainder is 4 */
.align 4
L(duP1):
@@ -687,7 +687,7 @@ L(duP1e):
bne cr0, L(duLcr0)
or rWORD6, rE, rF
cmplw cr6, rWORD5, rWORD6
- b L(duLoop3)
+ b L(duLoop3)
.align 4
/* At this point we exit early with the first word compare
complete and remainder of 0 to 3 bytes. See L(du14) for details on
@@ -751,7 +751,7 @@ L(duP2x):
lwz rWORD2, 4(rSTR2)
srw rA, rWORD2, rSHR
b L(dutrim)
-
+
/* Remainder is 12 */
.align 4
L(duP3):
@@ -801,7 +801,7 @@ L(duP3x):
lwz rWORD2, 4(rSTR2)
srw rA, rWORD2, rSHR
b L(dutrim)
-
+
/* Count is a multiple of 16, remainder is 0 */
.align 4
L(duP4):
@@ -867,8 +867,8 @@ L(duLoop3):
srw rG, rWORD8, rSHR
slw rB, rWORD8, rSHL
or rWORD8, rG, rH
- bdnz+ L(duLoop)
-
+ bdnz+ L(duLoop)
+
L(duL4):
bne cr1, L(duLcr1)
cmplw cr1, rWORD3, rWORD4
@@ -886,9 +886,9 @@ L(du14):
slwi. rN, rN, 3
bne cr5, L(duLcr5)
/* At this point we have a remainder of 1 to 3 bytes to compare. We use
- shift right to eliminate bits beyond the compare length.
+ shift right to eliminate bits beyond the compare length.
- However it may not be safe to load rWORD2 which may be beyond the
+ However it may not be safe to load rWORD2 which may be beyond the
string length. So we compare the bit length of the remainder to
the right shift count (rSHR). If the bit count is less than or equal
we do not need to load rWORD2 (all significant bits are already in
@@ -903,13 +903,13 @@ L(du14):
L(dutrim):
lwz rWORD1, 4(rSTR1)
lwz r31,48(1)
- subfic rN, rN, 32 /* Shift count is 32 - (rN * 8). */
+ subfic rN, rN, 32 /* Shift count is 32 - (rN * 8). */
or rWORD2, rA, rB
lwz r30,44(1)
lwz r29,40(r1)
srw rWORD1, rWORD1, rN
srw rWORD2, rWORD2, rN
- lwz r28,36(r1)
+ lwz r28,36(r1)
lwz r27,32(r1)
cmplw rWORD1,rWORD2
li rRTN,0
@@ -923,9 +923,9 @@ L(duLcr0):
lwz r31,48(1)
lwz r30,44(1)
li rRTN, 1
- bgt cr0, L(dureturn29)
+ bgt cr0, L(dureturn29)
lwz r29,40(r1)
- lwz r28,36(r1)
+ lwz r28,36(r1)
li rRTN, -1
b L(dureturn27)
.align 4
@@ -933,9 +933,9 @@ L(duLcr1):
lwz r31,48(1)
lwz r30,44(1)
li rRTN, 1
- bgt cr1, L(dureturn29)
+ bgt cr1, L(dureturn29)
lwz r29,40(r1)
- lwz r28,36(r1)
+ lwz r28,36(r1)
li rRTN, -1
b L(dureturn27)
.align 4
@@ -943,9 +943,9 @@ L(duLcr6):
lwz r31,48(1)
lwz r30,44(1)
li rRTN, 1
- bgt cr6, L(dureturn29)
+ bgt cr6, L(dureturn29)
lwz r29,40(r1)
- lwz r28,36(r1)
+ lwz r28,36(r1)
li rRTN, -1
b L(dureturn27)
.align 4
@@ -953,9 +953,9 @@ L(duLcr5):
lwz r31,48(1)
lwz r30,44(1)
li rRTN, 1
- bgt cr5, L(dureturn29)
+ bgt cr5, L(dureturn29)
lwz r29,40(r1)
- lwz r28,36(r1)
+ lwz r28,36(r1)
li rRTN, -1
b L(dureturn27)
.align 3
@@ -965,14 +965,14 @@ L(duZeroReturn):
L(dureturn):
lwz r31,48(1)
lwz r30,44(1)
-L(dureturn29):
+L(dureturn29):
lwz r29,40(r1)
- lwz r28,36(r1)
-L(dureturn27):
+ lwz r28,36(r1)
+L(dureturn27):
lwz r27,32(r1)
-L(dureturn26):
+L(dureturn26):
lwz r26,28(r1)
-L(dureturn25):
+L(dureturn25):
lwz r25,24(r1)
lwz r24,20(r1)
lwz 1,0(1)
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/memset.S b/libc/sysdeps/powerpc/powerpc32/power4/memset.S
index 1e8785cb4..c2d288b38 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/memset.S
+++ b/libc/sysdeps/powerpc/powerpc32/power4/memset.S
@@ -162,7 +162,7 @@ L(cacheAligned):
add rMEMP,rMEMP,rCLS
b L(cacheAligned)
-/* We are here because the cache line size was set and the remainder
+/* We are here because the cache line size was set and the remainder
(rLEN) is less than the actual cache line size.
So set up the preconditions for L(nondcbz) and go there. */
L(handletail32):
diff --git a/libc/sysdeps/powerpc/powerpc32/power4/wordcopy.c b/libc/sysdeps/powerpc/powerpc32/power4/wordcopy.c
index 6dd0fa392..5d857f61e 100644
--- a/libc/sysdeps/powerpc/powerpc32/power4/wordcopy.c
+++ b/libc/sysdeps/powerpc/powerpc32/power4/wordcopy.c
@@ -37,7 +37,7 @@ _wordcopy_fwd_aligned (dstp, srcp, len)
if (len & 1)
{
((op_t *) dstp)[0] = ((op_t *) srcp)[0];
-
+
if (len == 1)
return;
srcp += OPSIZ;
@@ -88,10 +88,10 @@ _wordcopy_fwd_dest_aligned (dstp, srcp, len)
{
a1 = ((op_t *) srcp)[1];
((op_t *) dstp)[0] = MERGE (a0, sh_1, a1, sh_2);
-
+
if (len == 1)
return;
-
+
a0 = a1;
srcp += OPSIZ;
dstp += OPSIZ;
@@ -131,7 +131,7 @@ _wordcopy_bwd_aligned (dstp, srcp, len)
srcp -= OPSIZ;
dstp -= OPSIZ;
((op_t *) dstp)[0] = ((op_t *) srcp)[0];
-
+
if (len == 1)
return;
len -= 1;
diff --git a/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_llround.S b/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_llround.S
index 558bd2acb..ecd37c3cd 100644
--- a/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_llround.S
+++ b/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_llround.S
@@ -18,13 +18,13 @@
#include <sysdep.h>
#include <math_ldbl_opt.h>
-
+
/* long [r3] llround (float x [fp1])
- IEEE 1003.1 lround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 lround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we pre-round using the V2.02 Floating Round to Integer Nearest
instruction before we use the Floating Convert to Integer Word with
round to zero instruction. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_lround.S b/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_lround.S
index 0fa359d07..d4da625bb 100644
--- a/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_lround.S
+++ b/libc/sysdeps/powerpc/powerpc32/power5+/fpu/s_lround.S
@@ -17,13 +17,13 @@
<http://www.gnu.org/licenses/>. */
#include <sysdep.h>
#include <math_ldbl_opt.h>
-
+
/* long [r3] lround (float x [fp1])
- IEEE 1003.1 lround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 lround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we pre-round using the V2.02 Floating Round to Integer Nearest
instruction before we use the Floating Convert to Integer Word with
round to zero instruction. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrt.S b/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrt.S
index 23559aa19..ed11d5aec 100644
--- a/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrt.S
+++ b/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrt.S
@@ -25,19 +25,19 @@
sets the appropriate floating point exceptions. Extended checking is
only needed to set errno (via __kernel_standard) if the input value
is negative.
-
+
So compare the input value against the absolute value of itself.
This will compare equal unless the value is negative (EDOM) or a NAN,
in which case we branch to the extend wrapper. If equal we can return
the result directly.
-
+
This part of the function looks like a leaf routine, so no need to
stack a frame or execute prologue/epilogue code. It is safe to
branch directly to w_sqrt as long as the input value (f1) is
preserved. Putting the sqrt result into f2 (float parameter 2)
allows passing both the input value and sqrt result into the extended
wrapper so there is no need to recompute.
-
+
This tactic avoids the overhead of stacking a frame for the normal
(non-error) case. Until gcc supports prologue shrink-wrapping
this is the best we can do. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrtf.S b/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrtf.S
index 590c24caf..204917288 100644
--- a/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrtf.S
+++ b/libc/sysdeps/powerpc/powerpc32/power5/fpu/w_sqrtf.S
@@ -25,19 +25,19 @@
sets the appropriate floating point exceptions. Extended checking is
only needed to set errno (via __kernel_standard) if the input value
is negative.
-
+
So compare the input value against the absolute value of itself.
This will compare equal unless the value is negative (EDOM) or a NAN,
in which case we branch to the extend wrapper. If equal we can return
the result directly.
-
+
This part of the function looks like a leaf routine, so no need to
stack a frame or execute prologue/epilogue code. It is safe to
branch directly to w_sqrt as long as the input value (f1) is
preserved. Putting the sqrt result into f2 (float parameter 2)
allows passing both the input value and sqrt result into the extended
wrapper so there is no need to recompute.
-
+
This tactic avoids the overhead of stacking a frame for the normal
(non-error) case. Until gcc supports prologue shrink-wrapping
this is the best we can do. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrint.S b/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrint.S
index 86f51bb4f..3344b312e 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrint.S
+++ b/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrint.S
@@ -20,7 +20,7 @@
#include <math_ldbl_opt.h>
/* long long int[r3, r4] __llrint (double x[fp1]) */
-ENTRY (__llrint)
+ENTRY (__llrint)
CALL_MCOUNT
stwu r1,-16(r1)
cfi_adjust_cfa_offset (16)
@@ -31,7 +31,7 @@ ENTRY (__llrint)
ori r1,r1,0
lwz r3,8(r1)
lwz r4,12(r1)
- addi r1,r1,16
+ addi r1,r1,16
blr
END (__llrint)
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrintf.S b/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrintf.S
index 1b06e2b9b..7f64f8d12 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrintf.S
+++ b/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llrintf.S
@@ -19,7 +19,7 @@
#include <sysdep.h>
/* long long int[r3, r4] __llrintf (float x[fp1]) */
-ENTRY (__llrintf)
+ENTRY (__llrintf)
CALL_MCOUNT
stwu r1,-16(r1)
cfi_adjust_cfa_offset (16)
@@ -30,7 +30,7 @@ ENTRY (__llrintf)
ori r1,r1,0
lwz r3,8(r1)
lwz r4,12(r1)
- addi r1,r1,16
+ addi r1,r1,16
blr
END (__llrintf)
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llround.S b/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llround.S
index bfc5efbe8..0ff04cb71 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llround.S
+++ b/libc/sysdeps/powerpc/powerpc32/power6/fpu/s_llround.S
@@ -18,13 +18,13 @@
#include <sysdep.h>
#include <math_ldbl_opt.h>
-
+
/* long [r3] llround (float x [fp1])
- IEEE 1003.1 lround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 lround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we pre-round using the V2.02 Floating Round to Integer Nearest
instruction before we use the Floating Convert to Integer Word with
round to zero instruction. */
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/memcpy.S b/libc/sysdeps/powerpc/powerpc32/power6/memcpy.S
index d900028cb..c3d55b768 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/memcpy.S
+++ b/libc/sysdeps/powerpc/powerpc32/power6/memcpy.S
@@ -21,7 +21,7 @@
/* __ptr_t [r3] memcpy (__ptr_t dst [r3], __ptr_t src [r4], size_t len [r5]);
Returns 'dst'.
- Memcpy handles short copies (< 32-bytes) using a binary move blocks
+ Memcpy handles short copies (< 32-bytes) using a binary move blocks
(no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
with the appropriate combination of byte and halfword load/stores.
There is minimal effort to optimize the alignment of short moves.
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/wcschr.c b/libc/sysdeps/powerpc/powerpc32/power6/wcschr.c
index a989877bb..1ba64e113 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/wcschr.c
+++ b/libc/sysdeps/powerpc/powerpc32/power6/wcschr.c
@@ -22,10 +22,10 @@
/* Find the first occurrence of WC in WCS. */
wchar_t *
wcschr (wcs, wc)
- register const wchar_t *wcs;
- register const wchar_t wc;
+ const wchar_t *wcs;
+ const wchar_t wc;
{
- register const wchar_t *wcs2 = wcs + 1;
+ const wchar_t *wcs2 = wcs + 1;
if (*wcs == wc)
return (wchar_t *) wcs;
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/wcsrchr.c b/libc/sysdeps/powerpc/powerpc32/power6/wcsrchr.c
index d0ff9f69c..8d26af05f 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/wcsrchr.c
+++ b/libc/sysdeps/powerpc/powerpc32/power6/wcsrchr.c
@@ -22,10 +22,10 @@
/* Find the last occurrence of WC in WCS. */
wchar_t *
wcsrchr (wcs, wc)
- register const wchar_t *wcs;
- register const wchar_t wc;
+ const wchar_t *wcs;
+ const wchar_t wc;
{
- register const wchar_t *wcs2 = wcs + 1;
+ const wchar_t *wcs2 = wcs + 1;
const wchar_t *retval = NULL;
if (*wcs == wc)
diff --git a/libc/sysdeps/powerpc/powerpc32/power6/wordcopy.c b/libc/sysdeps/powerpc/powerpc32/power6/wordcopy.c
index bcb6176a3..4106e5c6e 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6/wordcopy.c
+++ b/libc/sysdeps/powerpc/powerpc32/power6/wordcopy.c
@@ -38,7 +38,7 @@ _wordcopy_fwd_aligned (dstp, srcp, len)
if (len & 1)
{
((op_t *) dstp)[0] = ((op_t *) srcp)[0];
-
+
if (len == 1)
return;
srcp += OPSIZ;
@@ -105,10 +105,10 @@ _wordcopy_fwd_dest_aligned (dstp, srcp, len)
{
a1 = ((op_t *) srcp)[1];
((op_t *) dstp)[0] = MERGE (a0, sh_1, a1, sh_2);
-
+
if (len == 1)
return;
-
+
a0 = a1;
srcp += OPSIZ;
dstp += OPSIZ;
@@ -137,7 +137,7 @@ _wordcopy_bwd_aligned (dstp, srcp, len)
srcp -= OPSIZ;
dstp -= OPSIZ;
((op_t *) dstp)[0] = ((op_t *) srcp)[0];
-
+
if (len == 1)
return;
len -= 1;
diff --git a/libc/sysdeps/powerpc/powerpc32/power6x/fpu/s_lround.S b/libc/sysdeps/powerpc/powerpc32/power6x/fpu/s_lround.S
index 950b69536..0a04051ff 100644
--- a/libc/sysdeps/powerpc/powerpc32/power6x/fpu/s_lround.S
+++ b/libc/sysdeps/powerpc/powerpc32/power6x/fpu/s_lround.S
@@ -18,13 +18,13 @@
#include <sysdep.h>
#include <math_ldbl_opt.h>
-
+
/* long [r3] lround (float x [fp1])
- IEEE 1003.1 lround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 lround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we pre-round using the V2.02 Floating Round to Integer Nearest
instruction before we use the Floating Convert to Integer Word with
round to zero instruction. */
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_ceill.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_ceill.S
index bffac3962..3ef7b9f17 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_ceill.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_ceill.S
@@ -69,7 +69,7 @@ ENTRY (__ceill)
mtfsf 0x01,fp11 /* restore previous rounding mode. */
fnabs fp1,fp1 /* if (x == 0.0) */
blr /* x = -0.0; */
-
+
/* The high double is > TWO52 so we need to round the low double and
perhaps the high double. In this case we have to round the low
double and handle any adjustment to the high double that may be
@@ -93,7 +93,7 @@ ENTRY (__ceill)
beqlr- cr0
mtfsfi 7,2 /* Set rounding mode toward +inf. */
fdiv fp8,fp1,fp13 /* x_high/TWO52 */
-
+
bng- cr6,.L6 /* if (x > 0.0) */
fctidz fp0,fp8
fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
@@ -110,7 +110,7 @@ ENTRY (__ceill)
b .L9
.L6: /* if (x < 0.0) */
fctidz fp0,fp8
- fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
+ fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
bnl cr5,.L7 /* if (x_low < 0.0) */
fmr fp3,fp1
fmr fp4,fp2
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_floor.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_floor.S
index a8e79e936..44bd83233 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_floor.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_floor.S
@@ -48,7 +48,7 @@ EALIGN (__floor, 4, 0)
fnabs fp1,fp1 /* if (x == 0.0) */
/* x = -0.0; */
.L9:
- mtfsf 0x01,fp11 /* restore previous rounding mode. */
+ mtfsf 0x01,fp11 /* restore previous rounding mode. */
blr
END (__floor)
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_floorf.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_floorf.S
index f65e3b8a7..a0a22e7eb 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_floorf.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_floorf.S
@@ -47,7 +47,7 @@ EALIGN (__floorf, 4, 0)
fnabs fp1,fp1 /* if (x == 0.0) */
/* x = -0.0; */
.L9:
- mtfsf 0x01,fp11 /* restore previous rounding mode. */
+ mtfsf 0x01,fp11 /* restore previous rounding mode. */
blr
END (__floorf)
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_isnan.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_isnan.S
index 3fd62ae51..95eb81eef 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_isnan.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_isnan.S
@@ -21,7 +21,7 @@
/* int __isnan(x) */
.machine power4
-EALIGN (__isnan, 4, 0)
+EALIGN (__isnan, 4, 0)
CALL_MCOUNT 0
mffs fp0
mtfsb0 4*cr6+lt /* reset_fpscr_bit (FPSCR_VE) */
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_llrint.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_llrint.S
index 82b8df41d..7019347b5 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_llrint.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_llrint.S
@@ -20,14 +20,14 @@
#include <math_ldbl_opt.h>
/* long long int[r3] __llrint (double x[fp1]) */
-ENTRY (__llrint)
+ENTRY (__llrint)
CALL_MCOUNT 0
fctid fp13,fp1
stfd fp13,-16(r1)
nop /* Insure the following load is in a different dispatch group */
nop /* to avoid pipe stall on POWER4&5. */
nop
- ld r3,-16(r1)
+ ld r3,-16(r1)
blr
END (__llrint)
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_llrintf.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_llrintf.S
index 3de36c22f..467396722 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_llrintf.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_llrintf.S
@@ -19,14 +19,14 @@
#include <sysdep.h>
/* long long int[r3] __llrintf (float x[fp1]) */
-ENTRY (__llrintf)
+ENTRY (__llrintf)
CALL_MCOUNT 0
fctid fp13,fp1
stfd fp13,-16(r1)
nop /* Insure the following load is in a different dispatch group */
nop /* to avoid pipe stall on POWER4&5. */
nop
- ld r3,-16(r1)
+ ld r3,-16(r1)
blr
END (__llrintf)
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_llround.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_llround.S
index 286aae3d6..54b8341b4 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_llround.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_llround.S
@@ -25,13 +25,13 @@
.LC1: /* 0.5 */
.tc FD_3fe00000_0[TC],0x3fe0000000000000
.section ".text"
-
+
/* long long [r3] llround (double x [fp1])
- IEEE 1003.1 llround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 llround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "round to Nearest" mode. Instead we set
"round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value.
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_llroundf.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_llroundf.S
index b1083a81a..25c61f245 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_llroundf.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_llroundf.S
@@ -24,13 +24,13 @@
.LC1: /* 0.5 */
.tc FD_3fe00000_0[TC],0x3fe0000000000000
.section ".text"
-
+
/* long long [r3] llroundf (float x [fp1])
- IEEE 1003.1 llroundf function. IEEE specifies "roundf to the nearest
+ IEEE 1003.1 llroundf function. IEEE specifies "roundf to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "roundf to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "roundf to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "round to Nearest" mode. Instead we set
"round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value.
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_round.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_round.S
index 022be84a2..3c7437a8f 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_round.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_round.S
@@ -25,13 +25,13 @@
.LC1: /* 0.5 */
.tc FD_3fe00000_0[TC],0x3fe0000000000000
.section ".text"
-
+
/* double [fp1] round (double x [fp1])
- IEEE 1003.1 round function. IEEE specifies "round to the nearest
+ IEEE 1003.1 round function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "Round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "Round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "Round to Nearest" mode. Instead we set
"Round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value. */
@@ -63,7 +63,7 @@ EALIGN (__round, 4, 0)
fnabs fp1,fp1 /* if (x == 0.0) */
/* x = -0.0; */
.L9:
- mtfsf 0x01,fp11 /* restore previous rounding mode. */
+ mtfsf 0x01,fp11 /* restore previous rounding mode. */
blr
END (__round)
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_roundf.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_roundf.S
index dc50d1a77..980a77bde 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_roundf.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_roundf.S
@@ -24,13 +24,13 @@
.LC1: /* 0.5 */
.tc FD_3f000000_0[TC],0x3f00000000000000
.section ".text"
-
+
/* float [fp1] roundf (float x [fp1])
- IEEE 1003.1 round function. IEEE specifies "round to the nearest
+ IEEE 1003.1 round function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "Round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "Round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we can't use the PowerPC "Round to Nearest" mode. Instead we set
"Round toward Zero" mode and round by adding +-0.5 before rounding
to the integer value. */
@@ -62,7 +62,7 @@ EALIGN (__roundf, 4, 0)
fnabs fp1,fp1 /* if (x == 0.0) */
/* x = -0.0; */
.L9:
- mtfsf 0x01,fp11 /* restore previous rounding mode. */
+ mtfsf 0x01,fp11 /* restore previous rounding mode. */
blr
END (__roundf)
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_roundl.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_roundl.S
index 5cc41c03c..547b72150 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_roundl.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_roundl.S
@@ -91,7 +91,7 @@ ENTRY (__roundl)
beqlr- cr0
mtfsfi 7,1 /* Set rounding mode toward 0. */
fdiv fp8,fp1,fp13 /* x_high/TWO52 */
-
+
bng- cr6,.L6 /* if (x > 0.0) */
fctidz fp0,fp8
fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
@@ -109,7 +109,7 @@ ENTRY (__roundl)
b .L9
.L6: /* if (x < 0.0) */
fctidz fp0,fp8
- fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
+ fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
bnl cr5,.L7 /* if (x_low < 0.0) */
fmr fp3,fp1
fmr fp4,fp2
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_trunc.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_trunc.S
index 5b018fb3e..f12387366 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_trunc.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_trunc.S
@@ -23,10 +23,10 @@
.LC0: /* 2**52 */
.tc FD_43300000_0[TC],0x4330000000000000
.section ".text"
-
+
/* double [fp1] trunc (double x [fp1])
IEEE 1003.1 trunc function. IEEE specifies "trunc to the integer
- value, in floating format, nearest to but no larger in magnitude
+ value, in floating format, nearest to but no larger in magnitude
then the argument."
We set "round toward Zero" mode and trunc by adding +-2**52 then
subtracting +-2**52. */
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_truncf.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_truncf.S
index 9f3524024..5ea5f3d04 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_truncf.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_truncf.S
@@ -22,10 +22,10 @@
.LC0: /* 2**23 */
.tc FD_4b000000_0[TC],0x4b00000000000000
.section ".text"
-
+
/* float [fp1] truncf (float x [fp1])
IEEE 1003.1 trunc function. IEEE specifies "trunc to the integer
- value, in floating format, nearest to but no larger in magnitude
+ value, in floating format, nearest to but no larger in magnitude
then the argument."
We set "round toward Zero" mode and trunc by adding +-2**23 then
subtracting +-2**23. */
diff --git a/libc/sysdeps/powerpc/powerpc64/fpu/s_truncl.S b/libc/sysdeps/powerpc/powerpc64/fpu/s_truncl.S
index 03f45270a..06fd7dbe4 100644
--- a/libc/sysdeps/powerpc/powerpc64/fpu/s_truncl.S
+++ b/libc/sysdeps/powerpc/powerpc64/fpu/s_truncl.S
@@ -79,7 +79,7 @@ ENTRY (__truncl)
beqlr- cr0
mtfsfi 7,1 /* Set rounding mode toward 0. */
fdiv fp8,fp1,fp13 /* x_high/TWO52 */
-
+
bng- cr6,.L6 /* if (x > 0.0) */
fctidz fp0,fp8
fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
@@ -98,7 +98,7 @@ ENTRY (__truncl)
.L6: /* if (x < 0.0) */
fctidz fp0,fp8
fcfid fp8,fp0 /* tau = floor(x_high/TWO52); */
- fadd fp8,fp8,fp8 /* tau++; Make tau even */
+ fadd fp8,fp8,fp8 /* tau++; Make tau even */
bnl cr5,.L7 /* if (x_low < 0.0) */
fmr fp3,fp1
fmr fp4,fp2
diff --git a/libc/sysdeps/powerpc/powerpc64/hp-timing.c b/libc/sysdeps/powerpc/powerpc64/hp-timing.c
index f54a5f879..5073adb0e 100644
--- a/libc/sysdeps/powerpc/powerpc64/hp-timing.c
+++ b/libc/sysdeps/powerpc/powerpc64/hp-timing.c
@@ -1,4 +1,4 @@
-/* Support for high precision, low overhead timing functions.
+/* Support for high precision, low overhead timing functions.
powerpc64 version.
Copyright (C) 2005-2013 Free Software Foundation, Inc.
This file is part of the GNU C Library.
diff --git a/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrt.c b/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrt.c
index bd0f9f04f..78bba57a2 100644
--- a/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrt.c
+++ b/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrt.c
@@ -35,10 +35,10 @@ __sqrt (double x) /* wrapper sqrt */
#else
if (__builtin_expect (_LIB_VERSION == _IEEE_, 0))
return z;
-
+
if (__builtin_expect (x != x, 0))
return z;
-
+
if (__builtin_expect (x < 0.0, 0))
return __kernel_standard (x, x, 26); /* sqrt(negative) */
else
diff --git a/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrtf.c b/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrtf.c
index 07c4dc156..12d9f6273 100644
--- a/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrtf.c
+++ b/libc/sysdeps/powerpc/powerpc64/power4/fpu/w_sqrtf.c
@@ -38,10 +38,10 @@ __sqrtf (float x) /* wrapper sqrtf */
if (__builtin_expect (_LIB_VERSION == _IEEE_, 0))
return z;
-
+
if (__builtin_expect (x != x, 0))
return z;
-
+
if (__builtin_expect (x < 0.0, 0))
/* sqrtf(negative) */
return (float) __kernel_standard ((double) x, (double) x, 126);
diff --git a/libc/sysdeps/powerpc/powerpc64/power4/memcmp.S b/libc/sysdeps/powerpc/powerpc64/power4/memcmp.S
index 6378ecb2d..69caedc9f 100644
--- a/libc/sysdeps/powerpc/powerpc64/power4/memcmp.S
+++ b/libc/sysdeps/powerpc/powerpc64/power4/memcmp.S
@@ -51,17 +51,17 @@ EALIGN (memcmp, 4, 0)
/* If less than 8 bytes or not aligned, use the unaligned
byte loop. */
blt cr1, L(bytealigned)
- std rWORD8,-8(r1)
+ std rWORD8,-8(r1)
cfi_offset(rWORD8,-8)
- std rWORD7,-16(r1)
+ std rWORD7,-16(r1)
cfi_offset(rWORD7,-16)
bne L(unaligned)
/* At this point we know both strings have the same alignment and the
compare length is at least 8 bytes. rBITDIF contains the low order
3 bits of rSTR1 and cr5 contains the result of the logical compare
- of rBITDIF to 0. If rBITDIF == 0 then we are already double word
+ of rBITDIF to 0. If rBITDIF == 0 then we are already double word
aligned and can perform the DWaligned loop.
-
+
Otherwise we know the two strings have the same alignment (but not
yet DW). So we can force the string addresses to the next lower DW
boundary and special case this first DW word using shift left to
@@ -141,7 +141,7 @@ L(DWaligned):
beq L(dP4)
bgt cr1, L(dP3)
beq cr1, L(dP2)
-
+
/* Remainder is 8 */
.align 4
L(dP1):
@@ -150,7 +150,7 @@ L(dP1):
(8-15 byte compare), we want to use only volatile registers. This
means we can avoid restoring non-volatile registers since we did not
change any on the early exit path. The key here is the non-early
- exit path only cares about the condition code (cr5), not about which
+ exit path only cares about the condition code (cr5), not about which
register pair was used. */
ld rWORD5, 0(rSTR1)
ld rWORD6, 0(rSTR2)
@@ -168,7 +168,7 @@ L(dP1e):
cmpld cr6, rWORD5, rWORD6
bne cr5, L(dLcr5)
bne cr0, L(dLcr0)
-
+
ldu rWORD7, 32(rSTR1)
ldu rWORD8, 32(rSTR2)
bne cr1, L(dLcr1)
@@ -185,7 +185,7 @@ L(dP1x):
bne L(d00)
li rRTN, 0
blr
-
+
/* Remainder is 16 */
.align 4
L(dP2):
@@ -226,7 +226,7 @@ L(dP2x):
bne L(d00)
li rRTN, 0
blr
-
+
/* Remainder is 24 */
.align 4
L(dP3):
@@ -268,7 +268,7 @@ L(dP3x):
bne L(d00)
li rRTN, 0
blr
-
+
/* Count is a multiple of 32, remainder is 0 */
.align 4
L(dP4):
@@ -311,8 +311,8 @@ L(dLoop3):
ldu rWORD8, 32(rSTR2)
bne- cr1, L(dLcr1)
cmpld cr0, rWORD1, rWORD2
- bdnz+ L(dLoop)
-
+ bdnz+ L(dLoop)
+
L(dL4):
cmpld cr1, rWORD3, rWORD4
bne cr6, L(dLcr6)
@@ -327,7 +327,7 @@ L(d24):
bne cr6, L(dLcr6)
L(d14):
sldi. r12, rN, 3
- bne cr5, L(dLcr5)
+ bne cr5, L(dLcr5)
L(d04):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
@@ -338,7 +338,7 @@ L(d04):
shift right double to eliminate bits beyond the compare length. */
L(d00):
ld rWORD1, 8(rSTR1)
- ld rWORD2, 8(rSTR2)
+ ld rWORD2, 8(rSTR2)
srd rWORD1, rWORD1, rN
srd rWORD2, rWORD2, rN
cmpld cr5, rWORD1, rWORD2
@@ -378,22 +378,22 @@ L(dLcr5x):
bgtlr cr5
li rRTN, -1
blr
-
+
.align 4
L(bytealigned):
mtctr rN /* Power4 wants mtctr 1st in dispatch group */
beq- cr6, L(zeroLength)
/* We need to prime this loop. This loop is swing modulo scheduled
- to avoid pipe delays. The dependent instruction latencies (load to
+ to avoid pipe delays. The dependent instruction latencies (load to
compare to conditional branch) is 2 to 3 cycles. In this loop each
dispatch group ends in a branch and takes 1 cycle. Effectively
- the first iteration of the loop only serves to load operands and
- branches based on compares are delayed until the next loop.
+ the first iteration of the loop only serves to load operands and
+ branches based on compares are delayed until the next loop.
So we must precondition some registers and condition codes so that
we don't exit the loop early on the first iteration. */
-
+
lbz rWORD1, 0(rSTR1)
lbz rWORD2, 0(rSTR2)
bdz- L(b11)
@@ -413,7 +413,7 @@ L(bLoop):
cmpld cr6, rWORD5, rWORD6
bdz- L(b3i)
-
+
lbzu rWORD3, 1(rSTR1)
lbzu rWORD4, 1(rSTR2)
bne- cr1, L(bLcr1)
@@ -427,10 +427,10 @@ L(bLoop):
cmpld cr1, rWORD3, rWORD4
bdnz+ L(bLoop)
-
+
/* We speculatively loading bytes before we have tested the previous
bytes. But we must avoid overrunning the length (in the ctr) to
- prevent these speculative loads from causing a segfault. In this
+ prevent these speculative loads from causing a segfault. In this
case the loop will exit early (before the all pending bytes are
tested. In this case we must complete the pending operations
before returning. */
@@ -474,14 +474,14 @@ L(bx56):
nop
L(b12):
bne- cr0, L(bx12)
-L(bx34):
+L(bx34):
sub rRTN, rWORD3, rWORD4
blr
L(b11):
L(bx12):
sub rRTN, rWORD1, rWORD2
blr
- .align 4
+ .align 4
L(zeroLengthReturn):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
@@ -493,9 +493,9 @@ L(zeroLength):
/* At this point we know the strings have different alignment and the
compare length is at least 8 bytes. rBITDIF contains the low order
3 bits of rSTR1 and cr5 contains the result of the logical compare
- of rBITDIF to 0. If rBITDIF == 0 then rStr1 is double word
+ of rBITDIF to 0. If rBITDIF == 0 then rStr1 is double word
aligned and can perform the DWunaligned loop.
-
+
Otherwise we know that rSTR1 is not already DW aligned yet.
So we can force the string addresses to the next lower DW
boundary and special case this first DW word using shift left to
@@ -515,14 +515,14 @@ L(zeroLength):
#define rE r0 /* Right rotation temp for rWORD6. */
#define rG r12 /* Right rotation temp for rWORD8. */
L(unaligned):
- std r29,-24(r1)
+ std r29,-24(r1)
cfi_offset(r29,-24)
clrldi rSHL, rSTR2, 61
beq- cr6, L(duzeroLength)
- std r28,-32(r1)
+ std r28,-32(r1)
cfi_offset(r28,-32)
beq cr5, L(DWunaligned)
- std r27,-40(r1)
+ std r27,-40(r1)
cfi_offset(r27,-40)
/* Adjust the logical start of rSTR2 ro compensate for the extra bits
in the 1st rSTR1 DW. */
@@ -530,19 +530,19 @@ L(unaligned):
/* But do not attempt to address the DW before that DW that contains
the actual start of rSTR2. */
clrrdi rSTR2, rSTR2, 3
- std r26,-48(r1)
+ std r26,-48(r1)
cfi_offset(r26,-48)
/* Compute the left/right shift counts for the unalign rSTR2,
- compensating for the logical (DW aligned) start of rSTR1. */
+ compensating for the logical (DW aligned) start of rSTR1. */
clrldi rSHL, r27, 61
- clrrdi rSTR1, rSTR1, 3
- std r25,-56(r1)
+ clrrdi rSTR1, rSTR1, 3
+ std r25,-56(r1)
cfi_offset(r25,-56)
sldi rSHL, rSHL, 3
cmpld cr5, r27, rSTR2
add rN, rN, rBITDIF
sldi r11, rBITDIF, 3
- std r24,-64(r1)
+ std r24,-64(r1)
cfi_offset(r24,-64)
subfic rSHR, rSHL, 64
srdi rTMP, rN, 5 /* Divide by 32 */
@@ -618,16 +618,16 @@ L(duPs4):
compare length is at least 8 bytes. */
.align 4
L(DWunaligned):
- std r27,-40(r1)
+ std r27,-40(r1)
cfi_offset(r27,-40)
clrrdi rSTR2, rSTR2, 3
- std r26,-48(r1)
+ std r26,-48(r1)
cfi_offset(r26,-48)
srdi rTMP, rN, 5 /* Divide by 32 */
- std r25,-56(r1)
+ std r25,-56(r1)
cfi_offset(r25,-56)
andi. rBITDIF, rN, 24 /* Get the DW remainder */
- std r24,-64(r1)
+ std r24,-64(r1)
cfi_offset(r24,-64)
sldi rSHL, rSHL, 3
ld rWORD6, 0(rSTR2)
@@ -641,7 +641,7 @@ L(DWunaligned):
mtctr rTMP /* Power4 wants mtctr 1st in dispatch group */
bgt cr1, L(duP3)
beq cr1, L(duP2)
-
+
/* Remainder is 8 */
.align 4
L(duP1):
@@ -672,7 +672,7 @@ L(duP1e):
bne cr0, L(duLcr0)
or rWORD6, rE, rF
cmpld cr6, rWORD5, rWORD6
- b L(duLoop3)
+ b L(duLoop3)
.align 4
/* At this point we exit early with the first double word compare
complete and remainder of 0 to 7 bytes. See L(du14) for details on
@@ -736,7 +736,7 @@ L(duP2x):
ld rWORD2, 8(rSTR2)
srd rA, rWORD2, rSHR
b L(dutrim)
-
+
/* Remainder is 24 */
.align 4
L(duP3):
@@ -786,7 +786,7 @@ L(duP3x):
ld rWORD2, 8(rSTR2)
srd rA, rWORD2, rSHR
b L(dutrim)
-
+
/* Count is a multiple of 32, remainder is 0 */
.align 4
L(duP4):
@@ -852,8 +852,8 @@ L(duLoop3):
srd rG, rWORD8, rSHR
sld rB, rWORD8, rSHL
or rWORD8, rG, rH
- bdnz+ L(duLoop)
-
+ bdnz+ L(duLoop)
+
L(duL4):
bne cr1, L(duLcr1)
cmpld cr1, rWORD3, rWORD4
@@ -875,7 +875,7 @@ L(du14):
This allows the use of double word subtract to compute the final
result.
- However it may not be safe to load rWORD2 which may be beyond the
+ However it may not be safe to load rWORD2 which may be beyond the
string length. So we compare the bit length of the remainder to
the right shift count (rSHR). If the bit count is less than or equal
we do not need to load rWORD2 (all significant bits are already in
@@ -890,16 +890,16 @@ L(du14):
L(dutrim):
ld rWORD1, 8(rSTR1)
ld rWORD8,-8(r1)
- subfic rN, rN, 64 /* Shift count is 64 - (rN * 8). */
+ subfic rN, rN, 64 /* Shift count is 64 - (rN * 8). */
or rWORD2, rA, rB
- ld rWORD7,-16(r1)
+ ld rWORD7,-16(r1)
ld r29,-24(r1)
srd rWORD1, rWORD1, rN
srd rWORD2, rWORD2, rN
- ld r28,-32(r1)
+ ld r28,-32(r1)
ld r27,-40(r1)
li rRTN, 0
- cmpld cr0, rWORD1, rWORD2
+ cmpld cr0, rWORD1, rWORD2
ld r26,-48(r1)
ld r25,-56(r1)
beq cr0, L(dureturn24)
@@ -913,7 +913,7 @@ L(duLcr0):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
li rRTN, 1
- bgt cr0, L(dureturn29)
+ bgt cr0, L(dureturn29)
ld r29,-24(r1)
ld r28,-32(r1)
li rRTN, -1
@@ -923,7 +923,7 @@ L(duLcr1):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
li rRTN, 1
- bgt cr1, L(dureturn29)
+ bgt cr1, L(dureturn29)
ld r29,-24(r1)
ld r28,-32(r1)
li rRTN, -1
@@ -933,7 +933,7 @@ L(duLcr6):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
li rRTN, 1
- bgt cr6, L(dureturn29)
+ bgt cr6, L(dureturn29)
ld r29,-24(r1)
ld r28,-32(r1)
li rRTN, -1
@@ -943,7 +943,7 @@ L(duLcr5):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
li rRTN, 1
- bgt cr5, L(dureturn29)
+ bgt cr5, L(dureturn29)
ld r29,-24(r1)
ld r28,-32(r1)
li rRTN, -1
@@ -955,14 +955,14 @@ L(duZeroReturn):
L(dureturn):
ld rWORD8,-8(r1)
ld rWORD7,-16(r1)
-L(dureturn29):
+L(dureturn29):
ld r29,-24(r1)
ld r28,-32(r1)
-L(dureturn27):
+L(dureturn27):
ld r27,-40(r1)
-L(dureturn26):
+L(dureturn26):
ld r26,-48(r1)
-L(dureturn25):
+L(dureturn25):
ld r25,-56(r1)
L(dureturn24):
ld r24,-64(r1)
diff --git a/libc/sysdeps/powerpc/powerpc64/power4/memcpy.S b/libc/sysdeps/powerpc/powerpc64/power4/memcpy.S
index c43d1d2e4..4317c7e78 100644
--- a/libc/sysdeps/powerpc/powerpc64/power4/memcpy.S
+++ b/libc/sysdeps/powerpc/powerpc64/power4/memcpy.S
@@ -21,10 +21,10 @@
/* __ptr_t [r3] memcpy (__ptr_t dst [r3], __ptr_t src [r4], size_t len [r5]);
Returns 'dst'.
- Memcpy handles short copies (< 32-bytes) using a binary move blocks
- (no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
- with the appropriate combination of byte and halfword load/stores.
- There is minimal effort to optimize the alignment of short moves.
+ Memcpy handles short copies (< 32-bytes) using a binary move blocks
+ (no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
+ with the appropriate combination of byte and halfword load/stores.
+ There is minimal effort to optimize the alignment of short moves.
The 64-bit implementations of POWER3 and POWER4 do a reasonable job
of handling unaligned load/stores that do not cross 32-byte boundaries.
@@ -47,13 +47,13 @@ EALIGN (memcpy, 5, 0)
clrldi 10,4,61 /* check alignment of src. */
cmpldi cr6,5,8
ble- cr1,.L2 /* If move < 32 bytes use short move code. */
- cmpld cr6,10,11
+ cmpld cr6,10,11
mr 12,4
srdi 9,5,3 /* Number of full double words remaining. */
mtcrf 0x01,0
mr 31,5
beq .L0
-
+
subf 31,0,5
/* Move 0-7 bytes as needed to get the destination doubleword aligned. */
1: bf 31,2f
@@ -74,15 +74,15 @@ EALIGN (memcpy, 5, 0)
0:
clrldi 10,12,61 /* check alignment of src again. */
srdi 9,31,3 /* Number of full double words remaining. */
-
+
/* Copy doublewords from source to destination, assuming the
destination is aligned on a doubleword boundary.
At this point we know there are at least 25 bytes left (32-7) to copy.
- The next step is to determine if the source is also doubleword aligned.
+ The next step is to determine if the source is also doubleword aligned.
If not branch to the unaligned move code at .L6. which uses
a load, shift, store strategy.
-
+
Otherwise source and destination are doubleword aligned, and we can
the optimized doubleword copy loop. */
.L0:
@@ -95,14 +95,14 @@ EALIGN (memcpy, 5, 0)
Use a unrolled loop to copy 4 doubleword (32-bytes) per iteration.
If the copy is not an exact multiple of 32 bytes, 1-3
doublewords are copied as needed to set up the main loop. After
- the main loop exits there may be a tail of 1-7 bytes. These byte are
+ the main loop exits there may be a tail of 1-7 bytes. These byte are
copied a word/halfword/byte at a time as needed to preserve alignment. */
srdi 8,31,5
cmpldi cr1,9,4
cmpldi cr6,11,0
mr 11,12
-
+
bf 30,1f
ld 6,0(12)
ld 7,8(12)
@@ -113,7 +113,7 @@ EALIGN (memcpy, 5, 0)
addi 10,3,16
bf 31,4f
ld 0,16(12)
- std 0,16(3)
+ std 0,16(3)
blt cr1,3f
addi 11,12,24
addi 10,3,24
@@ -127,7 +127,7 @@ EALIGN (memcpy, 5, 0)
addi 11,12,8
std 6,0(3)
addi 10,3,8
-
+
.align 4
4:
ld 6,0(11)
@@ -142,7 +142,7 @@ EALIGN (memcpy, 5, 0)
std 0,24(10)
addi 10,10,32
bdnz 4b
-3:
+3:
rldicr 0,31,0,60
mtcrf 0x01,31
@@ -150,7 +150,7 @@ EALIGN (memcpy, 5, 0)
.L9:
add 3,3,0
add 12,12,0
-
+
/* At this point we have a tail of 0-7 bytes and we know that the
destination is double word aligned. */
4: bf 29,2f
@@ -171,29 +171,29 @@ EALIGN (memcpy, 5, 0)
ld 31,-8(1)
ld 3,-16(1)
blr
-
-/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
- bytes. Each case is handled without loops, using binary (1,2,4,8)
- tests.
-
+
+/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
+ bytes. Each case is handled without loops, using binary (1,2,4,8)
+ tests.
+
In the short (0-8 byte) case no attempt is made to force alignment
- of either source or destination. The hardware will handle the
- unaligned load/stores with small delays for crossing 32- 64-byte, and
+ of either source or destination. The hardware will handle the
+ unaligned load/stores with small delays for crossing 32- 64-byte, and
4096-byte boundaries. Since these short moves are unlikely to be
- unaligned or cross these boundaries, the overhead to force
+ unaligned or cross these boundaries, the overhead to force
alignment is not justified.
-
+
The longer (9-31 byte) move is more likely to cross 32- or 64-byte
boundaries. Since only loads are sensitive to the 32-/64-byte
- boundaries it is more important to align the source then the
+ boundaries it is more important to align the source then the
destination. If the source is not already word aligned, we first
- move 1-3 bytes as needed. Since we are only word aligned we don't
- use double word load/stores to insure that all loads are aligned.
+ move 1-3 bytes as needed. Since we are only word aligned we don't
+ use double word load/stores to insure that all loads are aligned.
While the destination and stores may still be unaligned, this
is only an issue for page (4096 byte boundary) crossing, which
should be rare for these short moves. The hardware handles this
- case automatically with a small delay. */
-
+ case automatically with a small delay. */
+
.align 4
.L2:
mtcrf 0x01,5
@@ -256,11 +256,11 @@ EALIGN (memcpy, 5, 0)
lwz 6,0(12)
addi 12,12,4
stw 6,0(3)
- addi 3,3,4
+ addi 3,3,4
2: /* Move 2-3 bytes. */
bf 30,1f
lhz 6,0(12)
- sth 6,0(3)
+ sth 6,0(3)
bf 31,0f
lbz 7,2(12)
stb 7,2(3)
@@ -281,7 +281,7 @@ EALIGN (memcpy, 5, 0)
mr 12,4
bne cr6,4f
/* Would have liked to use use ld/std here but the 630 processors are
- slow for load/store doubles that are not at least word aligned.
+ slow for load/store doubles that are not at least word aligned.
Unaligned Load/Store word execute with only a 1 cycle penalty. */
lwz 6,0(4)
lwz 7,4(4)
@@ -297,14 +297,14 @@ EALIGN (memcpy, 5, 0)
6:
bf 30,5f
lhz 7,4(4)
- sth 7,4(3)
+ sth 7,4(3)
bf 31,0f
lbz 8,6(4)
stb 8,6(3)
ld 3,-16(1)
blr
.align 4
-5:
+5:
bf 31,0f
lbz 6,4(4)
stb 6,4(3)
@@ -401,7 +401,7 @@ EALIGN (memcpy, 5, 0)
/* calculate and store the final DW */
sld 0,6,10
srd 8,7,9
- or 0,0,8
+ or 0,0,8
std 0,0(4)
3:
rldicr 0,31,0,60
diff --git a/libc/sysdeps/powerpc/powerpc64/power5+/fpu/s_llround.S b/libc/sysdeps/powerpc/powerpc64/power5+/fpu/s_llround.S
index 108910e77..28df00642 100644
--- a/libc/sysdeps/powerpc/powerpc64/power5+/fpu/s_llround.S
+++ b/libc/sysdeps/powerpc/powerpc64/power5+/fpu/s_llround.S
@@ -18,13 +18,13 @@
#include <sysdep.h>
#include <math_ldbl_opt.h>
-
+
/* long long [r3] llround (float x [fp1])
- IEEE 1003.1 llround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 llround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we pre-round using the V2.02 Floating Round to Integer Nearest
instruction before we use Floating Convert to Integer Word with
round to zero instruction. */
@@ -32,7 +32,7 @@
.machine "power5"
EALIGN (__llround, 4, 0)
CALL_MCOUNT 0
- frin fp2, fp1 /* Round to nearest +-0.5. */
+ frin fp2, fp1 /* Round to nearest +-0.5. */
fctidz fp3, fp2 /* Convert To Integer DW round toward 0. */
stfd fp3, -16(r1)
nop /* Insure the following load is in a different dispatch group */
diff --git a/libc/sysdeps/powerpc/powerpc64/power5/fpu/s_isnan.S b/libc/sysdeps/powerpc/powerpc64/power5/fpu/s_isnan.S
index 3afec2b6b..8319d6e17 100644
--- a/libc/sysdeps/powerpc/powerpc64/power5/fpu/s_isnan.S
+++ b/libc/sysdeps/powerpc/powerpc64/power5/fpu/s_isnan.S
@@ -21,13 +21,13 @@
/* int __isnan(x) */
.machine power5
-EALIGN (__isnan, 4, 0)
+EALIGN (__isnan, 4, 0)
CALL_MCOUNT 0
stfd fp1,-8(r1) /* copy FPR to GPR */
lis r0,0x7ff0
nop /* insure the following is in a different */
nop /* dispatch group */
- ld r4,-8(r1)
+ ld r4,-8(r1)
sldi r0,r0,32 /* const long r0 0x7ff00000 00000000 */
clrldi r4,r4,1 /* x = fabs(x) */
cmpd cr7,r4,r0 /* if (fabs(x) <= inf) */
diff --git a/libc/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S b/libc/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
index 8f0c80662..d4515d69d 100644
--- a/libc/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
+++ b/libc/sysdeps/powerpc/powerpc64/power6/fpu/s_isnan.S
@@ -21,7 +21,7 @@
/* int __isnan(x) */
.machine power6
-EALIGN (__isnan, 4, 0)
+EALIGN (__isnan, 4, 0)
CALL_MCOUNT 0
stfd fp1,-8(r1) /* copy FPR to GPR */
ori r1,r1,0
diff --git a/libc/sysdeps/powerpc/powerpc64/power6/memcpy.S b/libc/sysdeps/powerpc/powerpc64/power6/memcpy.S
index 55c0d7118..db29e2b06 100644
--- a/libc/sysdeps/powerpc/powerpc64/power6/memcpy.S
+++ b/libc/sysdeps/powerpc/powerpc64/power6/memcpy.S
@@ -21,22 +21,22 @@
/* __ptr_t [r3] memcpy (__ptr_t dst [r3], __ptr_t src [r4], size_t len [r5]);
Returns 'dst'.
- Memcpy handles short copies (< 32-bytes) using a binary move blocks
- (no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
- with the appropriate combination of byte and halfword load/stores.
- There is minimal effort to optimize the alignment of short moves.
+ Memcpy handles short copies (< 32-bytes) using a binary move blocks
+ (no loops) of lwz/stw. The tail (remaining 1-3) bytes is handled
+ with the appropriate combination of byte and halfword load/stores.
+ There is minimal effort to optimize the alignment of short moves.
The 64-bit implementations of POWER3 and POWER4 do a reasonable job
of handling unaligned load/stores that do not cross 32-byte boundaries.
Longer moves (>= 32-bytes) justify the effort to get at least the
destination doubleword (8-byte) aligned. Further optimization is
possible when both source and destination are doubleword aligned.
- Each case has a optimized unrolled loop.
-
+ Each case has a optimized unrolled loop.
+
For POWER6 unaligned loads will take a 20+ cycle hiccup for any
L1 cache miss that crosses a 32- or 128-byte boundary. Store
is more forgiving and does not take a hiccup until page or
- segment boundaries. So we require doubleword alignment for
+ segment boundaries. So we require doubleword alignment for
the source but may take a risk and only require word alignment
for the destination. */
@@ -54,10 +54,10 @@ EALIGN (memcpy, 7, 0)
cmpldi cr6,5,8
ble- cr1,.L2 /* If move < 32 bytes use short move code. */
mtcrf 0x01,0
- cmpld cr6,10,11
+ cmpld cr6,10,11
srdi 9,5,3 /* Number of full double words remaining. */
beq .L0
-
+
subf 5,0,5
/* Move 0-7 bytes as needed to get the destination doubleword aligned.
Duplicate some code to maximize fall-through and minimize agen delays. */
@@ -76,7 +76,7 @@ EALIGN (memcpy, 7, 0)
lwz 6,1(4)
stw 6,1(3)
b 0f
-
+
2: bf 30,4f
lhz 6,0(4)
sth 6,0(3)
@@ -84,26 +84,26 @@ EALIGN (memcpy, 7, 0)
lwz 6,2(4)
stw 6,2(3)
b 0f
-
+
4: bf 29,0f
lwz 6,0(4)
stw 6,0(3)
-0:
+0:
/* Add the number of bytes until the 1st doubleword of dst to src and dst. */
add 4,4,0
add 3,3,0
-
+
clrldi 10,4,61 /* check alignment of src again. */
srdi 9,5,3 /* Number of full double words remaining. */
-
+
/* Copy doublewords from source to destination, assuming the
destination is aligned on a doubleword boundary.
At this point we know there are at least 25 bytes left (32-7) to copy.
- The next step is to determine if the source is also doubleword aligned.
+ The next step is to determine if the source is also doubleword aligned.
If not branch to the unaligned move code at .L6. which uses
a load, shift, store strategy.
-
+
Otherwise source and destination are doubleword aligned, and we can
the optimized doubleword copy loop. */
.align 4
@@ -121,12 +121,12 @@ EALIGN (memcpy, 7, 0)
the main loop exits there may be a tail of 1-7 bytes. These byte
are copied a word/halfword/byte at a time as needed to preserve
alignment.
-
+
For POWER6 the L1 is store-through and the L2 is store-in. The
L2 is clocked at half CPU clock so we can store 16 bytes every
other cycle. POWER6 also has a load/store bypass so we can do
- load, load, store, store every 2 cycles.
-
+ load, load, store, store every 2 cycles.
+
The following code is sensitive to cache line alignment. Do not
make any change with out first making sure they don't result in
splitting ld/std pairs across a cache line. */
@@ -271,7 +271,7 @@ L(das_loop):
std 8,16+96(10)
std 0,24+96(10)
ble cr5,L(das_loop_e)
-
+
mtctr 12
.align 4
L(das_loop2):
@@ -324,7 +324,7 @@ L(das_loop_e):
.align 4
L(das_tail):
beq cr1,0f
-
+
L(das_tail2):
/* At this point we have a tail of 0-7 bytes and we know that the
destination is double word aligned. */
@@ -342,7 +342,7 @@ L(das_tail2):
lbz 6,4(4)
stb 6,4(3)
b 0f
-
+
2: bf 30,1f
lhz 6,0(4)
sth 6,0(3)
@@ -350,7 +350,7 @@ L(das_tail2):
lbz 6,2(4)
stb 6,2(3)
b 0f
-
+
1: bf 31,0f
lbz 6,0(4)
stb 6,0(3)
@@ -359,7 +359,7 @@ L(das_tail2):
ld 3,-16(1)
blr
-/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
+/* Copy up to 31 bytes. This divided into two cases 0-8 bytes and 9-31
bytes. Each case is handled without loops, using binary (1,2,4,8)
tests.
@@ -419,7 +419,7 @@ L(dus_tail):
/* At least 6 bytes left and the source is word aligned. This allows
some speculative loads up front. */
/* We need to special case the fall-through because the biggest delays
- are due to address computation not being ready in time for the
+ are due to address computation not being ready in time for the
AGEN. */
lwz 6,0(12)
lwz 7,4(12)
@@ -515,7 +515,7 @@ L(dus_tail4): /* Move 4 bytes. */
L(dus_tail2): /* Move 2-3 bytes. */
bf 30,L(dus_tail1)
lhz 6,0(12)
- sth 6,0(3)
+ sth 6,0(3)
bf 31,L(dus_tailX)
lbz 7,2(12)
stb 7,2(3)
@@ -550,7 +550,7 @@ L(dus_4):
stw 6,0(3)
bf 30,L(dus_5)
lhz 7,4(4)
- sth 7,4(3)
+ sth 7,4(3)
bf 31,L(dus_0)
lbz 8,6(4)
stb 8,6(3)
@@ -588,8 +588,8 @@ L(dus_0):
bge cr0, L(du4_do)
blt cr5, L(du1_do)
beq cr5, L(du2_do)
- b L(du3_do)
-
+ b L(du3_do)
+
.align 4
L(du1_do):
bf 30,L(du1_1dw)
@@ -663,7 +663,7 @@ L(du1_fini):
/* calculate and store the final DW */
sldi 0,6, 8
srdi 8,7, 64-8
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
@@ -740,7 +740,7 @@ L(du2_fini):
/* calculate and store the final DW */
sldi 0,6, 16
srdi 8,7, 64-16
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
@@ -817,7 +817,7 @@ L(du3_fini):
/* calculate and store the final DW */
sldi 0,6, 24
srdi 8,7, 64-24
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
@@ -900,7 +900,7 @@ L(du4_fini):
/* calculate and store the final DW */
sldi 0,6, 32
srdi 8,7, 64-32
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
@@ -977,7 +977,7 @@ L(du5_fini):
/* calculate and store the final DW */
sldi 0,6, 40
srdi 8,7, 64-40
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
@@ -1054,7 +1054,7 @@ L(du6_fini):
/* calculate and store the final DW */
sldi 0,6, 48
srdi 8,7, 64-48
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
@@ -1131,10 +1131,10 @@ L(du7_fini):
/* calculate and store the final DW */
sldi 0,6, 56
srdi 8,7, 64-56
- or 0,0,8
+ or 0,0,8
std 0,0(4)
b L(du_done)
-
+
.align 4
L(du_done):
rldicr 0,31,0,60
@@ -1142,7 +1142,7 @@ L(du_done):
beq cr1,0f /* If the tail is 0 bytes we are done! */
add 3,3,0
- add 12,12,0
+ add 12,12,0
/* At this point we have a tail of 0-7 bytes and we know that the
destination is double word aligned. */
4: bf 29,2f
diff --git a/libc/sysdeps/powerpc/powerpc64/power6/wordcopy.c b/libc/sysdeps/powerpc/powerpc64/power6/wordcopy.c
index f19829c2b..751789339 100644
--- a/libc/sysdeps/powerpc/powerpc64/power6/wordcopy.c
+++ b/libc/sysdeps/powerpc/powerpc64/power6/wordcopy.c
@@ -37,7 +37,7 @@ _wordcopy_fwd_aligned (dstp, srcp, len)
if (len & 1)
{
((op_t *) dstp)[0] = ((op_t *) srcp)[0];
-
+
if (len == 1)
return;
srcp += OPSIZ;
@@ -105,10 +105,10 @@ _wordcopy_fwd_dest_aligned (dstp, srcp, len)
{
a1 = ((op_t *) srcp)[1];
((op_t *) dstp)[0] = MERGE (a0, sh_1, a1, sh_2);
-
+
if (len == 1)
return;
-
+
a0 = a1;
srcp += OPSIZ;
dstp += OPSIZ;
@@ -137,7 +137,7 @@ _wordcopy_bwd_aligned (dstp, srcp, len)
srcp -= OPSIZ;
dstp -= OPSIZ;
((op_t *) dstp)[0] = ((op_t *) srcp)[0];
-
+
if (len == 1)
return;
len -= 1;
diff --git a/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_isnan.S b/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_isnan.S
index c09eb65c0..d29fe9e3c 100644
--- a/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_isnan.S
+++ b/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_isnan.S
@@ -21,7 +21,7 @@
/* int __isnan(x) */
.machine power6
-EALIGN (__isnan, 4, 0)
+EALIGN (__isnan, 4, 0)
CALL_MCOUNT 0
mftgpr r4,fp1 /* copy FPR to GPR */
lis r0,0x7ff0
diff --git a/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llrint.S b/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llrint.S
index bb1627fae..67d51ada6 100644
--- a/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llrint.S
+++ b/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llrint.S
@@ -21,7 +21,7 @@
.machine "power6"
/* long long int[r3] __llrint (double x[fp1]) */
-ENTRY (__llrint)
+ENTRY (__llrint)
CALL_MCOUNT 0
fctid fp13,fp1
mftgpr r3,fp13
diff --git a/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llround.S b/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llround.S
index 902db67a3..45aaceaa3 100644
--- a/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llround.S
+++ b/libc/sysdeps/powerpc/powerpc64/power6x/fpu/s_llround.S
@@ -18,13 +18,13 @@
#include <sysdep.h>
#include <math_ldbl_opt.h>
-
+
/* long long [r3] llround (float x [fp1])
- IEEE 1003.1 llround function. IEEE specifies "round to the nearest
+ IEEE 1003.1 llround function. IEEE specifies "round to the nearest
integer value, rounding halfway cases away from zero, regardless of
the current rounding mode." However PowerPC Architecture defines
- "round to Nearest" as "Choose the best approximation. In case of a
- tie, choose the one that is even (least significant bit o).".
+ "round to Nearest" as "Choose the best approximation. In case of a
+ tie, choose the one that is even (least significant bit o).".
So we pre-round using the V2.02 Floating Round to Integer Nearest
instruction before we use Floating Convert to Integer Word with
round to zero instruction. */
@@ -32,7 +32,7 @@
.machine "power6"
ENTRY (__llround)
CALL_MCOUNT 0
- frin fp2,fp1 /* Round to nearest +-0.5. */
+ frin fp2,fp1 /* Round to nearest +-0.5. */
fctidz fp3,fp2 /* Convert To Integer DW round toward 0. */
mftgpr r3,fp3 /* Transfer integer to R3. */
blr
diff --git a/libc/sysdeps/powerpc/powerpc64/strlen.S b/libc/sysdeps/powerpc/powerpc64/strlen.S
index dafd03387..0f9b5eea9 100644
--- a/libc/sysdeps/powerpc/powerpc64/strlen.S
+++ b/libc/sysdeps/powerpc/powerpc64/strlen.S
@@ -57,11 +57,11 @@
2) How popular are bytes with the high bit set? If they are very rare,
on some processors it might be useful to use the simpler expression
~((x - 0x01010101) | 0x7f7f7f7f) (that is, on processors with only one
- ALU), but this fails when any character has its high bit set.
-
+ ALU), but this fails when any character has its high bit set.
+
Answer:
- 1) Added a Data Cache Block Touch early to prefetch the first 128
- byte cache line. Adding dcbt instructions to the loop would not be
+ 1) Added a Data Cache Block Touch early to prefetch the first 128
+ byte cache line. Adding dcbt instructions to the loop would not be
effective since most strings will be shorter than the cache line.*/
/* Some notes on register usage: Under the SVR4 ABI, we can use registers
@@ -101,7 +101,7 @@ ENTRY (strlen)
li rMASK, -1
insrdi r7F7F, r7F7F, 32, 0
/* That's the setup done, now do the first pair of doublewords.
- We make an exception and use method (2) on the first two doublewords,
+ We make an exception and use method (2) on the first two doublewords,
to reduce overhead. */
srd rMASK, rMASK, rPADN
and rTMP1, r7F7F, rWORD1
diff --git a/libc/sysdeps/powerpc/powerpc64/strncmp.S b/libc/sysdeps/powerpc/powerpc64/strncmp.S
index e2726883f..779d9f7f6 100644
--- a/libc/sysdeps/powerpc/powerpc64/strncmp.S
+++ b/libc/sysdeps/powerpc/powerpc64/strncmp.S
@@ -51,7 +51,7 @@ EALIGN (strncmp, 4, 0)
clrldi rN, rN, 61
addi rFEFE, rFEFE, -0x101
addi r7F7F, r7F7F, 0x7f7f
- cmpldi cr1, rN, 0
+ cmpldi cr1, rN, 0
beq L(unaligned)
mtctr rTMP /* Power4 wants mtctr 1st in dispatch group. */
@@ -62,7 +62,7 @@ EALIGN (strncmp, 4, 0)
add rFEFE, rFEFE, rTMP
b L(g1)
-L(g0):
+L(g0):
ldu rWORD1, 8(rSTR1)
bne- cr1, L(different)
ldu rWORD2, 8(rSTR2)
@@ -72,11 +72,11 @@ L(g1): add rTMP, rFEFE, rWORD1
and. rTMP, rTMP, rNEG
cmpd cr1, rWORD1, rWORD2
beq+ L(g0)
-
+
/* OK. We've hit the end of the string. We need to be careful that
we don't compare two strings as different because of gunk beyond
the end of the strings... */
-
+
L(endstring):
and rTMP, r7F7F, rWORD1
beq cr1, L(equal)
@@ -152,7 +152,7 @@ L(u1):
lbzu rWORD1, 1(rSTR1)
bne+ cr1, L(u0)
-L(u2): lbzu rWORD1, -1(rSTR1)
+L(u2): lbzu rWORD1, -1(rSTR1)
L(u3): sub rRTN, rWORD1, rWORD2
blr
END (strncmp)
diff --git a/libc/sysdeps/powerpc/sysdep.h b/libc/sysdeps/powerpc/sysdep.h
index a2ed88f33..1b5334ad3 100644
--- a/libc/sysdeps/powerpc/sysdep.h
+++ b/libc/sysdeps/powerpc/sysdep.h
@@ -15,8 +15,8 @@
License along with the GNU C Library; if not, see
<http://www.gnu.org/licenses/>. */
-/*
- * Powerpc Feature masks for the Aux Vector Hardware Capabilities (AT_HWCAP).
+/*
+ * Powerpc Feature masks for the Aux Vector Hardware Capabilities (AT_HWCAP).
* This entry is copied to _dl_hwcap or rtld_global._dl_hwcap during startup.
*/
#define _SYS_AUXV_H 1