summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authormeissner <meissner@138bc75d-0d04-0410-961f-82ee72b054a4>2007-09-19 21:41:08 +0000
committermeissner <meissner@138bc75d-0d04-0410-961f-82ee72b054a4>2007-09-19 21:41:08 +0000
commit009b318f3d6363ef0a41b8e33c63d3e723c55fb3 (patch)
tree388fd26895e48eecd5226317111e04f710d0ed07
parent66ce45bdc273da0235b2532adc784c4b570ca667 (diff)
downloadgcc-009b318f3d6363ef0a41b8e33c63d3e723c55fb3.tar.gz
Eliminate trailing whitespace
git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@128605 138bc75d-0d04-0410-961f-82ee72b054a4
-rw-r--r--gcc/config/i386/athlon.md14
-rw-r--r--gcc/config/i386/bmmintrin.h246
-rw-r--r--gcc/config/i386/constraints.md4
-rw-r--r--gcc/config/i386/i386.c18
-rw-r--r--gcc/config/i386/i386.h10
-rw-r--r--gcc/config/i386/i386.md62
-rw-r--r--gcc/config/i386/mmx.md6
-rw-r--r--gcc/config/i386/ppro.md2
-rw-r--r--gcc/config/i386/sse.md36
-rw-r--r--gcc/config/i386/sync.md2
10 files changed, 200 insertions, 200 deletions
diff --git a/gcc/config/i386/athlon.md b/gcc/config/i386/athlon.md
index 04b1e1e2d34..c9860def7c6 100644
--- a/gcc/config/i386/athlon.md
+++ b/gcc/config/i386/athlon.md
@@ -1,7 +1,7 @@
;; AMD Athlon Scheduling
;;
;; The Athlon does contain three pipelined FP units, three integer units and
-;; three address generation units.
+;; three address generation units.
;;
;; The predecode logic is determining boundaries of instructions in the 64
;; byte cache line. So the cache line straddling problem of K6 might be issue
@@ -206,7 +206,7 @@
(and (eq_attr "type" "imul")
(and (eq_attr "mode" "HI")
(eq_attr "memory" "none,unknown"))))
- "athlon-vector,athlon-ieu0,athlon-mult,nothing,athlon-ieu0")
+ "athlon-vector,athlon-ieu0,athlon-mult,nothing,athlon-ieu0")
(define_insn_reservation "athlon_imul_mem" 8
(and (eq_attr "cpu" "athlon")
(and (eq_attr "type" "imul")
@@ -328,7 +328,7 @@
(eq_attr "memory" "both"))))
"athlon-direct,athlon-load,
athlon-ieu,athlon-store,
- athlon-store")
+ athlon-store")
(define_insn_reservation "athlon_ivector_both" 6
(and (eq_attr "cpu" "athlon,k8,generic64")
@@ -617,7 +617,7 @@
(and (eq_attr "cpu" "amdfam10")
(and (eq_attr "type" "mmxmov")
(eq_attr "memory" "load")))
- "athlon-direct,athlon-fploadk8, athlon-fany")
+ "athlon-direct,athlon-fploadk8, athlon-fany")
(define_insn_reservation "athlon_mmxssest" 3
(and (eq_attr "cpu" "k8,generic64")
(and (eq_attr "type" "mmxmov,ssemov")
@@ -926,7 +926,7 @@
(and (eq_attr "amdfam10_decode" "double")
(and (eq_attr "mode" "SF,DF")
(eq_attr "memory" "load")))))
- "athlon-double,athlon-fploadk8,(athlon-faddmul+athlon-fstore)")
+ "athlon-double,athlon-fploadk8,(athlon-faddmul+athlon-fstore)")
;; cvtsi2sd reg,reg is double decoded (vector on Athlon)
(define_insn_reservation "athlon_sseicvt_cvtsi2sd_k8" 11
(and (eq_attr "cpu" "k8,athlon,generic64")
@@ -1115,7 +1115,7 @@
(define_insn_reservation "athlon_ssemulvector_amdfam10" 4
(and (eq_attr "cpu" "amdfam10")
(eq_attr "type" "ssemul"))
- "athlon-direct,athlon-fpsched,athlon-fmul")
+ "athlon-direct,athlon-fpsched,athlon-fmul")
;; divsd timings. divss is faster
(define_insn_reservation "athlon_ssediv_load" 20
(and (eq_attr "cpu" "athlon")
@@ -1148,7 +1148,7 @@
(and (eq_attr "cpu" "amdfam10")
(and (eq_attr "type" "ssediv")
(eq_attr "memory" "load")))
- "athlon-direct,athlon-fploadk8,athlon-fmul*17")
+ "athlon-direct,athlon-fploadk8,athlon-fmul*17")
(define_insn_reservation "athlon_ssedivvector" 39
(and (eq_attr "cpu" "athlon")
(eq_attr "type" "ssediv"))
diff --git a/gcc/config/i386/bmmintrin.h b/gcc/config/i386/bmmintrin.h
index f321cee0d36..35c9831ded8 100644
--- a/gcc/config/i386/bmmintrin.h
+++ b/gcc/config/i386/bmmintrin.h
@@ -154,7 +154,7 @@ _mm_maccsd_epi16(__m128i __A, __m128i __B, __m128i __C)
static __inline __m128i __attribute__((__always_inline__))
_mm_maccd_epi16(__m128i __A, __m128i __B, __m128i __C)
{
- return (__m128i) __builtin_ia32_pmacswd ((__v8hi)__A, (__v8hi)__B, (__v4si)__C);
+ return (__m128i) __builtin_ia32_pmacswd ((__v8hi)__A, (__v8hi)__B, (__v4si)__C);
}
static __inline __m128i __attribute__((__always_inline__))
@@ -178,19 +178,19 @@ _mm_maccslo_epi32(__m128i __A, __m128i __B, __m128i __C)
static __inline __m128i __attribute__((__always_inline__))
_mm_macclo_epi32(__m128i __A, __m128i __B, __m128i __C)
{
- return (__m128i) __builtin_ia32_pmacsdql ((__v4si)__A, (__v4si)__B, (__v2di)__C);
+ return (__m128i) __builtin_ia32_pmacsdql ((__v4si)__A, (__v4si)__B, (__v2di)__C);
}
static __inline __m128i __attribute__((__always_inline__))
_mm_maccshi_epi32(__m128i __A, __m128i __B, __m128i __C)
{
- return (__m128i) __builtin_ia32_pmacssdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C);
+ return (__m128i) __builtin_ia32_pmacssdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C);
}
static __inline __m128i __attribute__((__always_inline__))
_mm_macchi_epi32(__m128i __A, __m128i __B, __m128i __C)
{
- return (__m128i) __builtin_ia32_pmacsdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C);
+ return (__m128i) __builtin_ia32_pmacsdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C);
}
static __inline __m128i __attribute__((__always_inline__))
@@ -324,25 +324,25 @@ _mm_perm_pd(__m128d __A, __m128d __B, __m128i __C)
/* Packed Integer Rotates and Shifts */
/* Rotates - Non-Immediate form */
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_rot_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_protb ((__v16qi)__A, (__v16qi)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_rot_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_protw ((__v8hi)__A, (__v8hi)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_rot_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_protd ((__v4si)__A, (__v4si)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_rot_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_protq ((__v2di)__A, (__v2di)__B);
@@ -351,25 +351,25 @@ _mm_rot_epi64(__m128i __A, __m128i __B)
/* Rotates - Immediate form */
#ifdef __OPTIMIZE__
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_roti_epi8(__m128i __A, int __B)
{
return (__m128i) __builtin_ia32_protbi ((__v16qi)__A, __B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_roti_epi16(__m128i __A, int __B)
{
return (__m128i) __builtin_ia32_protwi ((__v8hi)__A, __B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_roti_epi32(__m128i __A, int __B)
{
return (__m128i) __builtin_ia32_protdi ((__v4si)__A, __B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_roti_epi64(__m128i __A, int __B)
{
return (__m128i) __builtin_ia32_protqi ((__v2di)__A, __B);
@@ -383,50 +383,50 @@ _mm_roti_epi64(__m128i __A, int __B)
/* pshl */
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_shl_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshlb ((__v16qi)__A, (__v16qi)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_shl_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshlw ((__v8hi)__A, (__v8hi)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_shl_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshld ((__v4si)__A, (__v4si)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_shl_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshlq ((__v2di)__A, (__v2di)__B);
}
/* psha */
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_sha_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshab ((__v16qi)__A, (__v16qi)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_sha_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshaw ((__v8hi)__A, (__v8hi)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_sha_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshad ((__v4si)__A, (__v4si)__B);
}
-static __inline __m128i __attribute__((__always_inline__))
+static __inline __m128i __attribute__((__always_inline__))
_mm_sha_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pshaq ((__v2di)__A, (__v2di)__B);
@@ -465,14 +465,14 @@ _mm_comneq_ps(__m128 __A, __m128 __B)
return (__m128) __builtin_ia32_comuneqps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comnlt_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comunltps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
-_mm_comnle_ps(__m128 __A, __m128 __B)
+static __inline __m128 __attribute__((__always_inline__))
+_mm_comnle_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comunleps ((__v4sf)__A, (__v4sf)__B);
}
@@ -491,43 +491,43 @@ _mm_comueq_ps(__m128 __A, __m128 __B)
return (__m128) __builtin_ia32_comueqps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comnge_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comungeps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comngt_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comungtps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comfalse_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comfalseps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comoneq_ps(__m128 __A, __m128 __B)
{
- return (__m128) __builtin_ia32_comneqps ((__v4sf)__A, (__v4sf)__B);
+ return (__m128) __builtin_ia32_comneqps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comge_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comgeps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comgt_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comgtps ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comtrue_ps(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comtrueps ((__v4sf)__A, (__v4sf)__B);
@@ -565,14 +565,14 @@ _mm_comneq_pd(__m128d __A, __m128d __B)
return (__m128d) __builtin_ia32_comuneqpd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comnlt_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comunltpd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
-_mm_comnle_pd(__m128d __A, __m128d __B)
+static __inline __m128d __attribute__((__always_inline__))
+_mm_comnle_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comunlepd ((__v2df)__A, (__v2df)__B);
}
@@ -590,19 +590,19 @@ _mm_comueq_pd(__m128d __A, __m128d __B)
return (__m128d) __builtin_ia32_comueqpd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comnge_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comungepd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comngt_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comungtpd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comfalse_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comfalsepd ((__v2df)__A, (__v2df)__B);
@@ -614,19 +614,19 @@ _mm_comoneq_pd(__m128d __A, __m128d __B)
return (__m128d) __builtin_ia32_comneqpd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comge_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comgepd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comgt_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comgtpd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comtrue_pd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comtruepd ((__v2df)__A, (__v2df)__B);
@@ -663,14 +663,14 @@ _mm_comneq_ss(__m128 __A, __m128 __B)
return (__m128) __builtin_ia32_comuneqss ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comnlt_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comunltss ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
-_mm_comnle_ss(__m128 __A, __m128 __B)
+static __inline __m128 __attribute__((__always_inline__))
+_mm_comnle_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comunless ((__v4sf)__A, (__v4sf)__B);
}
@@ -688,43 +688,43 @@ _mm_comueq_ss(__m128 __A, __m128 __B)
return (__m128) __builtin_ia32_comueqss ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comnge_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comungess ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comngt_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comungtss ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comfalse_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comfalsess ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comoneq_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comneqss ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comge_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comgess ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comgt_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comgtss ((__v4sf)__A, (__v4sf)__B);
}
-static __inline __m128 __attribute__((__always_inline__))
+static __inline __m128 __attribute__((__always_inline__))
_mm_comtrue_ss(__m128 __A, __m128 __B)
{
return (__m128) __builtin_ia32_comtruess ((__v4sf)__A, (__v4sf)__B);
@@ -762,14 +762,14 @@ _mm_comneq_sd(__m128d __A, __m128d __B)
return (__m128d) __builtin_ia32_comuneqsd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comnlt_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comunltsd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
-_mm_comnle_sd(__m128d __A, __m128d __B)
+static __inline __m128d __attribute__((__always_inline__))
+_mm_comnle_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comunlesd ((__v2df)__A, (__v2df)__B);
}
@@ -787,19 +787,19 @@ _mm_comueq_sd(__m128d __A, __m128d __B)
return (__m128d) __builtin_ia32_comueqsd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comnge_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comungesd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comngt_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comungtsd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comfalse_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comfalsesd ((__v2df)__A, (__v2df)__B);
@@ -811,19 +811,19 @@ _mm_comoneq_sd(__m128d __A, __m128d __B)
return (__m128d) __builtin_ia32_comneqsd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comge_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comgesd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comgt_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comgtsd ((__v2df)__A, (__v2df)__B);
}
-static __inline __m128d __attribute__((__always_inline__))
+static __inline __m128d __attribute__((__always_inline__))
_mm_comtrue_sd(__m128d __A, __m128d __B)
{
return (__m128d) __builtin_ia32_comtruesd ((__v2df)__A, (__v2df)__B);
@@ -836,49 +836,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltub ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomleub ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtub ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgeub ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomequb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomnequb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalseub ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epu8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrueub ((__v16qi)__A, (__v16qi)__B);
-}
+}
/*pcom (integer, unsinged words) */
@@ -886,49 +886,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltuw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomleuw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtuw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgeuw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomequw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomnequw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalseuw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epu16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrueuw ((__v8hi)__A, (__v8hi)__B);
-}
+}
/*pcom (integer, unsinged double words) */
@@ -936,49 +936,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltud ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomleud ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtud ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgeud ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomequd ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomnequd ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalseud ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epu32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrueud ((__v4si)__A, (__v4si)__B);
-}
+}
/*pcom (integer, unsinged quad words) */
@@ -986,49 +986,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltuq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomleuq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtuq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgeuq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomequq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomnequq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalseuq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epu64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrueuq ((__v2di)__A, (__v2di)__B);
-}
+}
/*pcom (integer, signed bytes) */
@@ -1036,49 +1036,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomleb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgeb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomeqb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomneqb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalseb ((__v16qi)__A, (__v16qi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epi8(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrueb ((__v16qi)__A, (__v16qi)__B);
-}
+}
/*pcom (integer, signed words) */
@@ -1086,49 +1086,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomlew ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgew ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomeqw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomneqw ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalsew ((__v8hi)__A, (__v8hi)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epi16(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtruew ((__v8hi)__A, (__v8hi)__B);
-}
+}
/*pcom (integer, signed double words) */
@@ -1136,49 +1136,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltd ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomled ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtd ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomged ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomeqd ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomneqd ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalsed ((__v4si)__A, (__v4si)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epi32(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrued ((__v4si)__A, (__v4si)__B);
-}
+}
/*pcom (integer, signed quad words) */
@@ -1186,49 +1186,49 @@ static __inline __m128i __attribute__((__always_inline__))
_mm_comlt_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomltq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comle_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomleq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comgt_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgtq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comge_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomgeq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comeq_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomeqq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comneq_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomneqq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comfalse_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomfalseq ((__v2di)__A, (__v2di)__B);
-}
+}
static __inline __m128i __attribute__((__always_inline__))
_mm_comtrue_epi64(__m128i __A, __m128i __B)
{
return (__m128i) __builtin_ia32_pcomtrueq ((__v2di)__A, (__v2di)__B);
-}
+}
/* FRCZ */
static __inline __m128 __attribute__((__always_inline__))
diff --git a/gcc/config/i386/constraints.md b/gcc/config/i386/constraints.md
index 7f5737a627e..b9bec56e9af 100644
--- a/gcc/config/i386/constraints.md
+++ b/gcc/config/i386/constraints.md
@@ -18,7 +18,7 @@
;; <http://www.gnu.org/licenses/>.
;;; Unused letters:
-;;; B H TU W
+;;; B H TU W
;;; h jk vw z
;; Integer register constraints.
@@ -129,7 +129,7 @@
(match_test "IN_RANGE (ival, 0, 3)")))
(define_constraint "N"
- "Unsigned 8-bit integer constant (for @code{in} and @code{out}
+ "Unsigned 8-bit integer constant (for @code{in} and @code{out}
instructions)."
(and (match_code "const_int")
(match_test "IN_RANGE (ival, 0, 255)")))
diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c
index 0d46436c239..7bd5cb6a719 100644
--- a/gcc/config/i386/i386.c
+++ b/gcc/config/i386/i386.c
@@ -2011,7 +2011,7 @@ override_options (void)
{"pentium4", PROCESSOR_PENTIUM4, PTA_MMX |PTA_SSE | PTA_SSE2},
{"pentium4m", PROCESSOR_PENTIUM4, PTA_MMX | PTA_SSE | PTA_SSE2},
{"prescott", PROCESSOR_NOCONA, PTA_MMX | PTA_SSE | PTA_SSE2 | PTA_SSE3},
- {"nocona", PROCESSOR_NOCONA, (PTA_64BIT
+ {"nocona", PROCESSOR_NOCONA, (PTA_64BIT
| PTA_MMX | PTA_SSE | PTA_SSE2 | PTA_SSE3
| PTA_CX16 | PTA_NO_SAHF)},
{"core2", PROCESSOR_CORE2, (PTA_64BIT
@@ -2181,7 +2181,7 @@ override_options (void)
ix86_arch_string = TARGET_64BIT ? "x86-64" : "i386";
else
ix86_arch_specified = 1;
-
+
if (!strcmp (ix86_arch_string, "generic"))
error ("generic CPU can be used only for -mtune= switch");
if (!strncmp (ix86_arch_string, "generic", 7))
@@ -4558,7 +4558,7 @@ function_value_32 (enum machine_mode orig_mode, enum machine_mode mode,
else
/* Most things go in %eax. */
regno = 0;
-
+
/* Override FP return register with %xmm0 for local functions when
SSE math is enabled or for functions with sseregparm attribute. */
if ((fn || fntype) && (mode == SFmode || mode == DFmode))
@@ -4746,7 +4746,7 @@ ix86_return_in_memory (const_tree type)
but differs notably in that when MMX is available, 8-byte vectors
are returned in memory, rather than in MMX registers. */
-int
+int
ix86_sol10_return_in_memory (const_tree type)
{
int size;
@@ -4889,7 +4889,7 @@ setup_incoming_varargs_64 (CUMULATIVE_ARGS *cum)
ix86_save_varrargs_registers = 1;
/* We need 16-byte stack alignment to save SSE registers. If user
asked for lower preferred_stack_boundary, lets just hope that he knows
- what he is doing and won't varargs SSE values.
+ what he is doing and won't varargs SSE values.
We also may end up assuming that only 64bit values are stored in SSE
register let some floating point program work. */
@@ -19794,7 +19794,7 @@ ix86_expand_multi_arg_builtin (enum insn_code icode, tree exp, rtx target,
gcc_assert (GET_MODE (op) == mode || GET_MODE (op) == VOIDmode);
- if (optimize
+ if (optimize
|| ! (*insn_data[icode].operand[i+adjust+1].predicate) (op, mode)
|| num_memory > 1)
op = force_reg (mode, op);
@@ -21295,7 +21295,7 @@ ix86_veclibabi_acml (enum built_in_function fn, tree type_out, tree type_in)
|| n != 4)
return NULL_TREE;
break;
-
+
default:
return NULL_TREE;
}
@@ -24944,7 +24944,7 @@ bool ix86_sse5_valid_op_p (rtx operands[], rtx insn, int num, bool uses_oc0, int
/* format, example pmacsdd:
xmm1, xmm2, xmm3/mem, xmm1
-
+
For the integer multiply/add instructions be more restrictive and
require operands[2] and operands[3] to be the memory operands. */
else
@@ -25044,7 +25044,7 @@ static const struct attribute_spec ix86_attribute_table[] =
};
/* Implement targetm.vectorize.builtin_vectorization_cost. */
-static int
+static int
x86_builtin_vectorization_cost (bool runtime_test)
{
/* If the branch of the runtime test is taken - i.e. - the vectorized
diff --git a/gcc/config/i386/i386.h b/gcc/config/i386/i386.h
index cca89e30753..bc3e5f6dc9d 100644
--- a/gcc/config/i386/i386.h
+++ b/gcc/config/i386/i386.h
@@ -75,9 +75,9 @@ enum stringop_alg
When size is unknown, the UNKNOWN_SIZE alg is used. When size is
known at compile time or estimated via feedback, the SIZE array
is walked in order until MAX is greater then the estimate (or -1
- means infinity). Corresponding ALG is used then.
+ means infinity). Corresponding ALG is used then.
For example initializer:
- {{256, loop}, {-1, rep_prefix_4_byte}}
+ {{256, loop}, {-1, rep_prefix_4_byte}}
will use loop for blocks smaller or equal to 256 bytes, rep prefix will
be used otherwise. */
struct stringop_algs
@@ -153,7 +153,7 @@ struct processor_costs {
scalar-to-vector operation. */
const int vec_to_scalar_cost; /* Cost of vect-to-scalar operation. */
const int scalar_to_vec_cost; /* Cost of scalar-to-vector operation. */
- const int vec_align_load_cost; /* Cost of aligned vector load. */
+ const int vec_align_load_cost; /* Cost of aligned vector load. */
const int vec_unalign_load_cost; /* Cost of unaligned vector load. */
const int vec_store_cost; /* Cost of vector store. */
const int cond_taken_branch_cost; /* Cost of taken branch for vectorizer
@@ -375,7 +375,7 @@ enum ix86_arch_indices {
X86_ARCH_LAST
};
-
+
extern unsigned int ix86_arch_features[X86_ARCH_LAST];
#define TARGET_CMOVE ix86_arch_features[X86_ARCH_CMOVE]
@@ -2507,7 +2507,7 @@ struct machine_function GTY(())
#define TARG_SCALAR_STORE_COST ix86_cost->scalar_store_cost
/* Cost of any vector operation, excluding load, store or vector to scalar
- operation. */
+ operation. */
#undef TARG_VEC_STMT_COST
#define TARG_VEC_STMT_COST ix86_cost->vec_stmt_cost
diff --git a/gcc/config/i386/i386.md b/gcc/config/i386/i386.md
index cbf516175ff..89ae741d694 100644
--- a/gcc/config/i386/i386.md
+++ b/gcc/config/i386/i386.md
@@ -158,7 +158,7 @@
; For SSE4A support
(UNSPEC_EXTRQI 130)
- (UNSPEC_EXTRQ 131)
+ (UNSPEC_EXTRQ 131)
(UNSPEC_INSERTQI 132)
(UNSPEC_INSERTQ 133)
@@ -1166,7 +1166,7 @@
(set_attr "mode" "SI")])
;; Pentium Pro can do steps 1 through 3 in one go.
-;; comi*, ucomi*, fcomi*, ficomi*,fucomi* (i387 instructions set condition codes)
+;; comi*, ucomi*, fcomi*, ficomi*,fucomi* (i387 instructions set condition codes)
(define_insn "*cmpfp_i_mixed"
[(set (reg:CCFP FLAGS_REG)
(compare:CCFP (match_operand 0 "register_operand" "f,x")
@@ -1476,7 +1476,7 @@
(set_attr "mode" "SI")
(set_attr "pent_pair" "np")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "double")])
+ (set_attr "amdfam10_decode" "double")])
(define_expand "movhi"
[(set (match_operand:HI 0 "nonimmediate_operand" "")
@@ -1594,7 +1594,7 @@
(set_attr "mode" "SI")
(set_attr "pent_pair" "np")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "double")])
+ (set_attr "amdfam10_decode" "double")])
;; Not added amdfam10_decode since TARGET_PARTIAL_REG_STALL is disabled for AMDFAM10
(define_insn "*swaphi_2"
@@ -1770,7 +1770,7 @@
(set_attr "mode" "SI")
(set_attr "pent_pair" "np")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "vector")])
+ (set_attr "amdfam10_decode" "vector")])
;; Not added amdfam10_decode since TARGET_PARTIAL_REG_STALL is disabled for AMDFAM10
(define_insn "*swapqi_2"
@@ -2343,7 +2343,7 @@
(set_attr "mode" "DI")
(set_attr "pent_pair" "np")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "double")])
+ (set_attr "amdfam10_decode" "double")])
(define_expand "movti"
[(set (match_operand:TI 0 "nonimmediate_operand" "")
@@ -3950,7 +3950,7 @@
[(set (match_operand:DF 0 "register_operand" "")
(float_extend:DF
(match_operand:SF 1 "nonimmediate_operand" "")))]
- "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size
+ "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size
&& reload_completed && SSE_REG_P (operands[0])"
[(set (match_dup 2)
(float_extend:V2DF
@@ -3978,7 +3978,7 @@
emit_insn (gen_sse_unpcklps (operands[3], operands[3], operands[3]));
}
else
- emit_insn (gen_vec_setv4sf_0 (operands[3],
+ emit_insn (gen_vec_setv4sf_0 (operands[3],
CONST0_RTX (V4SFmode), operands[1]));
})
@@ -4086,7 +4086,7 @@
[(set (match_operand:SF 0 "register_operand" "")
(float_truncate:SF
(match_operand:DF 1 "nonimmediate_operand" "")))]
- "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size
+ "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size
&& reload_completed && SSE_REG_P (operands[0])"
[(set (match_dup 2)
(vec_concat:V4SF
@@ -4756,7 +4756,7 @@
(set_attr "mode" "HI")
(set_attr "unit" "i387")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "vector")])
+ (set_attr "amdfam10_decode" "vector")])
;; Conversion between fixed point and floating point.
@@ -4851,7 +4851,7 @@
(define_insn "*floatsisf2_mixed_vector"
[(set (match_operand:SF 0 "register_operand" "=x,f,?f")
(float:SF (match_operand:SI 1 "nonimmediate_operand" "x,m,r")))]
- "TARGET_MIX_SSE_I387 && !flag_trapping_math
+ "TARGET_MIX_SSE_I387 && !flag_trapping_math
&& TARGET_USE_VECTOR_CONVERTS && !optimize_size"
"@
cvtdq2ps\t{%1, %0|%0, %1}
@@ -4914,7 +4914,7 @@
"#"
[(set_attr "type" "multi")])
-(define_split
+(define_split
[(set (match_operand:SF 0 "register_operand" "")
(float:SF (match_operand:SI 1 "nonimmediate_operand" "")))]
"flag_trapping_math
@@ -4929,7 +4929,7 @@
emit_insn (gen_sse2_loadld (operands[2], CONST0_RTX (V4SImode), operands[1]));
})
-(define_split
+(define_split
[(set (match_operand:SF 0 "register_operand" "")
(float:SF (match_operand:SI 1 "register_operand" "")))]
"flag_trapping_math
@@ -5042,7 +5042,7 @@
(set_attr "amdfam10_decode" "double")
(set_attr "fp_int_src" "true")])
-(define_split
+(define_split
[(set (match_operand:DF 0 "register_operand" "")
(float:DF (match_operand:SI 1 "memory_operand" "")))]
"TARGET_USE_VECTOR_CONVERTS && reload_completed
@@ -7483,11 +7483,11 @@
"TARGET_64BIT"
"")
-;; On AMDFAM10
+;; On AMDFAM10
;; IMUL reg64, reg64, imm8 Direct
;; IMUL reg64, mem64, imm8 VectorPath
;; IMUL reg64, reg64, imm32 Direct
-;; IMUL reg64, mem64, imm32 VectorPath
+;; IMUL reg64, mem64, imm32 VectorPath
;; IMUL reg64, reg64 Direct
;; IMUL reg64, mem64 Direct
@@ -7517,7 +7517,7 @@
(cond [(and (eq_attr "alternative" "0,1")
(match_operand 1 "memory_operand" ""))
(const_string "vector")]
- (const_string "direct")))
+ (const_string "direct")))
(set_attr "mode" "DI")])
(define_expand "mulsi3"
@@ -7528,7 +7528,7 @@
""
"")
-;; On AMDFAM10
+;; On AMDFAM10
;; IMUL reg32, reg32, imm8 Direct
;; IMUL reg32, mem32, imm8 VectorPath
;; IMUL reg32, reg32, imm32 Direct
@@ -7561,7 +7561,7 @@
(cond [(and (eq_attr "alternative" "0,1")
(match_operand 1 "memory_operand" ""))
(const_string "vector")]
- (const_string "direct")))
+ (const_string "direct")))
(set_attr "mode" "SI")])
(define_insn "*mulsi3_1_zext"
@@ -7591,7 +7591,7 @@
(cond [(and (eq_attr "alternative" "0,1")
(match_operand 1 "memory_operand" ""))
(const_string "vector")]
- (const_string "direct")))
+ (const_string "direct")))
(set_attr "mode" "SI")])
(define_expand "mulhi3"
@@ -7659,7 +7659,7 @@
(if_then_else (eq_attr "cpu" "athlon")
(const_string "vector")
(const_string "direct")))
- (set_attr "amdfam10_decode" "direct")
+ (set_attr "amdfam10_decode" "direct")
(set_attr "mode" "QI")])
(define_expand "umulqihi3"
@@ -7711,7 +7711,7 @@
(if_then_else (eq_attr "cpu" "athlon")
(const_string "vector")
(const_string "direct")))
- (set_attr "amdfam10_decode" "direct")
+ (set_attr "amdfam10_decode" "direct")
(set_attr "mode" "QI")])
(define_expand "umulditi3"
@@ -7738,7 +7738,7 @@
(if_then_else (eq_attr "cpu" "athlon")
(const_string "vector")
(const_string "double")))
- (set_attr "amdfam10_decode" "double")
+ (set_attr "amdfam10_decode" "double")
(set_attr "mode" "DI")])
;; We can't use this pattern in 64bit mode, since it results in two separate 32bit registers
@@ -7766,7 +7766,7 @@
(if_then_else (eq_attr "cpu" "athlon")
(const_string "vector")
(const_string "double")))
- (set_attr "amdfam10_decode" "double")
+ (set_attr "amdfam10_decode" "double")
(set_attr "mode" "SI")])
(define_expand "mulditi3"
@@ -7820,7 +7820,7 @@
(if_then_else (eq_attr "cpu" "athlon")
(const_string "vector")
(const_string "double")))
- (set_attr "amdfam10_decode" "double")
+ (set_attr "amdfam10_decode" "double")
(set_attr "mode" "SI")])
(define_expand "umuldi3_highpart"
@@ -7857,7 +7857,7 @@
(if_then_else (eq_attr "cpu" "athlon")
(const_string "vector")
(const_string "double")))
- (set_attr "amdfam10_decode" "double")
+ (set_attr "amdfam10_decode" "double")
(set_attr "mode" "DI")])
(define_expand "umulsi3_highpart"
@@ -10991,7 +10991,7 @@
(set_attr "prefix_0f" "1")
(set_attr "mode" "DI")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "vector")])
+ (set_attr "amdfam10_decode" "vector")])
(define_expand "x86_64_shift_adj"
[(set (reg:CCZ FLAGS_REG)
@@ -11209,7 +11209,7 @@
(set_attr "mode" "SI")
(set_attr "pent_pair" "np")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "vector")])
+ (set_attr "amdfam10_decode" "vector")])
(define_expand "x86_shift_adj_1"
[(set (reg:CCZ FLAGS_REG)
@@ -11971,7 +11971,7 @@
(set_attr "prefix_0f" "1")
(set_attr "mode" "DI")
(set_attr "athlon_decode" "vector")
- (set_attr "amdfam10_decode" "vector")])
+ (set_attr "amdfam10_decode" "vector")])
(define_expand "ashrdi3"
[(set (match_operand:DI 0 "shiftdi_operand" "")
@@ -16606,7 +16606,7 @@
"fsqrt"
[(set_attr "type" "fpspc")
(set_attr "mode" "XF")
- (set_attr "athlon_decode" "direct")
+ (set_attr "athlon_decode" "direct")
(set_attr "amdfam10_decode" "direct")])
(define_insn "*rsqrtsf2_sse"
@@ -19734,7 +19734,7 @@
;; SSE5 conditional move
(define_insn "*sse5_pcmov_<mode>"
[(set (match_operand:MODEF 0 "register_operand" "=x,x,x,x")
- (if_then_else:MODEF
+ (if_then_else:MODEF
(match_operand:MODEF 1 "nonimmediate_operand" "xm,x,0,0")
(match_operand:MODEF 2 "nonimmediate_operand" "0,0,x,xm")
(match_operand:MODEF 3 "vector_move_operand" "x,xm,xm,x")))]
diff --git a/gcc/config/i386/mmx.md b/gcc/config/i386/mmx.md
index 4123bc33b22..20c4fb05b57 100644
--- a/gcc/config/i386/mmx.md
+++ b/gcc/config/i386/mmx.md
@@ -22,7 +22,7 @@
;; the same register file, and 3dNOW! adds a number of extensions to
;; the base integer MMX isa.
-;; Note! Except for the basic move instructions, *all* of these
+;; Note! Except for the basic move instructions, *all* of these
;; patterns are outside the normal optabs namespace. This is because
;; use of these registers requires the insertion of emms or femms
;; instructions to return to normal fpu mode. The compiler doesn't
@@ -301,7 +301,7 @@
"pfrsqrt\\t{%1, %0|%0, %1}"
[(set_attr "type" "mmx")
(set_attr "mode" "V2SF")])
-
+
(define_insn "mmx_rsqit1v2sf3"
[(set (match_operand:V2SF 0 "register_operand" "=y")
(unspec:V2SF [(match_operand:V2SF 1 "register_operand" "0")
@@ -1429,4 +1429,4 @@
"TARGET_3DNOW"
"femms"
[(set_attr "type" "mmx")
- (set_attr "memory" "none")])
+ (set_attr "memory" "none")])
diff --git a/gcc/config/i386/ppro.md b/gcc/config/i386/ppro.md
index f35a784f74f..b971e51089e 100644
--- a/gcc/config/i386/ppro.md
+++ b/gcc/config/i386/ppro.md
@@ -78,7 +78,7 @@
;; but still in only one cycle.
;; - a complex (microcode) instruction can also only be decoded by
;; decoder 0, and this takes an unspecified number of cycles.
-;;
+;;
;; The goal is to schedule such that we have a few-one-one uops sequence
;; in each cycle, to decode as many instructions per cycle as possible.
(define_cpu_unit "decoder0" "ppro_decoder")
diff --git a/gcc/config/i386/sse.md b/gcc/config/i386/sse.md
index e92ac733c99..a810ad68e5c 100644
--- a/gcc/config/i386/sse.md
+++ b/gcc/config/i386/sse.md
@@ -602,7 +602,7 @@
(set_attr "mode" "SF")])
;; ??? For !flag_finite_math_only, the representation with SMIN/SMAX
-;; isn't really correct, as those rtl operators aren't defined when
+;; isn't really correct, as those rtl operators aren't defined when
;; applied to NaNs. Hopefully the optimizers won't get too smart on us.
(define_expand "smaxv4sf3"
@@ -754,7 +754,7 @@
(vec_concat:V4SF
(vec_concat:V2SF
(plus:SF
- (vec_select:SF
+ (vec_select:SF
(match_operand:V4SF 1 "register_operand" "0")
(parallel [(const_int 0)]))
(vec_select:SF (match_dup 1) (parallel [(const_int 1)])))
@@ -781,7 +781,7 @@
(vec_concat:V4SF
(vec_concat:V2SF
(minus:SF
- (vec_select:SF
+ (vec_select:SF
(match_operand:V4SF 1 "register_operand" "0")
(parallel [(const_int 0)]))
(vec_select:SF (match_dup 1) (parallel [(const_int 1)])))
@@ -2290,7 +2290,7 @@
(set_attr "mode" "DF")])
;; ??? For !flag_finite_math_only, the representation with SMIN/SMAX
-;; isn't really correct, as those rtl operators aren't defined when
+;; isn't really correct, as those rtl operators aren't defined when
;; applied to NaNs. Hopefully the optimizers won't get too smart on us.
(define_expand "smaxv2df3"
@@ -3527,7 +3527,7 @@
emit_insn (gen_sse2_punpcklbw (t[3], operands[2], operands[2]));
/* Multiply words. The end-of-line annotations here give a picture of what
- the output of that instruction looks like. Dot means don't care; the
+ the output of that instruction looks like. Dot means don't care; the
letters are the bytes of the result with A being the most significant. */
emit_insn (gen_mulv8hi3 (gen_lowpart (V8HImode, t[4]), /* .A.B.C.D.E.F.G.H */
gen_lowpart (V8HImode, t[0]),
@@ -3569,8 +3569,8 @@
(define_expand "smulv8hi3_highpart"
[(set (match_operand:V8HI 0 "register_operand" "")
(truncate:V8HI
- (lshiftrt:V8SI
- (mult:V8SI
+ (lshiftrt:V8SI
+ (mult:V8SI
(sign_extend:V8SI
(match_operand:V8HI 1 "nonimmediate_operand" ""))
(sign_extend:V8SI
@@ -3578,7 +3578,7 @@
(const_int 16))))]
"TARGET_SSE2"
"ix86_fixup_binary_operands_no_copy (MULT, V8HImode, operands);")
-
+
(define_insn "*smulv8hi3_highpart"
[(set (match_operand:V8HI 0 "register_operand" "=x")
(truncate:V8HI
@@ -3776,7 +3776,7 @@
thirtytwo));
emit_insn (gen_sse2_lshrti3 (gen_lowpart (TImode, t3),
gen_lowpart (TImode, op2),
- thirtytwo));
+ thirtytwo));
/* Multiply elements 3 and 1. */
emit_insn (gen_sse2_umulv2siv2di3 (gen_lowpart (V2DImode, t4),
t2, t3));
@@ -4011,8 +4011,8 @@
})
(define_expand "udot_prodv4si"
- [(match_operand:V2DI 0 "register_operand" "")
- (match_operand:V4SI 1 "register_operand" "")
+ [(match_operand:V2DI 0 "register_operand" "")
+ (match_operand:V4SI 1 "register_operand" "")
(match_operand:V4SI 2 "register_operand" "")
(match_operand:V2DI 3 "register_operand" "")]
"TARGET_SSE2"
@@ -4875,7 +4875,7 @@
(const_int 9) (const_int 25)
(const_int 10) (const_int 26)
(const_int 11) (const_int 27)
- (const_int 12) (const_int 28)
+ (const_int 12) (const_int 28)
(const_int 13) (const_int 29)
(const_int 14) (const_int 30)
(const_int 15) (const_int 31)])))]
@@ -5283,7 +5283,7 @@
"TARGET_SSE"
"#"
"&& reload_completed
- && (TARGET_INTER_UNIT_MOVES
+ && (TARGET_INTER_UNIT_MOVES
|| MEM_P (operands [0])
|| !GENERAL_REGNO_P (true_regnum (operands [0])))"
[(set (match_dup 0) (match_dup 1))]
@@ -5343,7 +5343,7 @@
(parallel [(const_int 0)])))]
"TARGET_SSE
&& reload_completed
- && (TARGET_INTER_UNIT_MOVES
+ && (TARGET_INTER_UNIT_MOVES
|| MEM_P (operands [0])
|| !GENERAL_REGNO_P (true_regnum (operands [0])))"
[(set (match_dup 0) (match_dup 1))]
@@ -5839,7 +5839,7 @@
(set_attr "prefix_data16" "1")
(set_attr "mode" "TI")])
-;; The correct representation for this is absolutely enormous, and
+;; The correct representation for this is absolutely enormous, and
;; surely not generally useful.
(define_insn "sse2_psadbw"
[(set (match_operand:V2DI 0 "register_operand" "=x")
@@ -6641,7 +6641,7 @@
(define_insn "sse4a_vmmovntv2df"
[(set (match_operand:DF 0 "memory_operand" "=m")
- (unspec:DF [(vec_select:DF
+ (unspec:DF [(vec_select:DF
(match_operand:V2DF 1 "register_operand" "x")
(parallel [(const_int 0)]))]
UNSPEC_MOVNT))]
@@ -6661,7 +6661,7 @@
(define_insn "sse4a_vmmovntv4sf"
[(set (match_operand:SF 0 "memory_operand" "=m")
- (unspec:SF [(vec_select:SF
+ (unspec:SF [(vec_select:SF
(match_operand:V4SF 1 "register_operand" "x")
(parallel [(const_int 0)]))]
UNSPEC_MOVNT))]
@@ -7893,7 +7893,7 @@
;; SSE5 parallel XMM conditional moves
(define_insn "sse5_pcmov_<mode>"
[(set (match_operand:SSEMODE 0 "register_operand" "=x,x,x,x,x,x")
- (if_then_else:SSEMODE
+ (if_then_else:SSEMODE
(match_operand:SSEMODE 3 "nonimmediate_operand" "0,0,xm,xm,0,0")
(match_operand:SSEMODE 1 "vector_move_operand" "x,xm,0,x,C,x")
(match_operand:SSEMODE 2 "vector_move_operand" "xm,x,x,0,x,C")))]
diff --git a/gcc/config/i386/sync.md b/gcc/config/i386/sync.md
index 19295bbaff1..a4f6d38456c 100644
--- a/gcc/config/i386/sync.md
+++ b/gcc/config/i386/sync.md
@@ -34,7 +34,7 @@
;; ??? It would be possible to use cmpxchg8b on pentium for DImode
;; changes. It's complicated because the insn uses ecx:ebx as the
;; new value; note that the registers are reversed from the order
-;; that they'd be in with (reg:DI 2 ecx). Similarly for TImode
+;; that they'd be in with (reg:DI 2 ecx). Similarly for TImode
;; data in 64-bit mode.
(define_expand "sync_compare_and_swap<mode>"