diff options
author | meissner <meissner@138bc75d-0d04-0410-961f-82ee72b054a4> | 2007-09-19 21:41:08 +0000 |
---|---|---|
committer | meissner <meissner@138bc75d-0d04-0410-961f-82ee72b054a4> | 2007-09-19 21:41:08 +0000 |
commit | 009b318f3d6363ef0a41b8e33c63d3e723c55fb3 (patch) | |
tree | 388fd26895e48eecd5226317111e04f710d0ed07 | |
parent | 66ce45bdc273da0235b2532adc784c4b570ca667 (diff) | |
download | gcc-009b318f3d6363ef0a41b8e33c63d3e723c55fb3.tar.gz |
Eliminate trailing whitespace
git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@128605 138bc75d-0d04-0410-961f-82ee72b054a4
-rw-r--r-- | gcc/config/i386/athlon.md | 14 | ||||
-rw-r--r-- | gcc/config/i386/bmmintrin.h | 246 | ||||
-rw-r--r-- | gcc/config/i386/constraints.md | 4 | ||||
-rw-r--r-- | gcc/config/i386/i386.c | 18 | ||||
-rw-r--r-- | gcc/config/i386/i386.h | 10 | ||||
-rw-r--r-- | gcc/config/i386/i386.md | 62 | ||||
-rw-r--r-- | gcc/config/i386/mmx.md | 6 | ||||
-rw-r--r-- | gcc/config/i386/ppro.md | 2 | ||||
-rw-r--r-- | gcc/config/i386/sse.md | 36 | ||||
-rw-r--r-- | gcc/config/i386/sync.md | 2 |
10 files changed, 200 insertions, 200 deletions
diff --git a/gcc/config/i386/athlon.md b/gcc/config/i386/athlon.md index 04b1e1e2d34..c9860def7c6 100644 --- a/gcc/config/i386/athlon.md +++ b/gcc/config/i386/athlon.md @@ -1,7 +1,7 @@ ;; AMD Athlon Scheduling ;; ;; The Athlon does contain three pipelined FP units, three integer units and -;; three address generation units. +;; three address generation units. ;; ;; The predecode logic is determining boundaries of instructions in the 64 ;; byte cache line. So the cache line straddling problem of K6 might be issue @@ -206,7 +206,7 @@ (and (eq_attr "type" "imul") (and (eq_attr "mode" "HI") (eq_attr "memory" "none,unknown")))) - "athlon-vector,athlon-ieu0,athlon-mult,nothing,athlon-ieu0") + "athlon-vector,athlon-ieu0,athlon-mult,nothing,athlon-ieu0") (define_insn_reservation "athlon_imul_mem" 8 (and (eq_attr "cpu" "athlon") (and (eq_attr "type" "imul") @@ -328,7 +328,7 @@ (eq_attr "memory" "both")))) "athlon-direct,athlon-load, athlon-ieu,athlon-store, - athlon-store") + athlon-store") (define_insn_reservation "athlon_ivector_both" 6 (and (eq_attr "cpu" "athlon,k8,generic64") @@ -617,7 +617,7 @@ (and (eq_attr "cpu" "amdfam10") (and (eq_attr "type" "mmxmov") (eq_attr "memory" "load"))) - "athlon-direct,athlon-fploadk8, athlon-fany") + "athlon-direct,athlon-fploadk8, athlon-fany") (define_insn_reservation "athlon_mmxssest" 3 (and (eq_attr "cpu" "k8,generic64") (and (eq_attr "type" "mmxmov,ssemov") @@ -926,7 +926,7 @@ (and (eq_attr "amdfam10_decode" "double") (and (eq_attr "mode" "SF,DF") (eq_attr "memory" "load"))))) - "athlon-double,athlon-fploadk8,(athlon-faddmul+athlon-fstore)") + "athlon-double,athlon-fploadk8,(athlon-faddmul+athlon-fstore)") ;; cvtsi2sd reg,reg is double decoded (vector on Athlon) (define_insn_reservation "athlon_sseicvt_cvtsi2sd_k8" 11 (and (eq_attr "cpu" "k8,athlon,generic64") @@ -1115,7 +1115,7 @@ (define_insn_reservation "athlon_ssemulvector_amdfam10" 4 (and (eq_attr "cpu" "amdfam10") (eq_attr "type" "ssemul")) - "athlon-direct,athlon-fpsched,athlon-fmul") + "athlon-direct,athlon-fpsched,athlon-fmul") ;; divsd timings. divss is faster (define_insn_reservation "athlon_ssediv_load" 20 (and (eq_attr "cpu" "athlon") @@ -1148,7 +1148,7 @@ (and (eq_attr "cpu" "amdfam10") (and (eq_attr "type" "ssediv") (eq_attr "memory" "load"))) - "athlon-direct,athlon-fploadk8,athlon-fmul*17") + "athlon-direct,athlon-fploadk8,athlon-fmul*17") (define_insn_reservation "athlon_ssedivvector" 39 (and (eq_attr "cpu" "athlon") (eq_attr "type" "ssediv")) diff --git a/gcc/config/i386/bmmintrin.h b/gcc/config/i386/bmmintrin.h index f321cee0d36..35c9831ded8 100644 --- a/gcc/config/i386/bmmintrin.h +++ b/gcc/config/i386/bmmintrin.h @@ -154,7 +154,7 @@ _mm_maccsd_epi16(__m128i __A, __m128i __B, __m128i __C) static __inline __m128i __attribute__((__always_inline__)) _mm_maccd_epi16(__m128i __A, __m128i __B, __m128i __C) { - return (__m128i) __builtin_ia32_pmacswd ((__v8hi)__A, (__v8hi)__B, (__v4si)__C); + return (__m128i) __builtin_ia32_pmacswd ((__v8hi)__A, (__v8hi)__B, (__v4si)__C); } static __inline __m128i __attribute__((__always_inline__)) @@ -178,19 +178,19 @@ _mm_maccslo_epi32(__m128i __A, __m128i __B, __m128i __C) static __inline __m128i __attribute__((__always_inline__)) _mm_macclo_epi32(__m128i __A, __m128i __B, __m128i __C) { - return (__m128i) __builtin_ia32_pmacsdql ((__v4si)__A, (__v4si)__B, (__v2di)__C); + return (__m128i) __builtin_ia32_pmacsdql ((__v4si)__A, (__v4si)__B, (__v2di)__C); } static __inline __m128i __attribute__((__always_inline__)) _mm_maccshi_epi32(__m128i __A, __m128i __B, __m128i __C) { - return (__m128i) __builtin_ia32_pmacssdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C); + return (__m128i) __builtin_ia32_pmacssdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C); } static __inline __m128i __attribute__((__always_inline__)) _mm_macchi_epi32(__m128i __A, __m128i __B, __m128i __C) { - return (__m128i) __builtin_ia32_pmacsdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C); + return (__m128i) __builtin_ia32_pmacsdqh ((__v4si)__A, (__v4si)__B, (__v2di)__C); } static __inline __m128i __attribute__((__always_inline__)) @@ -324,25 +324,25 @@ _mm_perm_pd(__m128d __A, __m128d __B, __m128i __C) /* Packed Integer Rotates and Shifts */ /* Rotates - Non-Immediate form */ -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_rot_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_protb ((__v16qi)__A, (__v16qi)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_rot_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_protw ((__v8hi)__A, (__v8hi)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_rot_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_protd ((__v4si)__A, (__v4si)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_rot_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_protq ((__v2di)__A, (__v2di)__B); @@ -351,25 +351,25 @@ _mm_rot_epi64(__m128i __A, __m128i __B) /* Rotates - Immediate form */ #ifdef __OPTIMIZE__ -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_roti_epi8(__m128i __A, int __B) { return (__m128i) __builtin_ia32_protbi ((__v16qi)__A, __B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_roti_epi16(__m128i __A, int __B) { return (__m128i) __builtin_ia32_protwi ((__v8hi)__A, __B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_roti_epi32(__m128i __A, int __B) { return (__m128i) __builtin_ia32_protdi ((__v4si)__A, __B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_roti_epi64(__m128i __A, int __B) { return (__m128i) __builtin_ia32_protqi ((__v2di)__A, __B); @@ -383,50 +383,50 @@ _mm_roti_epi64(__m128i __A, int __B) /* pshl */ -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_shl_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshlb ((__v16qi)__A, (__v16qi)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_shl_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshlw ((__v8hi)__A, (__v8hi)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_shl_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshld ((__v4si)__A, (__v4si)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_shl_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshlq ((__v2di)__A, (__v2di)__B); } /* psha */ -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_sha_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshab ((__v16qi)__A, (__v16qi)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_sha_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshaw ((__v8hi)__A, (__v8hi)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_sha_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshad ((__v4si)__A, (__v4si)__B); } -static __inline __m128i __attribute__((__always_inline__)) +static __inline __m128i __attribute__((__always_inline__)) _mm_sha_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pshaq ((__v2di)__A, (__v2di)__B); @@ -465,14 +465,14 @@ _mm_comneq_ps(__m128 __A, __m128 __B) return (__m128) __builtin_ia32_comuneqps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comnlt_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comunltps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) -_mm_comnle_ps(__m128 __A, __m128 __B) +static __inline __m128 __attribute__((__always_inline__)) +_mm_comnle_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comunleps ((__v4sf)__A, (__v4sf)__B); } @@ -491,43 +491,43 @@ _mm_comueq_ps(__m128 __A, __m128 __B) return (__m128) __builtin_ia32_comueqps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comnge_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comungeps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comngt_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comungtps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comfalse_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comfalseps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comoneq_ps(__m128 __A, __m128 __B) { - return (__m128) __builtin_ia32_comneqps ((__v4sf)__A, (__v4sf)__B); + return (__m128) __builtin_ia32_comneqps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comge_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comgeps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comgt_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comgtps ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comtrue_ps(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comtrueps ((__v4sf)__A, (__v4sf)__B); @@ -565,14 +565,14 @@ _mm_comneq_pd(__m128d __A, __m128d __B) return (__m128d) __builtin_ia32_comuneqpd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comnlt_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comunltpd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) -_mm_comnle_pd(__m128d __A, __m128d __B) +static __inline __m128d __attribute__((__always_inline__)) +_mm_comnle_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comunlepd ((__v2df)__A, (__v2df)__B); } @@ -590,19 +590,19 @@ _mm_comueq_pd(__m128d __A, __m128d __B) return (__m128d) __builtin_ia32_comueqpd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comnge_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comungepd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comngt_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comungtpd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comfalse_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comfalsepd ((__v2df)__A, (__v2df)__B); @@ -614,19 +614,19 @@ _mm_comoneq_pd(__m128d __A, __m128d __B) return (__m128d) __builtin_ia32_comneqpd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comge_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comgepd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comgt_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comgtpd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comtrue_pd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comtruepd ((__v2df)__A, (__v2df)__B); @@ -663,14 +663,14 @@ _mm_comneq_ss(__m128 __A, __m128 __B) return (__m128) __builtin_ia32_comuneqss ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comnlt_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comunltss ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) -_mm_comnle_ss(__m128 __A, __m128 __B) +static __inline __m128 __attribute__((__always_inline__)) +_mm_comnle_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comunless ((__v4sf)__A, (__v4sf)__B); } @@ -688,43 +688,43 @@ _mm_comueq_ss(__m128 __A, __m128 __B) return (__m128) __builtin_ia32_comueqss ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comnge_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comungess ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comngt_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comungtss ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comfalse_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comfalsess ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comoneq_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comneqss ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comge_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comgess ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comgt_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comgtss ((__v4sf)__A, (__v4sf)__B); } -static __inline __m128 __attribute__((__always_inline__)) +static __inline __m128 __attribute__((__always_inline__)) _mm_comtrue_ss(__m128 __A, __m128 __B) { return (__m128) __builtin_ia32_comtruess ((__v4sf)__A, (__v4sf)__B); @@ -762,14 +762,14 @@ _mm_comneq_sd(__m128d __A, __m128d __B) return (__m128d) __builtin_ia32_comuneqsd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comnlt_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comunltsd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) -_mm_comnle_sd(__m128d __A, __m128d __B) +static __inline __m128d __attribute__((__always_inline__)) +_mm_comnle_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comunlesd ((__v2df)__A, (__v2df)__B); } @@ -787,19 +787,19 @@ _mm_comueq_sd(__m128d __A, __m128d __B) return (__m128d) __builtin_ia32_comueqsd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comnge_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comungesd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comngt_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comungtsd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comfalse_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comfalsesd ((__v2df)__A, (__v2df)__B); @@ -811,19 +811,19 @@ _mm_comoneq_sd(__m128d __A, __m128d __B) return (__m128d) __builtin_ia32_comneqsd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comge_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comgesd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comgt_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comgtsd ((__v2df)__A, (__v2df)__B); } -static __inline __m128d __attribute__((__always_inline__)) +static __inline __m128d __attribute__((__always_inline__)) _mm_comtrue_sd(__m128d __A, __m128d __B) { return (__m128d) __builtin_ia32_comtruesd ((__v2df)__A, (__v2df)__B); @@ -836,49 +836,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltub ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomleub ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtub ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgeub ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomequb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomnequb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalseub ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epu8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrueub ((__v16qi)__A, (__v16qi)__B); -} +} /*pcom (integer, unsinged words) */ @@ -886,49 +886,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltuw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomleuw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtuw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgeuw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomequw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomnequw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalseuw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epu16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrueuw ((__v8hi)__A, (__v8hi)__B); -} +} /*pcom (integer, unsinged double words) */ @@ -936,49 +936,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltud ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomleud ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtud ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgeud ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomequd ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomnequd ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalseud ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epu32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrueud ((__v4si)__A, (__v4si)__B); -} +} /*pcom (integer, unsinged quad words) */ @@ -986,49 +986,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltuq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomleuq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtuq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgeuq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomequq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomnequq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalseuq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epu64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrueuq ((__v2di)__A, (__v2di)__B); -} +} /*pcom (integer, signed bytes) */ @@ -1036,49 +1036,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomleb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgeb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomeqb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomneqb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalseb ((__v16qi)__A, (__v16qi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epi8(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrueb ((__v16qi)__A, (__v16qi)__B); -} +} /*pcom (integer, signed words) */ @@ -1086,49 +1086,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomlew ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgew ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomeqw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomneqw ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalsew ((__v8hi)__A, (__v8hi)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epi16(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtruew ((__v8hi)__A, (__v8hi)__B); -} +} /*pcom (integer, signed double words) */ @@ -1136,49 +1136,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltd ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomled ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtd ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomged ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomeqd ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomneqd ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalsed ((__v4si)__A, (__v4si)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epi32(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrued ((__v4si)__A, (__v4si)__B); -} +} /*pcom (integer, signed quad words) */ @@ -1186,49 +1186,49 @@ static __inline __m128i __attribute__((__always_inline__)) _mm_comlt_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomltq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comle_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomleq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comgt_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgtq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comge_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomgeq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comeq_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomeqq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comneq_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomneqq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comfalse_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomfalseq ((__v2di)__A, (__v2di)__B); -} +} static __inline __m128i __attribute__((__always_inline__)) _mm_comtrue_epi64(__m128i __A, __m128i __B) { return (__m128i) __builtin_ia32_pcomtrueq ((__v2di)__A, (__v2di)__B); -} +} /* FRCZ */ static __inline __m128 __attribute__((__always_inline__)) diff --git a/gcc/config/i386/constraints.md b/gcc/config/i386/constraints.md index 7f5737a627e..b9bec56e9af 100644 --- a/gcc/config/i386/constraints.md +++ b/gcc/config/i386/constraints.md @@ -18,7 +18,7 @@ ;; <http://www.gnu.org/licenses/>. ;;; Unused letters: -;;; B H TU W +;;; B H TU W ;;; h jk vw z ;; Integer register constraints. @@ -129,7 +129,7 @@ (match_test "IN_RANGE (ival, 0, 3)"))) (define_constraint "N" - "Unsigned 8-bit integer constant (for @code{in} and @code{out} + "Unsigned 8-bit integer constant (for @code{in} and @code{out} instructions)." (and (match_code "const_int") (match_test "IN_RANGE (ival, 0, 255)"))) diff --git a/gcc/config/i386/i386.c b/gcc/config/i386/i386.c index 0d46436c239..7bd5cb6a719 100644 --- a/gcc/config/i386/i386.c +++ b/gcc/config/i386/i386.c @@ -2011,7 +2011,7 @@ override_options (void) {"pentium4", PROCESSOR_PENTIUM4, PTA_MMX |PTA_SSE | PTA_SSE2}, {"pentium4m", PROCESSOR_PENTIUM4, PTA_MMX | PTA_SSE | PTA_SSE2}, {"prescott", PROCESSOR_NOCONA, PTA_MMX | PTA_SSE | PTA_SSE2 | PTA_SSE3}, - {"nocona", PROCESSOR_NOCONA, (PTA_64BIT + {"nocona", PROCESSOR_NOCONA, (PTA_64BIT | PTA_MMX | PTA_SSE | PTA_SSE2 | PTA_SSE3 | PTA_CX16 | PTA_NO_SAHF)}, {"core2", PROCESSOR_CORE2, (PTA_64BIT @@ -2181,7 +2181,7 @@ override_options (void) ix86_arch_string = TARGET_64BIT ? "x86-64" : "i386"; else ix86_arch_specified = 1; - + if (!strcmp (ix86_arch_string, "generic")) error ("generic CPU can be used only for -mtune= switch"); if (!strncmp (ix86_arch_string, "generic", 7)) @@ -4558,7 +4558,7 @@ function_value_32 (enum machine_mode orig_mode, enum machine_mode mode, else /* Most things go in %eax. */ regno = 0; - + /* Override FP return register with %xmm0 for local functions when SSE math is enabled or for functions with sseregparm attribute. */ if ((fn || fntype) && (mode == SFmode || mode == DFmode)) @@ -4746,7 +4746,7 @@ ix86_return_in_memory (const_tree type) but differs notably in that when MMX is available, 8-byte vectors are returned in memory, rather than in MMX registers. */ -int +int ix86_sol10_return_in_memory (const_tree type) { int size; @@ -4889,7 +4889,7 @@ setup_incoming_varargs_64 (CUMULATIVE_ARGS *cum) ix86_save_varrargs_registers = 1; /* We need 16-byte stack alignment to save SSE registers. If user asked for lower preferred_stack_boundary, lets just hope that he knows - what he is doing and won't varargs SSE values. + what he is doing and won't varargs SSE values. We also may end up assuming that only 64bit values are stored in SSE register let some floating point program work. */ @@ -19794,7 +19794,7 @@ ix86_expand_multi_arg_builtin (enum insn_code icode, tree exp, rtx target, gcc_assert (GET_MODE (op) == mode || GET_MODE (op) == VOIDmode); - if (optimize + if (optimize || ! (*insn_data[icode].operand[i+adjust+1].predicate) (op, mode) || num_memory > 1) op = force_reg (mode, op); @@ -21295,7 +21295,7 @@ ix86_veclibabi_acml (enum built_in_function fn, tree type_out, tree type_in) || n != 4) return NULL_TREE; break; - + default: return NULL_TREE; } @@ -24944,7 +24944,7 @@ bool ix86_sse5_valid_op_p (rtx operands[], rtx insn, int num, bool uses_oc0, int /* format, example pmacsdd: xmm1, xmm2, xmm3/mem, xmm1 - + For the integer multiply/add instructions be more restrictive and require operands[2] and operands[3] to be the memory operands. */ else @@ -25044,7 +25044,7 @@ static const struct attribute_spec ix86_attribute_table[] = }; /* Implement targetm.vectorize.builtin_vectorization_cost. */ -static int +static int x86_builtin_vectorization_cost (bool runtime_test) { /* If the branch of the runtime test is taken - i.e. - the vectorized diff --git a/gcc/config/i386/i386.h b/gcc/config/i386/i386.h index cca89e30753..bc3e5f6dc9d 100644 --- a/gcc/config/i386/i386.h +++ b/gcc/config/i386/i386.h @@ -75,9 +75,9 @@ enum stringop_alg When size is unknown, the UNKNOWN_SIZE alg is used. When size is known at compile time or estimated via feedback, the SIZE array is walked in order until MAX is greater then the estimate (or -1 - means infinity). Corresponding ALG is used then. + means infinity). Corresponding ALG is used then. For example initializer: - {{256, loop}, {-1, rep_prefix_4_byte}} + {{256, loop}, {-1, rep_prefix_4_byte}} will use loop for blocks smaller or equal to 256 bytes, rep prefix will be used otherwise. */ struct stringop_algs @@ -153,7 +153,7 @@ struct processor_costs { scalar-to-vector operation. */ const int vec_to_scalar_cost; /* Cost of vect-to-scalar operation. */ const int scalar_to_vec_cost; /* Cost of scalar-to-vector operation. */ - const int vec_align_load_cost; /* Cost of aligned vector load. */ + const int vec_align_load_cost; /* Cost of aligned vector load. */ const int vec_unalign_load_cost; /* Cost of unaligned vector load. */ const int vec_store_cost; /* Cost of vector store. */ const int cond_taken_branch_cost; /* Cost of taken branch for vectorizer @@ -375,7 +375,7 @@ enum ix86_arch_indices { X86_ARCH_LAST }; - + extern unsigned int ix86_arch_features[X86_ARCH_LAST]; #define TARGET_CMOVE ix86_arch_features[X86_ARCH_CMOVE] @@ -2507,7 +2507,7 @@ struct machine_function GTY(()) #define TARG_SCALAR_STORE_COST ix86_cost->scalar_store_cost /* Cost of any vector operation, excluding load, store or vector to scalar - operation. */ + operation. */ #undef TARG_VEC_STMT_COST #define TARG_VEC_STMT_COST ix86_cost->vec_stmt_cost diff --git a/gcc/config/i386/i386.md b/gcc/config/i386/i386.md index cbf516175ff..89ae741d694 100644 --- a/gcc/config/i386/i386.md +++ b/gcc/config/i386/i386.md @@ -158,7 +158,7 @@ ; For SSE4A support (UNSPEC_EXTRQI 130) - (UNSPEC_EXTRQ 131) + (UNSPEC_EXTRQ 131) (UNSPEC_INSERTQI 132) (UNSPEC_INSERTQ 133) @@ -1166,7 +1166,7 @@ (set_attr "mode" "SI")]) ;; Pentium Pro can do steps 1 through 3 in one go. -;; comi*, ucomi*, fcomi*, ficomi*,fucomi* (i387 instructions set condition codes) +;; comi*, ucomi*, fcomi*, ficomi*,fucomi* (i387 instructions set condition codes) (define_insn "*cmpfp_i_mixed" [(set (reg:CCFP FLAGS_REG) (compare:CCFP (match_operand 0 "register_operand" "f,x") @@ -1476,7 +1476,7 @@ (set_attr "mode" "SI") (set_attr "pent_pair" "np") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "double")]) + (set_attr "amdfam10_decode" "double")]) (define_expand "movhi" [(set (match_operand:HI 0 "nonimmediate_operand" "") @@ -1594,7 +1594,7 @@ (set_attr "mode" "SI") (set_attr "pent_pair" "np") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "double")]) + (set_attr "amdfam10_decode" "double")]) ;; Not added amdfam10_decode since TARGET_PARTIAL_REG_STALL is disabled for AMDFAM10 (define_insn "*swaphi_2" @@ -1770,7 +1770,7 @@ (set_attr "mode" "SI") (set_attr "pent_pair" "np") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "vector")]) + (set_attr "amdfam10_decode" "vector")]) ;; Not added amdfam10_decode since TARGET_PARTIAL_REG_STALL is disabled for AMDFAM10 (define_insn "*swapqi_2" @@ -2343,7 +2343,7 @@ (set_attr "mode" "DI") (set_attr "pent_pair" "np") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "double")]) + (set_attr "amdfam10_decode" "double")]) (define_expand "movti" [(set (match_operand:TI 0 "nonimmediate_operand" "") @@ -3950,7 +3950,7 @@ [(set (match_operand:DF 0 "register_operand" "") (float_extend:DF (match_operand:SF 1 "nonimmediate_operand" "")))] - "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size + "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size && reload_completed && SSE_REG_P (operands[0])" [(set (match_dup 2) (float_extend:V2DF @@ -3978,7 +3978,7 @@ emit_insn (gen_sse_unpcklps (operands[3], operands[3], operands[3])); } else - emit_insn (gen_vec_setv4sf_0 (operands[3], + emit_insn (gen_vec_setv4sf_0 (operands[3], CONST0_RTX (V4SFmode), operands[1])); }) @@ -4086,7 +4086,7 @@ [(set (match_operand:SF 0 "register_operand" "") (float_truncate:SF (match_operand:DF 1 "nonimmediate_operand" "")))] - "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size + "(TARGET_USE_VECTOR_CONVERTS || TARGET_GENERIC) && !optimize_size && reload_completed && SSE_REG_P (operands[0])" [(set (match_dup 2) (vec_concat:V4SF @@ -4756,7 +4756,7 @@ (set_attr "mode" "HI") (set_attr "unit" "i387") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "vector")]) + (set_attr "amdfam10_decode" "vector")]) ;; Conversion between fixed point and floating point. @@ -4851,7 +4851,7 @@ (define_insn "*floatsisf2_mixed_vector" [(set (match_operand:SF 0 "register_operand" "=x,f,?f") (float:SF (match_operand:SI 1 "nonimmediate_operand" "x,m,r")))] - "TARGET_MIX_SSE_I387 && !flag_trapping_math + "TARGET_MIX_SSE_I387 && !flag_trapping_math && TARGET_USE_VECTOR_CONVERTS && !optimize_size" "@ cvtdq2ps\t{%1, %0|%0, %1} @@ -4914,7 +4914,7 @@ "#" [(set_attr "type" "multi")]) -(define_split +(define_split [(set (match_operand:SF 0 "register_operand" "") (float:SF (match_operand:SI 1 "nonimmediate_operand" "")))] "flag_trapping_math @@ -4929,7 +4929,7 @@ emit_insn (gen_sse2_loadld (operands[2], CONST0_RTX (V4SImode), operands[1])); }) -(define_split +(define_split [(set (match_operand:SF 0 "register_operand" "") (float:SF (match_operand:SI 1 "register_operand" "")))] "flag_trapping_math @@ -5042,7 +5042,7 @@ (set_attr "amdfam10_decode" "double") (set_attr "fp_int_src" "true")]) -(define_split +(define_split [(set (match_operand:DF 0 "register_operand" "") (float:DF (match_operand:SI 1 "memory_operand" "")))] "TARGET_USE_VECTOR_CONVERTS && reload_completed @@ -7483,11 +7483,11 @@ "TARGET_64BIT" "") -;; On AMDFAM10 +;; On AMDFAM10 ;; IMUL reg64, reg64, imm8 Direct ;; IMUL reg64, mem64, imm8 VectorPath ;; IMUL reg64, reg64, imm32 Direct -;; IMUL reg64, mem64, imm32 VectorPath +;; IMUL reg64, mem64, imm32 VectorPath ;; IMUL reg64, reg64 Direct ;; IMUL reg64, mem64 Direct @@ -7517,7 +7517,7 @@ (cond [(and (eq_attr "alternative" "0,1") (match_operand 1 "memory_operand" "")) (const_string "vector")] - (const_string "direct"))) + (const_string "direct"))) (set_attr "mode" "DI")]) (define_expand "mulsi3" @@ -7528,7 +7528,7 @@ "" "") -;; On AMDFAM10 +;; On AMDFAM10 ;; IMUL reg32, reg32, imm8 Direct ;; IMUL reg32, mem32, imm8 VectorPath ;; IMUL reg32, reg32, imm32 Direct @@ -7561,7 +7561,7 @@ (cond [(and (eq_attr "alternative" "0,1") (match_operand 1 "memory_operand" "")) (const_string "vector")] - (const_string "direct"))) + (const_string "direct"))) (set_attr "mode" "SI")]) (define_insn "*mulsi3_1_zext" @@ -7591,7 +7591,7 @@ (cond [(and (eq_attr "alternative" "0,1") (match_operand 1 "memory_operand" "")) (const_string "vector")] - (const_string "direct"))) + (const_string "direct"))) (set_attr "mode" "SI")]) (define_expand "mulhi3" @@ -7659,7 +7659,7 @@ (if_then_else (eq_attr "cpu" "athlon") (const_string "vector") (const_string "direct"))) - (set_attr "amdfam10_decode" "direct") + (set_attr "amdfam10_decode" "direct") (set_attr "mode" "QI")]) (define_expand "umulqihi3" @@ -7711,7 +7711,7 @@ (if_then_else (eq_attr "cpu" "athlon") (const_string "vector") (const_string "direct"))) - (set_attr "amdfam10_decode" "direct") + (set_attr "amdfam10_decode" "direct") (set_attr "mode" "QI")]) (define_expand "umulditi3" @@ -7738,7 +7738,7 @@ (if_then_else (eq_attr "cpu" "athlon") (const_string "vector") (const_string "double"))) - (set_attr "amdfam10_decode" "double") + (set_attr "amdfam10_decode" "double") (set_attr "mode" "DI")]) ;; We can't use this pattern in 64bit mode, since it results in two separate 32bit registers @@ -7766,7 +7766,7 @@ (if_then_else (eq_attr "cpu" "athlon") (const_string "vector") (const_string "double"))) - (set_attr "amdfam10_decode" "double") + (set_attr "amdfam10_decode" "double") (set_attr "mode" "SI")]) (define_expand "mulditi3" @@ -7820,7 +7820,7 @@ (if_then_else (eq_attr "cpu" "athlon") (const_string "vector") (const_string "double"))) - (set_attr "amdfam10_decode" "double") + (set_attr "amdfam10_decode" "double") (set_attr "mode" "SI")]) (define_expand "umuldi3_highpart" @@ -7857,7 +7857,7 @@ (if_then_else (eq_attr "cpu" "athlon") (const_string "vector") (const_string "double"))) - (set_attr "amdfam10_decode" "double") + (set_attr "amdfam10_decode" "double") (set_attr "mode" "DI")]) (define_expand "umulsi3_highpart" @@ -10991,7 +10991,7 @@ (set_attr "prefix_0f" "1") (set_attr "mode" "DI") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "vector")]) + (set_attr "amdfam10_decode" "vector")]) (define_expand "x86_64_shift_adj" [(set (reg:CCZ FLAGS_REG) @@ -11209,7 +11209,7 @@ (set_attr "mode" "SI") (set_attr "pent_pair" "np") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "vector")]) + (set_attr "amdfam10_decode" "vector")]) (define_expand "x86_shift_adj_1" [(set (reg:CCZ FLAGS_REG) @@ -11971,7 +11971,7 @@ (set_attr "prefix_0f" "1") (set_attr "mode" "DI") (set_attr "athlon_decode" "vector") - (set_attr "amdfam10_decode" "vector")]) + (set_attr "amdfam10_decode" "vector")]) (define_expand "ashrdi3" [(set (match_operand:DI 0 "shiftdi_operand" "") @@ -16606,7 +16606,7 @@ "fsqrt" [(set_attr "type" "fpspc") (set_attr "mode" "XF") - (set_attr "athlon_decode" "direct") + (set_attr "athlon_decode" "direct") (set_attr "amdfam10_decode" "direct")]) (define_insn "*rsqrtsf2_sse" @@ -19734,7 +19734,7 @@ ;; SSE5 conditional move (define_insn "*sse5_pcmov_<mode>" [(set (match_operand:MODEF 0 "register_operand" "=x,x,x,x") - (if_then_else:MODEF + (if_then_else:MODEF (match_operand:MODEF 1 "nonimmediate_operand" "xm,x,0,0") (match_operand:MODEF 2 "nonimmediate_operand" "0,0,x,xm") (match_operand:MODEF 3 "vector_move_operand" "x,xm,xm,x")))] diff --git a/gcc/config/i386/mmx.md b/gcc/config/i386/mmx.md index 4123bc33b22..20c4fb05b57 100644 --- a/gcc/config/i386/mmx.md +++ b/gcc/config/i386/mmx.md @@ -22,7 +22,7 @@ ;; the same register file, and 3dNOW! adds a number of extensions to ;; the base integer MMX isa. -;; Note! Except for the basic move instructions, *all* of these +;; Note! Except for the basic move instructions, *all* of these ;; patterns are outside the normal optabs namespace. This is because ;; use of these registers requires the insertion of emms or femms ;; instructions to return to normal fpu mode. The compiler doesn't @@ -301,7 +301,7 @@ "pfrsqrt\\t{%1, %0|%0, %1}" [(set_attr "type" "mmx") (set_attr "mode" "V2SF")]) - + (define_insn "mmx_rsqit1v2sf3" [(set (match_operand:V2SF 0 "register_operand" "=y") (unspec:V2SF [(match_operand:V2SF 1 "register_operand" "0") @@ -1429,4 +1429,4 @@ "TARGET_3DNOW" "femms" [(set_attr "type" "mmx") - (set_attr "memory" "none")]) + (set_attr "memory" "none")]) diff --git a/gcc/config/i386/ppro.md b/gcc/config/i386/ppro.md index f35a784f74f..b971e51089e 100644 --- a/gcc/config/i386/ppro.md +++ b/gcc/config/i386/ppro.md @@ -78,7 +78,7 @@ ;; but still in only one cycle. ;; - a complex (microcode) instruction can also only be decoded by ;; decoder 0, and this takes an unspecified number of cycles. -;; +;; ;; The goal is to schedule such that we have a few-one-one uops sequence ;; in each cycle, to decode as many instructions per cycle as possible. (define_cpu_unit "decoder0" "ppro_decoder") diff --git a/gcc/config/i386/sse.md b/gcc/config/i386/sse.md index e92ac733c99..a810ad68e5c 100644 --- a/gcc/config/i386/sse.md +++ b/gcc/config/i386/sse.md @@ -602,7 +602,7 @@ (set_attr "mode" "SF")]) ;; ??? For !flag_finite_math_only, the representation with SMIN/SMAX -;; isn't really correct, as those rtl operators aren't defined when +;; isn't really correct, as those rtl operators aren't defined when ;; applied to NaNs. Hopefully the optimizers won't get too smart on us. (define_expand "smaxv4sf3" @@ -754,7 +754,7 @@ (vec_concat:V4SF (vec_concat:V2SF (plus:SF - (vec_select:SF + (vec_select:SF (match_operand:V4SF 1 "register_operand" "0") (parallel [(const_int 0)])) (vec_select:SF (match_dup 1) (parallel [(const_int 1)]))) @@ -781,7 +781,7 @@ (vec_concat:V4SF (vec_concat:V2SF (minus:SF - (vec_select:SF + (vec_select:SF (match_operand:V4SF 1 "register_operand" "0") (parallel [(const_int 0)])) (vec_select:SF (match_dup 1) (parallel [(const_int 1)]))) @@ -2290,7 +2290,7 @@ (set_attr "mode" "DF")]) ;; ??? For !flag_finite_math_only, the representation with SMIN/SMAX -;; isn't really correct, as those rtl operators aren't defined when +;; isn't really correct, as those rtl operators aren't defined when ;; applied to NaNs. Hopefully the optimizers won't get too smart on us. (define_expand "smaxv2df3" @@ -3527,7 +3527,7 @@ emit_insn (gen_sse2_punpcklbw (t[3], operands[2], operands[2])); /* Multiply words. The end-of-line annotations here give a picture of what - the output of that instruction looks like. Dot means don't care; the + the output of that instruction looks like. Dot means don't care; the letters are the bytes of the result with A being the most significant. */ emit_insn (gen_mulv8hi3 (gen_lowpart (V8HImode, t[4]), /* .A.B.C.D.E.F.G.H */ gen_lowpart (V8HImode, t[0]), @@ -3569,8 +3569,8 @@ (define_expand "smulv8hi3_highpart" [(set (match_operand:V8HI 0 "register_operand" "") (truncate:V8HI - (lshiftrt:V8SI - (mult:V8SI + (lshiftrt:V8SI + (mult:V8SI (sign_extend:V8SI (match_operand:V8HI 1 "nonimmediate_operand" "")) (sign_extend:V8SI @@ -3578,7 +3578,7 @@ (const_int 16))))] "TARGET_SSE2" "ix86_fixup_binary_operands_no_copy (MULT, V8HImode, operands);") - + (define_insn "*smulv8hi3_highpart" [(set (match_operand:V8HI 0 "register_operand" "=x") (truncate:V8HI @@ -3776,7 +3776,7 @@ thirtytwo)); emit_insn (gen_sse2_lshrti3 (gen_lowpart (TImode, t3), gen_lowpart (TImode, op2), - thirtytwo)); + thirtytwo)); /* Multiply elements 3 and 1. */ emit_insn (gen_sse2_umulv2siv2di3 (gen_lowpart (V2DImode, t4), t2, t3)); @@ -4011,8 +4011,8 @@ }) (define_expand "udot_prodv4si" - [(match_operand:V2DI 0 "register_operand" "") - (match_operand:V4SI 1 "register_operand" "") + [(match_operand:V2DI 0 "register_operand" "") + (match_operand:V4SI 1 "register_operand" "") (match_operand:V4SI 2 "register_operand" "") (match_operand:V2DI 3 "register_operand" "")] "TARGET_SSE2" @@ -4875,7 +4875,7 @@ (const_int 9) (const_int 25) (const_int 10) (const_int 26) (const_int 11) (const_int 27) - (const_int 12) (const_int 28) + (const_int 12) (const_int 28) (const_int 13) (const_int 29) (const_int 14) (const_int 30) (const_int 15) (const_int 31)])))] @@ -5283,7 +5283,7 @@ "TARGET_SSE" "#" "&& reload_completed - && (TARGET_INTER_UNIT_MOVES + && (TARGET_INTER_UNIT_MOVES || MEM_P (operands [0]) || !GENERAL_REGNO_P (true_regnum (operands [0])))" [(set (match_dup 0) (match_dup 1))] @@ -5343,7 +5343,7 @@ (parallel [(const_int 0)])))] "TARGET_SSE && reload_completed - && (TARGET_INTER_UNIT_MOVES + && (TARGET_INTER_UNIT_MOVES || MEM_P (operands [0]) || !GENERAL_REGNO_P (true_regnum (operands [0])))" [(set (match_dup 0) (match_dup 1))] @@ -5839,7 +5839,7 @@ (set_attr "prefix_data16" "1") (set_attr "mode" "TI")]) -;; The correct representation for this is absolutely enormous, and +;; The correct representation for this is absolutely enormous, and ;; surely not generally useful. (define_insn "sse2_psadbw" [(set (match_operand:V2DI 0 "register_operand" "=x") @@ -6641,7 +6641,7 @@ (define_insn "sse4a_vmmovntv2df" [(set (match_operand:DF 0 "memory_operand" "=m") - (unspec:DF [(vec_select:DF + (unspec:DF [(vec_select:DF (match_operand:V2DF 1 "register_operand" "x") (parallel [(const_int 0)]))] UNSPEC_MOVNT))] @@ -6661,7 +6661,7 @@ (define_insn "sse4a_vmmovntv4sf" [(set (match_operand:SF 0 "memory_operand" "=m") - (unspec:SF [(vec_select:SF + (unspec:SF [(vec_select:SF (match_operand:V4SF 1 "register_operand" "x") (parallel [(const_int 0)]))] UNSPEC_MOVNT))] @@ -7893,7 +7893,7 @@ ;; SSE5 parallel XMM conditional moves (define_insn "sse5_pcmov_<mode>" [(set (match_operand:SSEMODE 0 "register_operand" "=x,x,x,x,x,x") - (if_then_else:SSEMODE + (if_then_else:SSEMODE (match_operand:SSEMODE 3 "nonimmediate_operand" "0,0,xm,xm,0,0") (match_operand:SSEMODE 1 "vector_move_operand" "x,xm,0,x,C,x") (match_operand:SSEMODE 2 "vector_move_operand" "xm,x,x,0,x,C")))] diff --git a/gcc/config/i386/sync.md b/gcc/config/i386/sync.md index 19295bbaff1..a4f6d38456c 100644 --- a/gcc/config/i386/sync.md +++ b/gcc/config/i386/sync.md @@ -34,7 +34,7 @@ ;; ??? It would be possible to use cmpxchg8b on pentium for DImode ;; changes. It's complicated because the insn uses ecx:ebx as the ;; new value; note that the registers are reversed from the order -;; that they'd be in with (reg:DI 2 ecx). Similarly for TImode +;; that they'd be in with (reg:DI 2 ecx). Similarly for TImode ;; data in 64-bit mode. (define_expand "sync_compare_and_swap<mode>" |