Author: kromanova
Date: Fri May 27 19:18:59 2016
New Revision: 271077

URL: http://llvm.org/viewvc/llvm-project?rev=271077&view=rev
Log:
Clean up: remove trailing spaces in x86 intrinsic headers.

Differential Revision: http://reviews.llvm.org/D20614


Modified:
    cfe/trunk/lib/Headers/__wmmintrin_aes.h
    cfe/trunk/lib/Headers/__wmmintrin_pclmul.h
    cfe/trunk/lib/Headers/avx512fintrin.h
    cfe/trunk/lib/Headers/avx512vldqintrin.h
    cfe/trunk/lib/Headers/mwaitxintrin.h
    cfe/trunk/lib/Headers/pmmintrin.h

Modified: cfe/trunk/lib/Headers/__wmmintrin_aes.h
URL: 
http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/__wmmintrin_aes.h?rev=271077&r1=271076&r2=271077&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/__wmmintrin_aes.h (original)
+++ cfe/trunk/lib/Headers/__wmmintrin_aes.h Fri May 27 19:18:59 2016
@@ -125,9 +125,9 @@ _mm_aesimc_si128(__m128i __V)
   return (__m128i)__builtin_ia32_aesimc128((__v2di)__V);
 }
 
-/// \brief Generates a round key for AES encyption, operating on 128-bit data 
+/// \brief Generates a round key for AES encyption, operating on 128-bit data
 ///    specified in the first source operand and using an 8-bit round constant
-///    specified by the second source operand, and writes the result to the 
+///    specified by the second source operand, and writes the result to the
 ///    destination.
 ///
 /// \headerfile <x86intrin.h>

Modified: cfe/trunk/lib/Headers/__wmmintrin_pclmul.h
URL: 
http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/__wmmintrin_pclmul.h?rev=271077&r1=271076&r2=271077&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/__wmmintrin_pclmul.h (original)
+++ cfe/trunk/lib/Headers/__wmmintrin_pclmul.h Fri May 27 19:18:59 2016
@@ -24,15 +24,15 @@
 #define _WMMINTRIN_PCLMUL_H
 
 /// \brief Multiplies two 64-bit integer values, which are selected from source
-///    operands using the immediate-value operand. The multiplication is a 
+///    operands using the immediate-value operand. The multiplication is a
 ///    carry-less multiplication, and the 128-bit integer product is stored in
 ///    the destination.
 ///
 /// \headerfile <x86intrin.h>
 ///
-/// \code 
+/// \code
 /// __m128i _mm_clmulepi64_si128(__m128i __X, __m128i __Y, const int __I);
-/// \endcode 
+/// \endcode
 ///
 /// This intrinsic corresponds to the \c VPCLMULQDQ instruction.
 ///

Modified: cfe/trunk/lib/Headers/avx512fintrin.h
URL: 
http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/avx512fintrin.h?rev=271077&r1=271076&r2=271077&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/avx512fintrin.h (original)
+++ cfe/trunk/lib/Headers/avx512fintrin.h Fri May 27 19:18:59 2016
@@ -949,7 +949,7 @@ _mm_maskz_max_sd(__mmask8 __U,__m128d __
                                            (__v2df)_mm_setzero_pd(), \
                                            (__mmask8)(U), (int)(R)); })
 
-static __inline __m512i 
+static __inline __m512i
 __DEFAULT_FN_ATTRS
 _mm512_max_epi32(__m512i __A, __m512i __B)
 {
@@ -1508,7 +1508,7 @@ _mm_rsqrt14_ss(__m128 __A, __m128 __B)
 }
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_rsqrt14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_rsqrt14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_rsqrt14ss_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -1517,7 +1517,7 @@ _mm_mask_rsqrt14_ss (__m128 __W, __mmask
 }
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_rsqrt14_ss (__mmask8 __U, __m128 __A, __m128 __B) 
+_mm_maskz_rsqrt14_ss (__mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_rsqrt14ss_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -1536,7 +1536,7 @@ _mm_rsqrt14_sd(__m128d __A, __m128d __B)
 }
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_rsqrt14_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_rsqrt14_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_rsqrt14sd_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -1545,7 +1545,7 @@ _mm_mask_rsqrt14_sd (__m128d __W, __mmas
 }
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_rsqrt14_sd (__mmask8 __U, __m128d __A, __m128d __B) 
+_mm_maskz_rsqrt14_sd (__mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_rsqrt14sd_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -1616,7 +1616,7 @@ _mm_rcp14_ss(__m128 __A, __m128 __B)
 }
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_rcp14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_rcp14_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_rcp14ss_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -1625,7 +1625,7 @@ _mm_mask_rcp14_ss (__m128 __W, __mmask8
 }
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_rcp14_ss (__mmask8 __U, __m128 __A, __m128 __B) 
+_mm_maskz_rcp14_ss (__mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_rcp14ss_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -1644,7 +1644,7 @@ _mm_rcp14_sd(__m128d __A, __m128d __B)
 }
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_rcp14_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_rcp14_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_rcp14sd_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -1653,7 +1653,7 @@ _mm_mask_rcp14_sd (__m128d __W, __mmask8
 }
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_rcp14_sd (__mmask8 __U, __m128d __A, __m128d __B) 
+_mm_maskz_rcp14_sd (__mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_rcp14sd_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -3306,7 +3306,7 @@ _mm512_maskz_permutex2var_epi64 (__mmask
                                          (__v16si)(__m512i)(B), (int)(I), \
                                          (__v16si)_mm512_setzero_si512(), \
                                          (__mmask16)-1); })
-                                         
+
 #define _mm512_mask_alignr_epi32(W, U, A, B, imm) __extension__ ({\
   (__m512i)__builtin_ia32_alignd512_mask((__v16si)(__m512i)(A), \
                                          (__v16si)(__m512i)(B), (int)(imm), \
@@ -3488,7 +3488,7 @@ static __inline __m512d __DEFAULT_FN_ATT
 _mm512_cvtepi32_pd(__m256i __A)
 {
   return (__m512d) __builtin_ia32_cvtdq2pd512_mask ((__v8si) __A,
-                (__v8df) 
+                (__v8df)
                 _mm512_setzero_pd (),
                 (__mmask8) -1);
 }
@@ -3613,7 +3613,7 @@ static  __inline __m512 __DEFAULT_FN_ATT
 _mm512_cvtph_ps(__m256i __A)
 {
   return (__m512) __builtin_ia32_vcvtph2ps512_mask ((__v16hi) __A,
-                (__v16sf) 
+                (__v16sf)
                 _mm512_setzero_ps (),
                 (__mmask16) -1,
                 _MM_FROUND_CUR_DIRECTION);
@@ -3733,7 +3733,7 @@ _mm512_maskz_cvtps_epi32 (__mmask16 __U,
                  (__mmask16) __U,
                  _MM_FROUND_CUR_DIRECTION);
 }
-                                           
+
 #define _mm512_cvt_roundpd_epi32(A, R) __extension__ ({ \
   (__m256i)__builtin_ia32_cvtpd2dq512_mask((__v8df)(__m512d)(A), \
                                            (__v8si)_mm256_setzero_si256(), \
@@ -3774,7 +3774,7 @@ _mm512_maskz_cvtpd_epi32 (__mmask8 __U,
                                             (__mmask16)-1, (int)(R)); })
 
 static __inline__ __m512i __DEFAULT_FN_ATTRS
-_mm512_cvtps_epu32 ( __m512 __A) 
+_mm512_cvtps_epu32 ( __m512 __A)
 {
   return (__m512i) __builtin_ia32_cvtps2udq512_mask ((__v16sf) __A,\
                   (__v16si)\
@@ -5267,7 +5267,7 @@ _mm_getexp_sd (__m128d __A, __m128d __B)
 }
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_getexp_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_getexp_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_getexpsd128_round_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -5283,7 +5283,7 @@ _mm_mask_getexp_sd (__m128d __W, __mmask
                                                  (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_getexp_sd (__mmask8 __U, __m128d __A, __m128d __B) 
+_mm_maskz_getexp_sd (__mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_getexpsd128_round_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -5312,7 +5312,7 @@ _mm_getexp_ss (__m128 __A, __m128 __B)
 }
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_getexp_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_getexp_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_getexpss128_round_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -5328,7 +5328,7 @@ _mm_mask_getexp_ss (__m128 __W, __mmask8
                                                 (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_getexp_ss (__mmask8 __U, __m128 __A, __m128 __B) 
+_mm_maskz_getexp_ss (__mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_getexpss128_round_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -6654,7 +6654,7 @@ _mm_scalef_sd (__m128d __A, __m128d __B)
 }
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_scalef_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_scalef_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_scalefsd_round_mask ( (__v2df) __A,
                  (__v2df) __B,
@@ -6670,7 +6670,7 @@ _mm_mask_scalef_sd (__m128d __W, __mmask
                                               (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_scalef_sd (__mmask8 __U, __m128d __A, __m128d __B) 
+_mm_maskz_scalef_sd (__mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_scalefsd_round_mask ( (__v2df) __A,
                  (__v2df) __B,
@@ -6701,7 +6701,7 @@ _mm_scalef_ss (__m128 __A, __m128 __B)
 }
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_scalef_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_scalef_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_scalefss_round_mask ( (__v4sf) __A,
                 (__v4sf) __B,
@@ -6717,7 +6717,7 @@ _mm_mask_scalef_ss (__m128 __W, __mmask8
                                              (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_scalef_ss (__mmask8 __U, __m128 __A, __m128 __B) 
+_mm_maskz_scalef_ss (__mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_scalefss_round_mask ( (__v4sf) __A,
                  (__v4sf) __B,
@@ -6878,7 +6878,7 @@ _mm_maskz_scalef_ss (__mmask8 __U, __m12
                                             (__mmask8)-1, (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_sqrt_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_sqrt_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_sqrtsd_round_mask ( (__v2df) __B,
                  (__v2df) __A,
@@ -6894,7 +6894,7 @@ _mm_mask_sqrt_sd (__m128d __W, __mmask8
                                             (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_sqrt_sd (__mmask8 __U, __m128d __A, __m128d __B) 
+_mm_maskz_sqrt_sd (__mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_sqrtsd_round_mask ( (__v2df) __B,
                  (__v2df) __A,
@@ -6916,7 +6916,7 @@ _mm_maskz_sqrt_sd (__mmask8 __U, __m128d
                                            (__mmask8)-1, (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_sqrt_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_sqrt_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_sqrtss_round_mask ( (__v4sf) __B,
                  (__v4sf) __A,
@@ -6932,7 +6932,7 @@ _mm_mask_sqrt_ss (__m128 __W, __mmask8 _
                                            (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_sqrt_ss (__mmask8 __U, __m128 __A, __m128 __B) 
+_mm_maskz_sqrt_ss (__mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_sqrtss_round_mask ( (__v4sf) __A,
                  (__v4sf) __B,
@@ -7963,7 +7963,7 @@ __builtin_ia32_gatherdiv16sf ((__v8sf) _
                                (__v8di)(__m512i)(v1), (int)(scale)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_fmadd_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_fmadd_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask ((__v4sf) __A,
           (__v4sf) __B,
@@ -7979,7 +7979,7 @@ _mm_mask_fmadd_ss (__m128 __W, __mmask8
                                         (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_fmadd_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) 
+_mm_maskz_fmadd_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C)
 {
  return (__m128) __builtin_ia32_vfmaddss3_maskz ((__v4sf) __A,
           (__v4sf) __B,
@@ -7995,7 +7995,7 @@ _mm_maskz_fmadd_ss (__mmask8 __U, __m128
                                          _MM_FROUND_CUR_DIRECTION); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask3_fmadd_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U) 
+_mm_mask3_fmadd_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask3 ((__v4sf) __W,
           (__v4sf) __X,
@@ -8011,7 +8011,7 @@ _mm_mask3_fmadd_ss (__m128 __W, __m128 _
                                          (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_fmsub_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_fmsub_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask ((__v4sf) __A,
           -(__v4sf) __B,
@@ -8027,7 +8027,7 @@ _mm_mask_fmsub_ss (__m128 __W, __mmask8
                                         (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_fmsub_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) 
+_mm_maskz_fmsub_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C)
 {
  return (__m128) __builtin_ia32_vfmaddss3_maskz ((__v4sf) __A,
           (__v4sf) __B,
@@ -8043,7 +8043,7 @@ _mm_maskz_fmsub_ss (__mmask8 __U, __m128
                                          (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask3_fmsub_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U) 
+_mm_mask3_fmsub_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask3 ((__v4sf) __W,
           (__v4sf) __X,
@@ -8059,7 +8059,7 @@ _mm_mask3_fmsub_ss (__m128 __W, __m128 _
                                          (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_fnmadd_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_fnmadd_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask (-(__v4sf) __A,
           (__v4sf) __B,
@@ -8075,7 +8075,7 @@ _mm_mask_fnmadd_ss (__m128 __W, __mmask8
                                         (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_fnmadd_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) 
+_mm_maskz_fnmadd_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C)
 {
  return (__m128) __builtin_ia32_vfmaddss3_maskz (-(__v4sf) __A,
           (__v4sf) __B,
@@ -8091,7 +8091,7 @@ _mm_maskz_fnmadd_ss (__mmask8 __U, __m12
                                          (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask3_fnmadd_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U) 
+_mm_mask3_fnmadd_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask3 (-(__v4sf) __W,
           (__v4sf) __X,
@@ -8107,7 +8107,7 @@ _mm_mask3_fnmadd_ss (__m128 __W, __m128
                                          (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask_fnmsub_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) 
+_mm_mask_fnmsub_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask (-(__v4sf) __A,
           -(__v4sf) __B,
@@ -8123,7 +8123,7 @@ _mm_mask_fnmsub_ss (__m128 __W, __mmask8
                                         (int)(R)); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_maskz_fnmsub_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C) 
+_mm_maskz_fnmsub_ss (__mmask8 __U, __m128 __A, __m128 __B, __m128 __C)
 {
  return (__m128) __builtin_ia32_vfmaddss3_maskz (-(__v4sf) __A,
           (__v4sf) __B,
@@ -8139,7 +8139,7 @@ _mm_maskz_fnmsub_ss (__mmask8 __U, __m12
                                          _MM_FROUND_CUR_DIRECTION); })
 
 static __inline__ __m128 __DEFAULT_FN_ATTRS
-_mm_mask3_fnmsub_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U) 
+_mm_mask3_fnmsub_ss (__m128 __W, __m128 __X, __m128 __Y, __mmask8 __U)
 {
  return (__m128) __builtin_ia32_vfmaddss3_mask3 (-(__v4sf) __W,
           (__v4sf) __X,
@@ -8155,7 +8155,7 @@ _mm_mask3_fnmsub_ss (__m128 __W, __m128
                                          (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_fmadd_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_fmadd_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask ( (__v2df) __A,
           (__v2df) __B,
@@ -8171,7 +8171,7 @@ _mm_mask_fmadd_sd (__m128d __W, __mmask8
                                          (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_fmadd_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C) 
+_mm_maskz_fmadd_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_maskz ( (__v2df) __A,
           (__v2df) __B,
@@ -8187,7 +8187,7 @@ _mm_maskz_fmadd_sd (__mmask8 __U, __m128
                                           _MM_FROUND_CUR_DIRECTION); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask3_fmadd_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U) 
+_mm_mask3_fmadd_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask3 ((__v2df) __W,
           (__v2df) __X,
@@ -8203,7 +8203,7 @@ _mm_mask3_fmadd_sd (__m128d __W, __m128d
                                           (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_fmsub_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_fmsub_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask ( (__v2df) __A,
           -(__v2df) __B,
@@ -8219,7 +8219,7 @@ _mm_mask_fmsub_sd (__m128d __W, __mmask8
                                          (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_fmsub_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C) 
+_mm_maskz_fmsub_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_maskz ( (__v2df) __A,
           (__v2df) __B,
@@ -8235,7 +8235,7 @@ _mm_maskz_fmsub_sd (__mmask8 __U, __m128
                                           (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask3_fmsub_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U) 
+_mm_mask3_fmsub_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask3 ((__v2df) __W,
           (__v2df) __X,
@@ -8251,7 +8251,7 @@ _mm_mask3_fmsub_sd (__m128d __W, __m128d
                                           (__mmask8)(U), (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_fnmadd_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_fnmadd_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask ( -(__v2df) __A,
           (__v2df) __B,
@@ -8267,7 +8267,7 @@ _mm_mask_fnmadd_sd (__m128d __W, __mmask
                                          (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_fnmadd_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C) 
+_mm_maskz_fnmadd_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_maskz ( -(__v2df) __A,
           (__v2df) __B,
@@ -8283,7 +8283,7 @@ _mm_maskz_fnmadd_sd (__mmask8 __U, __m12
                                           (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask3_fnmadd_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U) 
+_mm_mask3_fnmadd_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask3 (-(__v2df) __W,
           (__v2df) __X,
@@ -8299,7 +8299,7 @@ _mm_mask3_fnmadd_sd (__m128d __W, __m128
                                           (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask_fnmsub_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) 
+_mm_mask_fnmsub_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask ( -(__v2df) __A,
           -(__v2df) __B,
@@ -8315,7 +8315,7 @@ _mm_mask_fnmsub_sd (__m128d __W, __mmask
                                          (int)(R)); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_maskz_fnmsub_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C) 
+_mm_maskz_fnmsub_sd (__mmask8 __U, __m128d __A, __m128d __B, __m128d __C)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_maskz ( -(__v2df) __A,
           (__v2df) __B,
@@ -8332,7 +8332,7 @@ _mm_maskz_fnmsub_sd (__mmask8 __U, __m12
                                           _MM_FROUND_CUR_DIRECTION); })
 
 static __inline__ __m128d __DEFAULT_FN_ATTRS
-_mm_mask3_fnmsub_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U) 
+_mm_mask3_fnmsub_sd (__m128d __W, __m128d __X, __m128d __Y, __mmask8 __U)
 {
  return (__m128d) __builtin_ia32_vfmaddsd3_mask3 (-(__v2df) (__W),
           (__v2df) __X,
@@ -9097,7 +9097,7 @@ _mm512_set_epi32 (int __A, int __B, int
 #define _mm512_setr_epi32(e0,e1,e2,e3,e4,e5,e6,e7,           \
        e8,e9,e10,e11,e12,e13,e14,e15)          \
   _mm512_set_epi32(e15,e14,e13,e12,e11,e10,e9,e8,e7,e6,e5,e4,e3,e2,e1,e0)
-  
+
 static __inline__ __m512i __DEFAULT_FN_ATTRS
 _mm512_set_epi64 (long long __A, long long __B, long long __C,
      long long __D, long long __E, long long __F,

Modified: cfe/trunk/lib/Headers/avx512vldqintrin.h
URL: 
http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/avx512vldqintrin.h?rev=271077&r1=271076&r2=271077&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/avx512vldqintrin.h (original)
+++ cfe/trunk/lib/Headers/avx512vldqintrin.h Fri May 27 19:18:59 2016
@@ -1066,7 +1066,7 @@ static __inline__ __m256d __DEFAULT_FN_A
 _mm256_mask_broadcast_f64x2 (__m256d __O, __mmask8 __M, __m128d __A)
 {
   return (__m256d) __builtin_ia32_broadcastf64x2_256_mask ((__v2df) __A,
-                 (__v4df) __O, 
+                 (__v4df) __O,
                  __M);
 }
 
@@ -1138,7 +1138,7 @@ static __inline__ __m256i __DEFAULT_FN_A
 _mm256_mask_broadcast_i64x2 (__m256i __O, __mmask8 __M, __m128i __A)
 {
   return (__m256i) __builtin_ia32_broadcasti64x2_256_mask ((__v2di) __A,
-                 (__v4di) __O, 
+                 (__v4di) __O,
                  __M);
 }
 

Modified: cfe/trunk/lib/Headers/mwaitxintrin.h
URL: 
http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/mwaitxintrin.h?rev=271077&r1=271076&r2=271077&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/mwaitxintrin.h (original)
+++ cfe/trunk/lib/Headers/mwaitxintrin.h Fri May 27 19:18:59 2016
@@ -42,6 +42,6 @@ _mm_mwaitx(unsigned __extensions, unsign
   __builtin_ia32_mwaitx(__extensions, __hints, __clock);
 }
 
-#undef __DEFAULT_FN_ATTRS 
+#undef __DEFAULT_FN_ATTRS
 
 #endif /* _MWAITXINTRIN_H */

Modified: cfe/trunk/lib/Headers/pmmintrin.h
URL: 
http://llvm.org/viewvc/llvm-project/cfe/trunk/lib/Headers/pmmintrin.h?rev=271077&r1=271076&r2=271077&view=diff
==============================================================================
--- cfe/trunk/lib/Headers/pmmintrin.h (original)
+++ cfe/trunk/lib/Headers/pmmintrin.h Fri May 27 19:18:59 2016
@@ -31,9 +31,9 @@
   __attribute__((__always_inline__, __nodebug__, __target__("sse3")))
 
 /// \brief Loads data from an unaligned memory location to elements in a 
128-bit
-///    vector. If the address of the data is not 16-byte aligned, the 
-///    instruction may read two adjacent aligned blocks of memory to retrieve 
-///    the requested data. 
+///    vector. If the address of the data is not 16-byte aligned, the
+///    instruction may read two adjacent aligned blocks of memory to retrieve
+///    the requested data.
 ///
 /// \headerfile <x86intrin.h>
 ///
@@ -75,14 +75,14 @@ _mm_addsub_ps(__m128 __a, __m128 __b)
 /// This intrinsic corresponds to the \c VHADDPS instruction.
 ///
 /// \param __a
-///    A 128-bit vector of [4 x float] containing one of the source operands. 
-///    The horizontal sums of the values are stored in the lower bits of the 
+///    A 128-bit vector of [4 x float] containing one of the source operands.
+///    The horizontal sums of the values are stored in the lower bits of the
 ///    destination.
 /// \param __b
-///    A 128-bit vector of [4 x float] containing one of the source operands. 
-///    The horizontal sums of the values are stored in the upper bits of the 
+///    A 128-bit vector of [4 x float] containing one of the source operands.
+///    The horizontal sums of the values are stored in the upper bits of the
 ///    destination.
-/// \returns A 128-bit vector of [4 x float] containing the horizontal sums of 
+/// \returns A 128-bit vector of [4 x float] containing the horizontal sums of
 ///    both operands.
 static __inline__ __m128 __DEFAULT_FN_ATTRS
 _mm_hadd_ps(__m128 __a, __m128 __b)
@@ -98,14 +98,14 @@ _mm_hadd_ps(__m128 __a, __m128 __b)
 /// This intrinsic corresponds to the \c VHSUBPS instruction.
 ///
 /// \param __a
-///    A 128-bit vector of [4 x float] containing one of the source operands. 
-///    The horizontal differences between the values are stored in the lower 
+///    A 128-bit vector of [4 x float] containing one of the source operands.
+///    The horizontal differences between the values are stored in the lower
 ///    bits of the destination.
 /// \param __b
-///    A 128-bit vector of [4 x float] containing one of the source operands. 
-///    The horizontal differences between the values are stored in the upper 
+///    A 128-bit vector of [4 x float] containing one of the source operands.
+///    The horizontal differences between the values are stored in the upper
 ///    bits of the destination.
-/// \returns A 128-bit vector of [4 x float] containing the horizontal 
+/// \returns A 128-bit vector of [4 x float] containing the horizontal
 ///    differences of both operands.
 static __inline__ __m128 __DEFAULT_FN_ATTRS
 _mm_hsub_ps(__m128 __a, __m128 __b)
@@ -168,7 +168,7 @@ _mm_moveldup_ps(__m128 __a)
 ///    A 128-bit vector of [2 x double] containing the left source operand.
 /// \param __b
 ///    A 128-bit vector of [2 x double] containing the right source operand.
-/// \returns A 128-bit vector of [2 x double] containing the alternating sums 
+/// \returns A 128-bit vector of [2 x double] containing the alternating sums
 ///    and differences of both operands.
 static __inline__ __m128d __DEFAULT_FN_ATTRS
 _mm_addsub_pd(__m128d __a, __m128d __b)
@@ -176,7 +176,7 @@ _mm_addsub_pd(__m128d __a, __m128d __b)
   return __builtin_ia32_addsubpd((__v2df)__a, (__v2df)__b);
 }
 
-/// \brief Horizontally adds the pairs of values contained in two 128-bit 
+/// \brief Horizontally adds the pairs of values contained in two 128-bit
 ///    vectors of [2 x double].
 ///
 /// \headerfile <x86intrin.h>
@@ -184,12 +184,12 @@ _mm_addsub_pd(__m128d __a, __m128d __b)
 /// This intrinsic corresponds to the \c VHADDPD instruction.
 ///
 /// \param __a
-///    A 128-bit vector of [2 x double] containing one of the source operands. 
-///    The horizontal sum of the values is stored in the lower bits of the 
+///    A 128-bit vector of [2 x double] containing one of the source operands.
+///    The horizontal sum of the values is stored in the lower bits of the
 ///    destination.
 /// \param __b
-///    A 128-bit vector of [2 x double] containing one of the source operands. 
-///    The horizontal sum of the values is stored in the upper bits of the 
+///    A 128-bit vector of [2 x double] containing one of the source operands.
+///    The horizontal sum of the values is stored in the upper bits of the
 ///    destination.
 /// \returns A 128-bit vector of [2 x double] containing the horizontal sums of
 ///    both operands.
@@ -207,14 +207,14 @@ _mm_hadd_pd(__m128d __a, __m128d __b)
 /// This intrinsic corresponds to the \c VHSUBPD instruction.
 ///
 /// \param __a
-///    A 128-bit vector of [2 x double] containing one of the source operands. 
+///    A 128-bit vector of [2 x double] containing one of the source operands.
 ///    The horizontal difference of the values is stored in the lower bits of
 ///    the destination.
 /// \param __b
-///    A 128-bit vector of [2 x double] containing one of the source operands. 
+///    A 128-bit vector of [2 x double] containing one of the source operands.
 ///    The horizontal difference of the values is stored in the upper bits of
 ///    the destination.
-/// \returns A 128-bit vector of [2 x double] containing the horizontal 
+/// \returns A 128-bit vector of [2 x double] containing the horizontal
 ///    differences of both operands.
 static __inline__ __m128d __DEFAULT_FN_ATTRS
 _mm_hsub_pd(__m128d __a, __m128d __b)
@@ -227,19 +227,19 @@ _mm_hsub_pd(__m128d __a, __m128d __b)
 ///
 /// \headerfile <x86intrin.h>
 ///
-/// \code 
+/// \code
 /// __m128d _mm_loaddup_pd(double const * dp);
-/// \endcode 
+/// \endcode
 ///
 /// This intrinsic corresponds to the \c VMOVDDUP instruction.
 ///
 /// \param dp
 ///    A pointer to a double-precision value to be moved and duplicated.
-/// \returns A 128-bit vector of [2 x double] containing the moved and 
+/// \returns A 128-bit vector of [2 x double] containing the moved and
 ///    duplicated values.
 #define        _mm_loaddup_pd(dp)        _mm_load1_pd(dp)
 
-/// \brief Moves and duplicates the double-precision value in the lower bits 
of 
+/// \brief Moves and duplicates the double-precision value in the lower bits of
 ///    a 128-bit vector of [2 x double] to double-precision values stored in a
 ///    128-bit vector of [2 x double].
 ///
@@ -248,9 +248,9 @@ _mm_hsub_pd(__m128d __a, __m128d __b)
 /// This intrinsic corresponds to the \c VMOVDDUP instruction.
 ///
 /// \param __a
-///    A 128-bit vector of [2 x double]. Bits [63:0] are written to bits 
+///    A 128-bit vector of [2 x double]. Bits [63:0] are written to bits
 ///    [127:64] and [63:0] of the destination.
-/// \returns A 128-bit vector of [2 x double] containing the moved and 
+/// \returns A 128-bit vector of [2 x double] containing the moved and
 ///    duplicated values.
 static __inline__ __m128d __DEFAULT_FN_ATTRS
 _mm_movedup_pd(__m128d __a)
@@ -266,8 +266,8 @@ _mm_movedup_pd(__m128d __a)
 #define _MM_GET_DENORMALS_ZERO_MODE() (_mm_getcsr() & _MM_DENORMALS_ZERO_MASK)
 #define _MM_SET_DENORMALS_ZERO_MODE(x) (_mm_setcsr((_mm_getcsr() & 
~_MM_DENORMALS_ZERO_MASK) | (x)))
 
-/// \brief Establishes a linear address memory range to be monitored and puts 
-///    the processor in the monitor event pending state. Data stored in the 
+/// \brief Establishes a linear address memory range to be monitored and puts
+///    the processor in the monitor event pending state. Data stored in the
 ///    monitored address range causes the processor to exit the pending state.
 ///
 /// \headerfile <x86intrin.h>
@@ -275,7 +275,7 @@ _mm_movedup_pd(__m128d __a)
 /// This intrinsic corresponds to the \c MONITOR instruction.
 ///
 /// \param __p
-///    The memory range to be monitored. The size of the range is determined 
by 
+///    The memory range to be monitored. The size of the range is determined by
 ///    CPUID function 0000_0005h.
 /// \param __extensions
 ///    Optional extensions for the monitoring state.
@@ -296,7 +296,7 @@ _mm_monitor(void const *__p, unsigned __
 /// This intrinsic corresponds to the \c MWAIT instruction.
 ///
 /// \param __extensions
-///    Optional extensions for the monitoring state, which may vary by 
+///    Optional extensions for the monitoring state, which may vary by
 ///    processor.
 /// \param __hints
 ///    Optional hints for the monitoring state, which may vary by processor.


_______________________________________________
cfe-commits mailing list
cfe-commits@lists.llvm.org
http://lists.llvm.org/cgi-bin/mailman/listinfo/cfe-commits

Reply via email to