1  /* { dg-final { check-function-bodies "**" "" "-DCHECK_ASM" } } */
       2  
       3  #include "test_sve_acle.h"
       4  
       5  /*
       6  ** qadd_s16_tied1:
       7  **	sqadd	z0\.h, (z0\.h, z1\.h|z1\.h, z0\.h)
       8  **	ret
       9  */
      10  TEST_UNIFORM_Z (qadd_s16_tied1, svint16_t,
      11  		z0 = svqadd_s16 (z0, z1),
      12  		z0 = svqadd (z0, z1))
      13  
      14  /*
      15  ** qadd_s16_tied2:
      16  **	sqadd	z0\.h, (z0\.h, z1\.h|z1\.h, z0\.h)
      17  **	ret
      18  */
      19  TEST_UNIFORM_Z (qadd_s16_tied2, svint16_t,
      20  		z0 = svqadd_s16 (z1, z0),
      21  		z0 = svqadd (z1, z0))
      22  
      23  /*
      24  ** qadd_s16_untied:
      25  **	sqadd	z0\.h, (z1\.h, z2\.h|z2\.h, z1\.h)
      26  **	ret
      27  */
      28  TEST_UNIFORM_Z (qadd_s16_untied, svint16_t,
      29  		z0 = svqadd_s16 (z1, z2),
      30  		z0 = svqadd (z1, z2))
      31  
      32  /*
      33  ** qadd_w0_s16_tied1:
      34  **	mov	(z[0-9]+\.h), w0
      35  **	sqadd	z0\.h, (z0\.h, \1|\1, z0\.h)
      36  **	ret
      37  */
      38  TEST_UNIFORM_ZX (qadd_w0_s16_tied1, svint16_t, int16_t,
      39  		 z0 = svqadd_n_s16 (z0, x0),
      40  		 z0 = svqadd (z0, x0))
      41  
      42  /*
      43  ** qadd_w0_s16_untied:
      44  **	mov	(z[0-9]+\.h), w0
      45  **	sqadd	z0\.h, (z1\.h, \1|\1, z1\.h)
      46  **	ret
      47  */
      48  TEST_UNIFORM_ZX (qadd_w0_s16_untied, svint16_t, int16_t,
      49  		 z0 = svqadd_n_s16 (z1, x0),
      50  		 z0 = svqadd (z1, x0))
      51  
      52  /*
      53  ** qadd_1_s16_tied1:
      54  **	sqadd	z0\.h, z0\.h, #1
      55  **	ret
      56  */
      57  TEST_UNIFORM_Z (qadd_1_s16_tied1, svint16_t,
      58  		z0 = svqadd_n_s16 (z0, 1),
      59  		z0 = svqadd (z0, 1))
      60  
      61  /*
      62  ** qadd_1_s16_untied:
      63  **	movprfx	z0, z1
      64  **	sqadd	z0\.h, z0\.h, #1
      65  **	ret
      66  */
      67  TEST_UNIFORM_Z (qadd_1_s16_untied, svint16_t,
      68  		z0 = svqadd_n_s16 (z1, 1),
      69  		z0 = svqadd (z1, 1))
      70  
      71  /*
      72  ** qadd_127_s16:
      73  **	sqadd	z0\.h, z0\.h, #127
      74  **	ret
      75  */
      76  TEST_UNIFORM_Z (qadd_127_s16, svint16_t,
      77  		z0 = svqadd_n_s16 (z0, 127),
      78  		z0 = svqadd (z0, 127))
      79  
      80  /*
      81  ** qadd_128_s16:
      82  **	sqadd	z0\.h, z0\.h, #128
      83  **	ret
      84  */
      85  TEST_UNIFORM_Z (qadd_128_s16, svint16_t,
      86  		z0 = svqadd_n_s16 (z0, 128),
      87  		z0 = svqadd (z0, 128))
      88  
      89  /*
      90  ** qadd_255_s16:
      91  **	sqadd	z0\.h, z0\.h, #255
      92  **	ret
      93  */
      94  TEST_UNIFORM_Z (qadd_255_s16, svint16_t,
      95  		z0 = svqadd_n_s16 (z0, 255),
      96  		z0 = svqadd (z0, 255))
      97  
      98  /*
      99  ** qadd_m1_s16:
     100  **	sqsub	z0\.h, z0\.h, #1
     101  **	ret
     102  */
     103  TEST_UNIFORM_Z (qadd_m1_s16, svint16_t,
     104  		z0 = svqadd_n_s16 (z0, -1),
     105  		z0 = svqadd (z0, -1))
     106  
     107  /*
     108  ** qadd_m127_s16:
     109  **	sqsub	z0\.h, z0\.h, #127
     110  **	ret
     111  */
     112  TEST_UNIFORM_Z (qadd_m127_s16, svint16_t,
     113  		z0 = svqadd_n_s16 (z0, -127),
     114  		z0 = svqadd (z0, -127))
     115  
     116  /*
     117  ** qadd_m128_s16:
     118  **	sqsub	z0\.h, z0\.h, #128
     119  **	ret
     120  */
     121  TEST_UNIFORM_Z (qadd_m128_s16, svint16_t,
     122  		z0 = svqadd_n_s16 (z0, -128),
     123  		z0 = svqadd (z0, -128))