(root)/
gcc-13.2.0/
gcc/
testsuite/
gcc.target/
aarch64/
sve2/
acle/
asm/
qshlu_s16.c
       1  /* { dg-final { check-function-bodies "**" "" "-DCHECK_ASM" } } */
       2  
       3  #include "test_sve_acle.h"
       4  
       5  /*
       6  ** qshlu_0_s16_m_tied1:
       7  **	sqshlu	z0\.h, p0/m, z0\.h, #0
       8  **	ret
       9  */
      10  TEST_DUAL_Z_REV (qshlu_0_s16_m_tied1, svuint16_t, svint16_t,
      11  		 z0_res = svqshlu_n_s16_m (p0, z0, 0),
      12  		 z0_res = svqshlu_m (p0, z0, 0))
      13  
      14  /*
      15  ** qshlu_0_s16_m_untied:
      16  **	movprfx	z0, z1
      17  **	sqshlu	z0\.h, p0/m, z0\.h, #0
      18  **	ret
      19  */
      20  TEST_DUAL_Z_REV (qshlu_0_s16_m_untied, svuint16_t, svint16_t,
      21  		 z0_res = svqshlu_n_s16_m (p0, z1, 0),
      22  		 z0_res = svqshlu_m (p0, z1, 0))
      23  
      24  /*
      25  ** qshlu_1_s16_m_tied1:
      26  **	sqshlu	z0\.h, p0/m, z0\.h, #1
      27  **	ret
      28  */
      29  TEST_DUAL_Z_REV (qshlu_1_s16_m_tied1, svuint16_t, svint16_t,
      30  		 z0_res = svqshlu_n_s16_m (p0, z0, 1),
      31  		 z0_res = svqshlu_m (p0, z0, 1))
      32  
      33  /*
      34  ** qshlu_1_s16_m_untied:
      35  **	movprfx	z0, z1
      36  **	sqshlu	z0\.h, p0/m, z0\.h, #1
      37  **	ret
      38  */
      39  TEST_DUAL_Z_REV (qshlu_1_s16_m_untied, svuint16_t, svint16_t,
      40  		 z0_res = svqshlu_n_s16_m (p0, z1, 1),
      41  		 z0_res = svqshlu_m (p0, z1, 1))
      42  
      43  /*
      44  ** qshlu_15_s16_m_tied1:
      45  **	sqshlu	z0\.h, p0/m, z0\.h, #15
      46  **	ret
      47  */
      48  TEST_DUAL_Z_REV (qshlu_15_s16_m_tied1, svuint16_t, svint16_t,
      49  		 z0_res = svqshlu_n_s16_m (p0, z0, 15),
      50  		 z0_res = svqshlu_m (p0, z0, 15))
      51  
      52  /*
      53  ** qshlu_15_s16_m_untied:
      54  **	movprfx	z0, z1
      55  **	sqshlu	z0\.h, p0/m, z0\.h, #15
      56  **	ret
      57  */
      58  TEST_DUAL_Z_REV (qshlu_15_s16_m_untied, svuint16_t, svint16_t,
      59  		 z0_res = svqshlu_n_s16_m (p0, z1, 15),
      60  		 z0_res = svqshlu_m (p0, z1, 15))
      61  
      62  /*
      63  ** qshlu_0_s16_z_tied1:
      64  **	movprfx	z0\.h, p0/z, z0\.h
      65  **	sqshlu	z0\.h, p0/m, z0\.h, #0
      66  **	ret
      67  */
      68  TEST_DUAL_Z_REV (qshlu_0_s16_z_tied1, svuint16_t, svint16_t,
      69  		 z0_res = svqshlu_n_s16_z (p0, z0, 0),
      70  		 z0_res = svqshlu_z (p0, z0, 0))
      71  
      72  /*
      73  ** qshlu_0_s16_z_untied:
      74  **	movprfx	z0\.h, p0/z, z1\.h
      75  **	sqshlu	z0\.h, p0/m, z0\.h, #0
      76  **	ret
      77  */
      78  TEST_DUAL_Z_REV (qshlu_0_s16_z_untied, svuint16_t, svint16_t,
      79  		 z0_res = svqshlu_n_s16_z (p0, z1, 0),
      80  		 z0_res = svqshlu_z (p0, z1, 0))
      81  
      82  /*
      83  ** qshlu_1_s16_z_tied1:
      84  **	movprfx	z0\.h, p0/z, z0\.h
      85  **	sqshlu	z0\.h, p0/m, z0\.h, #1
      86  **	ret
      87  */
      88  TEST_DUAL_Z_REV (qshlu_1_s16_z_tied1, svuint16_t, svint16_t,
      89  		 z0_res = svqshlu_n_s16_z (p0, z0, 1),
      90  		 z0_res = svqshlu_z (p0, z0, 1))
      91  
      92  /*
      93  ** qshlu_1_s16_z_untied:
      94  **	movprfx	z0\.h, p0/z, z1\.h
      95  **	sqshlu	z0\.h, p0/m, z0\.h, #1
      96  **	ret
      97  */
      98  TEST_DUAL_Z_REV (qshlu_1_s16_z_untied, svuint16_t, svint16_t,
      99  		 z0_res = svqshlu_n_s16_z (p0, z1, 1),
     100  		 z0_res = svqshlu_z (p0, z1, 1))
     101  
     102  /*
     103  ** qshlu_15_s16_z_tied1:
     104  **	movprfx	z0\.h, p0/z, z0\.h
     105  **	sqshlu	z0\.h, p0/m, z0\.h, #15
     106  **	ret
     107  */
     108  TEST_DUAL_Z_REV (qshlu_15_s16_z_tied1, svuint16_t, svint16_t,
     109  		 z0_res = svqshlu_n_s16_z (p0, z0, 15),
     110  		 z0_res = svqshlu_z (p0, z0, 15))
     111  
     112  /*
     113  ** qshlu_15_s16_z_untied:
     114  **	movprfx	z0\.h, p0/z, z1\.h
     115  **	sqshlu	z0\.h, p0/m, z0\.h, #15
     116  **	ret
     117  */
     118  TEST_DUAL_Z_REV (qshlu_15_s16_z_untied, svuint16_t, svint16_t,
     119  		 z0_res = svqshlu_n_s16_z (p0, z1, 15),
     120  		 z0_res = svqshlu_z (p0, z1, 15))
     121  
     122  /*
     123  ** qshlu_0_s16_x_tied1:
     124  **	sqshlu	z0\.h, p0/m, z0\.h, #0
     125  **	ret
     126  */
     127  TEST_DUAL_Z_REV (qshlu_0_s16_x_tied1, svuint16_t, svint16_t,
     128  		 z0_res = svqshlu_n_s16_x (p0, z0, 0),
     129  		 z0_res = svqshlu_x (p0, z0, 0))
     130  
     131  /*
     132  ** qshlu_0_s16_x_untied:
     133  **	movprfx	z0, z1
     134  **	sqshlu	z0\.h, p0/m, z0\.h, #0
     135  **	ret
     136  */
     137  TEST_DUAL_Z_REV (qshlu_0_s16_x_untied, svuint16_t, svint16_t,
     138  		 z0_res = svqshlu_n_s16_x (p0, z1, 0),
     139  		 z0_res = svqshlu_x (p0, z1, 0))
     140  
     141  /*
     142  ** qshlu_1_s16_x_tied1:
     143  **	sqshlu	z0\.h, p0/m, z0\.h, #1
     144  **	ret
     145  */
     146  TEST_DUAL_Z_REV (qshlu_1_s16_x_tied1, svuint16_t, svint16_t,
     147  		 z0_res = svqshlu_n_s16_x (p0, z0, 1),
     148  		 z0_res = svqshlu_x (p0, z0, 1))
     149  
     150  /*
     151  ** qshlu_1_s16_x_untied:
     152  **	movprfx	z0, z1
     153  **	sqshlu	z0\.h, p0/m, z0\.h, #1
     154  **	ret
     155  */
     156  TEST_DUAL_Z_REV (qshlu_1_s16_x_untied, svuint16_t, svint16_t,
     157  		 z0_res = svqshlu_n_s16_x (p0, z1, 1),
     158  		 z0_res = svqshlu_x (p0, z1, 1))
     159  
     160  /*
     161  ** qshlu_15_s16_x_tied1:
     162  **	sqshlu	z0\.h, p0/m, z0\.h, #15
     163  **	ret
     164  */
     165  TEST_DUAL_Z_REV (qshlu_15_s16_x_tied1, svuint16_t, svint16_t,
     166  		 z0_res = svqshlu_n_s16_x (p0, z0, 15),
     167  		 z0_res = svqshlu_x (p0, z0, 15))
     168  
     169  /*
     170  ** qshlu_15_s16_x_untied:
     171  **	movprfx	z0, z1
     172  **	sqshlu	z0\.h, p0/m, z0\.h, #15
     173  **	ret
     174  */
     175  TEST_DUAL_Z_REV (qshlu_15_s16_x_untied, svuint16_t, svint16_t,
     176  		 z0_res = svqshlu_n_s16_x (p0, z1, 15),
     177  		 z0_res = svqshlu_x (p0, z1, 15))