This document is complementary to the main Arm C Language Extensions (ACLE) specification, which can be found on the ACLE project on GitHub.
For the latest release of this document, see the ACLE project on GitHub.
Please report defects in this specification to the issue tracker page on GitHub.
This work is licensed under the Creative Commons Attribution-ShareAlike 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by-sa/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA.
Grant of Patent License. Subject to the terms and conditions of this license (both the Public License and this Patent License), each Licensor hereby grants to You a perpetual, worldwide, non-exclusive, no-charge, royalty-free, irrevocable (except as stated in this section) patent license to make, have made, use, offer to sell, sell, import, and otherwise transfer the Licensed Material, where such license applies only to those patent claims licensable by such Licensor that are necessarily infringed by their contribution(s) alone or by combination of their contribution(s) with the Licensed Material to which such contribution(s) was submitted. If You institute patent litigation against any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the Licensed Material or a contribution incorporated within the Licensed Material constitutes direct or contributory patent infringement, then any licenses granted to You under this license for that Licensed Material shall terminate as of the date such litigation is filed.
As identified more fully in the License section, this project is licensed under CC-BY-SA-4.0 along with an additional patent license. The language in the additional patent license is largely identical to that in Apache-2.0 (specifically, Section 3 of Apache-2.0 as reflected at https://www.apache.org/licenses/LICENSE-2.0) with two exceptions.
First, several changes were made related to the defined terms so as to reflect the fact that such defined terms need to align with the terminology in CC-BY-SA-4.0 rather than Apache-2.0 (e.g., changing “Work” to “Licensed Material”).
Second, the defensive termination clause was changed such that the scope of defensive termination applies to “any licenses granted to You” (rather than “any patent licenses granted to You”). This change is intended to help maintain a healthy ecosystem by providing additional protection to the community against patent litigation claims.
Contributions to this project are licensed under an inbound=outbound model such that any such contributions are licensed by the contributor under the same terms as those in the LICENSE file.
We do not require copyright assignment. The original contributor will retain the copyright.
The text of and illustrations in this document are licensed by Arm under a Creative Commons Attribution–Share Alike 4.0 International license ("CC-BY-SA-4.0”), with an additional clause on patents. The Arm trademarks featured here are registered trademarks or trademarks of Arm Limited (or its subsidiaries) in the US and/or elsewhere. All rights reserved. Please visit https://www.arm.com/company/policies/trademarks for more information about Arm’s trademarks.
Copyright 2019-2022 Arm Limited and/or its affiliates open-source-office@arm.com.
| Issue | Date | Change | 
|---|---|---|
| Q219-00 | 30 June 2019 | Version ACLE Q2 2019 | 
| Q319-00 | 30 September 2019 | Version ACLE Q3 2019 | 
| Q419-00 | 31 December 2019 | Version ACLE Q4 2019 | 
| Q220-00 | 30 May 2020 | Version ACLE Q2 2020 | 
| 2021Q2 | 02 July 2021 | Open source release | 
| 2021Q4 | 11 January 2022 | See Changes for 2021Q4 | 
.rst) to
Markdown (.md). The tool pandoc is now
used to render the PDF of the specs. The PDF is rendered using the
standard layout used in Arm specifications.| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vcreateq_f16( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcreateq_f32( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| int8x16_t [__arm_]vcreateq_s8( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcreateq_s16( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcreateq_s32( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| int64x2_t [__arm_]vcreateq_s64( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcreateq_u8( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcreateq_u16( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcreateq_u32( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| uint64x2_t [__arm_]vcreateq_u64( | a -> [Rt0, Rt1]b -> [Rt2, Rt3] | VMOV Qd[0], Rt0VMOV Qd[1], Rt1VMOV Qd[2], Rt2VMOV Qd[3], Rt3 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vddupq[_n]_u8( | a -> Rnimm in [1,2,4,8] | VDDUP.U8 Qd, Rn, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vddupq[_n]_u16( | a -> Rnimm in [1,2,4,8] | VDDUP.U16 Qd, Rn, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vddupq[_n]_u32( | a -> Rnimm in [1,2,4,8] | VDDUP.U32 Qd, Rn, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vddupq[_wb]_u8( | *a -> Rnimm in [1,2,4,8] | VDDUP.U8 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vddupq[_wb]_u16( | *a -> Rnimm in [1,2,4,8] | VDDUP.U16 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vddupq[_wb]_u32( | *a -> Rnimm in [1,2,4,8] | VDDUP.U32 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vddupq_m[_n_u8]( | inactive -> Qda -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U8 Qd, Rn, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vddupq_m[_n_u16]( | inactive -> Qda -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U16 Qd, Rn, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vddupq_m[_n_u32]( | inactive -> Qda -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U32 Qd, Rn, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vddupq_m[_wb_u8]( | inactive -> Qd*a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U8 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vddupq_m[_wb_u16]( | inactive -> Qd*a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U16 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vddupq_m[_wb_u32]( | inactive -> Qd*a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U32 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vddupq_x[_n]_u8( | a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U8 Qd, Rn, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vddupq_x[_n]_u16( | a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U16 Qd, Rn, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vddupq_x[_n]_u32( | a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U32 Qd, Rn, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vddupq_x[_wb]_u8( | *a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U8 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vddupq_x[_wb]_u16( | *a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U16 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vddupq_x[_wb]_u32( | *a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDDUPT.U32 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vdwdupq[_n]_u8( | a -> Rnb -> Rmimm in [1,2,4,8] | VDWDUP.U8 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vdwdupq[_n]_u16( | a -> Rnb -> Rmimm in [1,2,4,8] | VDWDUP.U16 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vdwdupq[_n]_u32( | a -> Rnb -> Rmimm in [1,2,4,8] | VDWDUP.U32 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vdwdupq[_wb]_u8( | *a -> Rnb -> Rmimm in [1,2,4,8] | VDWDUP.U8 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vdwdupq[_wb]_u16( | *a -> Rnb -> Rmimm in [1,2,4,8] | VDWDUP.U16 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vdwdupq[_wb]_u32( | *a -> Rnb -> Rmimm in [1,2,4,8] | VDWDUP.U32 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vdwdupq_m[_n_u8]( | inactive -> Qda -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U8 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vdwdupq_m[_n_u16]( | inactive -> Qda -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U16 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vdwdupq_m[_n_u32]( | inactive -> Qda -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U32 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vdwdupq_m[_wb_u8]( | inactive -> Qd*a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U8 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vdwdupq_m[_wb_u16]( | inactive -> Qd*a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U16 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vdwdupq_m[_wb_u32]( | inactive -> Qd*a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U32 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vdwdupq_x[_n]_u8( | a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U8 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vdwdupq_x[_n]_u16( | a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U16 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vdwdupq_x[_n]_u32( | a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U32 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vdwdupq_x[_wb]_u8( | *a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U8 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vdwdupq_x[_wb]_u16( | *a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U16 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vdwdupq_x[_wb]_u32( | *a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVDWDUPT.U32 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vidupq[_n]_u8( | a -> Rnimm in [1,2,4,8] | VIDUP.U8 Qd, Rn, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vidupq[_n]_u16( | a -> Rnimm in [1,2,4,8] | VIDUP.U16 Qd, Rn, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vidupq[_n]_u32( | a -> Rnimm in [1,2,4,8] | VIDUP.U32 Qd, Rn, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vidupq[_wb]_u8( | *a -> Rnimm in [1,2,4,8] | VIDUP.U8 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vidupq[_wb]_u16( | *a -> Rnimm in [1,2,4,8] | VIDUP.U16 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vidupq[_wb]_u32( | *a -> Rnimm in [1,2,4,8] | VIDUP.U32 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vidupq_m[_n_u8]( | inactive -> Qda -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U8 Qd, Rn, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vidupq_m[_n_u16]( | inactive -> Qda -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U16 Qd, Rn, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vidupq_m[_n_u32]( | inactive -> Qda -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U32 Qd, Rn, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vidupq_m[_wb_u8]( | inactive -> Qd*a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U8 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vidupq_m[_wb_u16]( | inactive -> Qd*a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U16 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vidupq_m[_wb_u32]( | inactive -> Qd*a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U32 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]vidupq_x[_n]_u8( | a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U8 Qd, Rn, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vidupq_x[_n]_u16( | a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U16 Qd, Rn, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vidupq_x[_n]_u32( | a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U32 Qd, Rn, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vidupq_x[_wb]_u8( | *a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U8 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]vidupq_x[_wb]_u16( | *a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U16 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]vidupq_x[_wb]_u32( | *a -> Rnimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIDUPT.U32 Qd, Rn, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]viwdupq[_n]_u8( | a -> Rnb -> Rmimm in [1,2,4,8] | VIWDUP.U8 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]viwdupq[_n]_u16( | a -> Rnb -> Rmimm in [1,2,4,8] | VIWDUP.U16 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]viwdupq[_n]_u32( | a -> Rnb -> Rmimm in [1,2,4,8] | VIWDUP.U32 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]viwdupq[_wb]_u8( | *a -> Rnb -> Rmimm in [1,2,4,8] | VIWDUP.U8 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]viwdupq[_wb]_u16( | *a -> Rnb -> Rmimm in [1,2,4,8] | VIWDUP.U16 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]viwdupq[_wb]_u32( | *a -> Rnb -> Rmimm in [1,2,4,8] | VIWDUP.U32 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]viwdupq_m[_n_u8]( | inactive -> Qda -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U8 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]viwdupq_m[_n_u16]( | inactive -> Qda -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U16 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]viwdupq_m[_n_u32]( | inactive -> Qda -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U32 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]viwdupq_m[_wb_u8]( | inactive -> Qd*a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U8 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]viwdupq_m[_wb_u16]( | inactive -> Qd*a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U16 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]viwdupq_m[_wb_u32]( | inactive -> Qd*a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U32 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint8x16_t [__arm_]viwdupq_x[_n]_u8( | a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U8 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]viwdupq_x[_n]_u16( | a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U16 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]viwdupq_x[_n]_u32( | a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U32 Qd, Rn, Rm, imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]viwdupq_x[_wb]_u8( | *a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U8 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint16x8_t [__arm_]viwdupq_x[_wb]_u16( | *a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U16 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| uint32x4_t [__arm_]viwdupq_x[_wb]_u32( | *a -> Rnb -> Rmimm in [1,2,4,8]p -> Rp | VMSR P0, RpVPSTVIWDUPT.U32 Qd, Rn, Rm, imm | Qd -> resultRn -> *a | MVE | 
| int8x16_t [__arm_]vdupq_n_s8(int8_t a) | a -> Rt | VDUP.8 Qd, Rt | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vdupq_n_s16(int16_t a) | a -> Rt | VDUP.16 Qd, Rt | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vdupq_n_s32(int32_t a) | a -> Rt | VDUP.32 Qd, Rt | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vdupq_n_u8(uint8_t a) | a -> Rt | VDUP.8 Qd, Rt | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vdupq_n_u16(uint16_t a) | a -> Rt | VDUP.16 Qd, Rt | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vdupq_n_u32(uint32_t a) | a -> Rt | VDUP.32 Qd, Rt | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vdupq_n_f16(float16_t a) | a -> Rt | VDUP.16 Qd, Rt | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vdupq_n_f32(float32_t a) | a -> Rt | VDUP.32 Qd, Rt | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vdupq_m[_n_s8]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.8 Qd, Rt | Qd -> result | MVE | 
| int16x8_t [__arm_]vdupq_m[_n_s16]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.16 Qd, Rt | Qd -> result | MVE | 
| int32x4_t [__arm_]vdupq_m[_n_s32]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.32 Qd, Rt | Qd -> result | MVE | 
| uint8x16_t [__arm_]vdupq_m[_n_u8]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.8 Qd, Rt | Qd -> result | MVE | 
| uint16x8_t [__arm_]vdupq_m[_n_u16]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.16 Qd, Rt | Qd -> result | MVE | 
| uint32x4_t [__arm_]vdupq_m[_n_u32]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.32 Qd, Rt | Qd -> result | MVE | 
| float16x8_t [__arm_]vdupq_m[_n_f16]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.16 Qd, Rt | Qd -> result | MVE | 
| float32x4_t [__arm_]vdupq_m[_n_f32]( | inactive -> Qda -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.32 Qd, Rt | Qd -> result | MVE | 
| int8x16_t [__arm_]vdupq_x_n_s8( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.8 Qd, Rt | Qd -> result | MVE | 
| int16x8_t [__arm_]vdupq_x_n_s16( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.16 Qd, Rt | Qd -> result | MVE | 
| int32x4_t [__arm_]vdupq_x_n_s32( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.32 Qd, Rt | Qd -> result | MVE | 
| uint8x16_t [__arm_]vdupq_x_n_u8( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.8 Qd, Rt | Qd -> result | MVE | 
| uint16x8_t [__arm_]vdupq_x_n_u16( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.16 Qd, Rt | Qd -> result | MVE | 
| uint32x4_t [__arm_]vdupq_x_n_u32( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.32 Qd, Rt | Qd -> result | MVE | 
| float16x8_t [__arm_]vdupq_x_n_f16( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.16 Qd, Rt | Qd -> result | MVE | 
| float32x4_t [__arm_]vdupq_x_n_f32( | a -> Rtp -> Rp | VMSR P0, RpVPSTVDUPT.32 Qd, Rt | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vrev16q[_s8](int8x16_t a) | a -> Qm | VREV16.8 Qd, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vrev16q[_u8](uint8x16_t a) | a -> Qm | VREV16.8 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vrev16q_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV16T.8 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrev16q_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV16T.8 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrev16q_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV16T.8 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrev16q_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV16T.8 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrev32q[_s8](int8x16_t a) | a -> Qm | VREV32.8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vrev32q[_s16](int16x8_t a) | a -> Qm | VREV32.16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vrev32q[_u8](uint8x16_t a) | a -> Qm | VREV32.8 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vrev32q[_u16](uint16x8_t a) | a -> Qm | VREV32.16 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrev32q[_f16](float16x8_t a) | a -> Qm | VREV32.16 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vrev32q_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrev32q_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.16 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrev32q_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.8 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrev32q_m[_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrev32q_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.16 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrev32q_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrev32q_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.16 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrev32q_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.8 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrev32q_x[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrev32q_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV32T.16 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrev64q[_s8](int8x16_t a) | a -> Qm | VREV64.8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vrev64q[_s16](int16x8_t a) | a -> Qm | VREV64.16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vrev64q[_s32](int32x4_t a) | a -> Qm | VREV64.32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vrev64q[_u8](uint8x16_t a) | a -> Qm | VREV64.8 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vrev64q[_u16](uint16x8_t a) | a -> Qm | VREV64.16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vrev64q[_u32](uint32x4_t a) | a -> Qm | VREV64.32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrev64q[_f16](float16x8_t a) | a -> Qm | VREV64.16 Qd, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vrev64q[_f32](float32x4_t a) | a -> Qm | VREV64.32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vrev64q_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrev64q_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrev64q_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrev64q_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.8 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrev64q_m[_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrev64q_m[_u32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrev64q_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrev64q_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrev64q_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrev64q_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrev64q_x[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrev64q_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.8 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrev64q_x[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrev64q_x[_u32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrev64q_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrev64q_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVREV64T.32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16_t [__arm_]vgetq_lane[_f16]( | a -> Qn0 <= idx <= 7 | VMOV.U16 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| float32_t [__arm_]vgetq_lane[_f32]( | a -> Qn0 <= idx <= 3 | VMOV.32 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| int8_t [__arm_]vgetq_lane[_s8]( | a -> Qn0 <= idx <= 15 | VMOV.S8 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| int16_t [__arm_]vgetq_lane[_s16]( | a -> Qn0 <= idx <= 7 | VMOV.S16 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| int32_t [__arm_]vgetq_lane[_s32]( | a -> Qn0 <= idx <= 3 | VMOV.32 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| int64_t [__arm_]vgetq_lane[_s64]( | a -> Qn0 <= idx <= 1 | VMOV Rt1, Rt2, D(2*n+idx) | [Rt1,Rt2] -> result | MVE/NEON | 
| uint8_t [__arm_]vgetq_lane[_u8]( | a -> Qn0 <= idx <= 15 | VMOV.U8 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| uint16_t [__arm_]vgetq_lane[_u16]( | a -> Qn0 <= idx <= 7 | VMOV.U16 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| uint32_t [__arm_]vgetq_lane[_u32]( | a -> Qn0 <= idx <= 3 | VMOV.32 Rt, Qn[idx] | Rt -> result | MVE/NEON | 
| uint64_t [__arm_]vgetq_lane[_u64]( | a -> Qn0 <= idx <= 1 | VMOV Rt1, Rt2, D(2*n+idx) | [Rt1,Rt2] -> result | MVE/NEON | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vsetq_lane[_f16]( | a -> Rtb -> Qd0 <= idx <= 7 | VMOV.16 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vsetq_lane[_f32]( | a -> Rtb -> Qd0 <= idx <= 3 | VMOV.32 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vsetq_lane[_s8]( | a -> Rtb -> Qd0 <= idx <= 15 | VMOV.8 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vsetq_lane[_s16]( | a -> Rtb -> Qd0 <= idx <= 7 | VMOV.16 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vsetq_lane[_s32]( | a -> Rtb -> Qd0 <= idx <= 3 | VMOV.32 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vsetq_lane[_s64]( | a -> [Rt1,Rt2]b -> Qd0 <= idx <= 1 | VMOV D(2*d+idx), Rt1, Rt2 | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vsetq_lane[_u8]( | a -> Rtb -> Qd0 <= idx <= 15 | VMOV.8 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vsetq_lane[_u16]( | a -> Rtb -> Qd0 <= idx <= 7 | VMOV.16 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vsetq_lane[_u32]( | a -> Rtb -> Qd0 <= idx <= 3 | VMOV.32 Qd[idx], Rt | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vsetq_lane[_u64]( | a -> [Rt1,Rt2]b -> Qd0 <= idx <= 1 | VMOV D(2*d+idx), Rt1, Rt2 | Qd -> result | MVE/NEON | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vuninitializedq_s8(void) | Qd -> result | MVE | ||
| int16x8_t [__arm_]vuninitializedq_s16(void) | Qd -> result | MVE | ||
| int32x4_t [__arm_]vuninitializedq_s32(void) | Qd -> result | MVE | ||
| int64x2_t [__arm_]vuninitializedq_s64(void) | Qd -> result | MVE | ||
| uint8x16_t [__arm_]vuninitializedq_u8(void) | Qd -> result | MVE | ||
| uint16x8_t [__arm_]vuninitializedq_u16(void) | Qd -> result | MVE | ||
| uint32x4_t [__arm_]vuninitializedq_u32(void) | Qd -> result | MVE | ||
| uint64x2_t [__arm_]vuninitializedq_u64(void) | Qd -> result | MVE | ||
| float16x8_t [__arm_]vuninitializedq_f16(void) | Qd -> result | MVE | ||
| float32x4_t [__arm_]vuninitializedq_f32(void) | Qd -> result | MVE | ||
| int8x16_t [__arm_]vuninitializedq(int8x16_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| int16x8_t [__arm_]vuninitializedq(int16x8_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| int32x4_t [__arm_]vuninitializedq(int32x4_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| int64x2_t [__arm_]vuninitializedq(int64x2_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| uint8x16_t [__arm_]vuninitializedq(uint8x16_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| uint16x8_t [__arm_]vuninitializedq(uint16x8_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| uint32x4_t [__arm_]vuninitializedq(uint32x4_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| uint64x2_t [__arm_]vuninitializedq(uint64x2_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| float16x8_t [__arm_]vuninitializedq(float16x8_t t) | t -> Do Not Evaluate | Qd -> result | MVE | |
| float32x4_t [__arm_]vuninitializedq(float32x4_t t) | t -> Do Not Evaluate | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vcmpeqq[_f16]( | a -> Qnb -> Qm | VCMP.F16 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_f32]( | a -> Qnb -> Qm | VCMP.F32 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_s8]( | a -> Qnb -> Qm | VCMP.I8 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_s16]( | a -> Qnb -> Qm | VCMP.I16 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_s32]( | a -> Qnb -> Qm | VCMP.I32 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_u8]( | a -> Qnb -> Qm | VCMP.I8 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_u16]( | a -> Qnb -> Qm | VCMP.I16 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_u32]( | a -> Qnb -> Qm | VCMP.I32 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_f16]( | a -> Qnb -> Rm | VCMP.F16 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_f32]( | a -> Qnb -> Rm | VCMP.F32 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_s8]( | a -> Qnb -> Rm | VCMP.I8 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_s16]( | a -> Qnb -> Rm | VCMP.I16 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_s32]( | a -> Qnb -> Rm | VCMP.I32 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_u8]( | a -> Qnb -> Rm | VCMP.I8 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_u16]( | a -> Qnb -> Rm | VCMP.I16 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq[_n_u32]( | a -> Qnb -> Rm | VCMP.I32 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 eq, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpeqq_m[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 eq, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vcmpneq[_f16]( | a -> Qnb -> Qm | VCMP.F16 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_f32]( | a -> Qnb -> Qm | VCMP.F32 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_s8]( | a -> Qnb -> Qm | VCMP.I8 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_s16]( | a -> Qnb -> Qm | VCMP.I16 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_s32]( | a -> Qnb -> Qm | VCMP.I32 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_u8]( | a -> Qnb -> Qm | VCMP.I8 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_u16]( | a -> Qnb -> Qm | VCMP.I16 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_u32]( | a -> Qnb -> Qm | VCMP.I32 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 ne, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_f16]( | a -> Qnb -> Rm | VCMP.F16 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_f32]( | a -> Qnb -> Rm | VCMP.F32 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_s8]( | a -> Qnb -> Rm | VCMP.I8 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_s16]( | a -> Qnb -> Rm | VCMP.I16 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_s32]( | a -> Qnb -> Rm | VCMP.I32 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_u8]( | a -> Qnb -> Rm | VCMP.I8 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_u16]( | a -> Qnb -> Rm | VCMP.I16 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq[_n_u32]( | a -> Qnb -> Rm | VCMP.I32 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I8 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I16 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpneq_m[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.I32 ne, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vcmpgeq[_f16]( | a -> Qnb -> Qm | VCMP.F16 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_f32]( | a -> Qnb -> Qm | VCMP.F32 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_s8]( | a -> Qnb -> Qm | VCMP.S8 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_s16]( | a -> Qnb -> Qm | VCMP.S16 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_s32]( | a -> Qnb -> Qm | VCMP.S32 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 ge, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_n_f16]( | a -> Qnb -> Rm | VCMP.F16 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_n_f32]( | a -> Qnb -> Rm | VCMP.F32 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_n_s8]( | a -> Qnb -> Rm | VCMP.S8 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_n_s16]( | a -> Qnb -> Rm | VCMP.S16 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq[_n_s32]( | a -> Qnb -> Rm | VCMP.S32 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgeq_m[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 ge, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq[_u8]( | a -> Qnb -> Qm | VCMP.U8 cs, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq[_u16]( | a -> Qnb -> Qm | VCMP.U16 cs, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq[_u32]( | a -> Qnb -> Qm | VCMP.U32 cs, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq_m[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.U8 cs, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq_m[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.U16 cs, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq_m[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.U32 cs, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq[_n_u8]( | a -> Qnb -> Rm | VCMP.U8 cs, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq[_n_u16]( | a -> Qnb -> Rm | VCMP.U16 cs, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq[_n_u32]( | a -> Qnb -> Rm | VCMP.U32 cs, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq_m[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.U8 cs, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq_m[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.U16 cs, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpcsq_m[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.U32 cs, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vcmpgtq[_f16]( | a -> Qnb -> Qm | VCMP.F16 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_f32]( | a -> Qnb -> Qm | VCMP.F32 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_s8]( | a -> Qnb -> Qm | VCMP.S8 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_s16]( | a -> Qnb -> Qm | VCMP.S16 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_s32]( | a -> Qnb -> Qm | VCMP.S32 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 gt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_n_f16]( | a -> Qnb -> Rm | VCMP.F16 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_n_f32]( | a -> Qnb -> Rm | VCMP.F32 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_n_s8]( | a -> Qnb -> Rm | VCMP.S8 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_n_s16]( | a -> Qnb -> Rm | VCMP.S16 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq[_n_s32]( | a -> Qnb -> Rm | VCMP.S32 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpgtq_m[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 gt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq[_u8]( | a -> Qnb -> Qm | VCMP.U8 hi, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq[_u16]( | a -> Qnb -> Qm | VCMP.U16 hi, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq[_u32]( | a -> Qnb -> Qm | VCMP.U32 hi, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq_m[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.U8 hi, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq_m[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.U16 hi, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq_m[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.U32 hi, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq[_n_u8]( | a -> Qnb -> Rm | VCMP.U8 hi, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq[_n_u16]( | a -> Qnb -> Rm | VCMP.U16 hi, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq[_n_u32]( | a -> Qnb -> Rm | VCMP.U32 hi, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq_m[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.U8 hi, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq_m[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.U16 hi, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmphiq_m[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.U32 hi, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vcmpleq[_f16]( | a -> Qnb -> Qm | VCMP.F16 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_f32]( | a -> Qnb -> Qm | VCMP.F32 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_s8]( | a -> Qnb -> Qm | VCMP.S8 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_s16]( | a -> Qnb -> Qm | VCMP.S16 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_s32]( | a -> Qnb -> Qm | VCMP.S32 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 le, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_n_f16]( | a -> Qnb -> Rm | VCMP.F16 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_n_f32]( | a -> Qnb -> Rm | VCMP.F32 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_n_s8]( | a -> Qnb -> Rm | VCMP.S8 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_n_s16]( | a -> Qnb -> Rm | VCMP.S16 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq[_n_s32]( | a -> Qnb -> Rm | VCMP.S32 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpleq_m[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 le, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vcmpltq[_f16]( | a -> Qnb -> Qm | VCMP.F16 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_f32]( | a -> Qnb -> Qm | VCMP.F32 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_s8]( | a -> Qnb -> Qm | VCMP.S8 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_s16]( | a -> Qnb -> Qm | VCMP.S16 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_s32]( | a -> Qnb -> Qm | VCMP.S32 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 lt, Qn, QmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_n_f16]( | a -> Qnb -> Rm | VCMP.F16 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_n_f32]( | a -> Qnb -> Rm | VCMP.F32 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_n_s8]( | a -> Qnb -> Rm | VCMP.S8 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_n_s16]( | a -> Qnb -> Rm | VCMP.S16 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq[_n_s32]( | a -> Qnb -> Rm | VCMP.S32 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F16 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.F32 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S8 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S16 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vcmpltq_m[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVCMPT.S32 lt, Qn, RmVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vminq[_s8]( | a -> Qnb -> Qm | VMIN.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vminq[_s16]( | a -> Qnb -> Qm | VMIN.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vminq[_s32]( | a -> Qnb -> Qm | VMIN.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vminq[_u8]( | a -> Qnb -> Qm | VMIN.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vminq[_u16]( | a -> Qnb -> Qm | VMIN.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vminq[_u32]( | a -> Qnb -> Qm | VMIN.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vminq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vminq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vminq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vminq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vminq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vminq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vminq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vminq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vminq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vminq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vminq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vminq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vminaq[_s8]( | a -> Qdab -> Qm | VMINA.S8 Qda, Qm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vminaq[_s16]( | a -> Qdab -> Qm | VMINA.S16 Qda, Qm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vminaq[_s32]( | a -> Qdab -> Qm | VMINA.S32 Qda, Qm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vminaq_m[_s8]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINAT.S8 Qda, Qm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vminaq_m[_s16]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINAT.S16 Qda, Qm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vminaq_m[_s32]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINAT.S32 Qda, Qm | Qda -> result | MVE | 
| int8_t [__arm_]vminvq[_s8]( | a -> Rdab -> Qm | VMINV.S8 Rda, Qm | Rda -> result | MVE | 
| int16_t [__arm_]vminvq[_s16]( | a -> Rdab -> Qm | VMINV.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vminvq[_s32]( | a -> Rdab -> Qm | VMINV.S32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vminvq[_u8]( | a -> Rdab -> Qm | VMINV.U8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vminvq[_u16]( | a -> Rdab -> Qm | VMINV.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vminvq[_u32]( | a -> Rdab -> Qm | VMINV.U32 Rda, Qm | Rda -> result | MVE | 
| int8_t [__arm_]vminvq_p[_s8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINVT.S8 Rda, Qm | Rda -> result | MVE | 
| int16_t [__arm_]vminvq_p[_s16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINVT.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vminvq_p[_s32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINVT.S32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vminvq_p[_u8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINVT.U8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vminvq_p[_u16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINVT.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vminvq_p[_u32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINVT.U32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vminavq[_s8]( | a -> Rdab -> Qm | VMINAV.S8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vminavq[_s16]( | a -> Rdab -> Qm | VMINAV.S16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vminavq[_s32]( | a -> Rdab -> Qm | VMINAV.S32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vminavq_p[_s8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINAVT.S8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vminavq_p[_s16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINAVT.S16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vminavq_p[_s32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINAVT.S32 Rda, Qm | Rda -> result | MVE | 
| float16x8_t [__arm_]vminnmq[_f16]( | a -> Qnb -> Qm | VMINNM.F16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vminnmq[_f32]( | a -> Qnb -> Qm | VMINNM.F32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vminnmq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vminnmq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vminnmq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vminnmq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vminnmaq[_f16]( | a -> Qdab -> Qm | VMINNMA.F16 Qda, Qm | Qda -> result | MVE | 
| float32x4_t [__arm_]vminnmaq[_f32]( | a -> Qdab -> Qm | VMINNMA.F32 Qda, Qm | Qda -> result | MVE | 
| float16x8_t [__arm_]vminnmaq_m[_f16]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMAT.F16 Qda, Qm | Qda -> result | MVE | 
| float32x4_t [__arm_]vminnmaq_m[_f32]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMAT.F32 Qda, Qm | Qda -> result | MVE | 
| float16_t [__arm_]vminnmvq[_f16]( | a -> Rdab -> Qm | VMINNMV.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vminnmvq[_f32]( | a -> Rdab -> Qm | VMINNMV.F32 Rda, Qm | Rda -> result | MVE | 
| float16_t [__arm_]vminnmvq_p[_f16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMVT.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vminnmvq_p[_f32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMVT.F32 Rda, Qm | Rda -> result | MVE | 
| float16_t [__arm_]vminnmavq[_f16]( | a -> Rdab -> Qm | VMINNMAV.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vminnmavq[_f32]( | a -> Rdab -> Qm | VMINNMAV.F32 Rda, Qm | Rda -> result | MVE | 
| float16_t [__arm_]vminnmavq_p[_f16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMAVT.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vminnmavq_p[_f32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMINNMAVT.F32 Rda, Qm | Rda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vmaxq[_s8]( | a -> Qnb -> Qm | VMAX.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vmaxq[_s16]( | a -> Qnb -> Qm | VMAX.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vmaxq[_s32]( | a -> Qnb -> Qm | VMAX.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vmaxq[_u8]( | a -> Qnb -> Qm | VMAX.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vmaxq[_u16]( | a -> Qnb -> Qm | VMAX.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vmaxq[_u32]( | a -> Qnb -> Qm | VMAX.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vmaxq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmaxq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmaxq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmaxq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmaxq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmaxq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmaxq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmaxq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmaxq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmaxq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmaxq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmaxq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmaxaq[_s8]( | a -> Qdab -> Qm | VMAXA.S8 Qda, Qm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vmaxaq[_s16]( | a -> Qdab -> Qm | VMAXA.S16 Qda, Qm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vmaxaq[_s32]( | a -> Qdab -> Qm | VMAXA.S32 Qda, Qm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vmaxaq_m[_s8]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXAT.S8 Qda, Qm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vmaxaq_m[_s16]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXAT.S16 Qda, Qm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vmaxaq_m[_s32]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXAT.S32 Qda, Qm | Qda -> result | MVE | 
| int8_t [__arm_]vmaxvq[_s8]( | a -> Rdab -> Qm | VMAXV.S8 Rda, Qm | Rda -> result | MVE | 
| int16_t [__arm_]vmaxvq[_s16]( | a -> Rdab -> Qm | VMAXV.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmaxvq[_s32]( | a -> Rdab -> Qm | VMAXV.S32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vmaxvq[_u8]( | a -> Rdab -> Qm | VMAXV.U8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vmaxvq[_u16]( | a -> Rdab -> Qm | VMAXV.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmaxvq[_u32]( | a -> Rdab -> Qm | VMAXV.U32 Rda, Qm | Rda -> result | MVE | 
| int8_t [__arm_]vmaxvq_p[_s8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXVT.S8 Rda, Qm | Rda -> result | MVE | 
| int16_t [__arm_]vmaxvq_p[_s16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXVT.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmaxvq_p[_s32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXVT.S32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vmaxvq_p[_u8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXVT.U8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vmaxvq_p[_u16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXVT.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmaxvq_p[_u32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXVT.U32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vmaxavq[_s8]( | a -> Rdab -> Qm | VMAXAV.S8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vmaxavq[_s16]( | a -> Rdab -> Qm | VMAXAV.S16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmaxavq[_s32]( | a -> Rdab -> Qm | VMAXAV.S32 Rda, Qm | Rda -> result | MVE | 
| uint8_t [__arm_]vmaxavq_p[_s8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXAVT.S8 Rda, Qm | Rda -> result | MVE | 
| uint16_t [__arm_]vmaxavq_p[_s16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXAVT.S16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmaxavq_p[_s32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXAVT.S32 Rda, Qm | Rda -> result | MVE | 
| float16x8_t [__arm_]vmaxnmq[_f16]( | a -> Qnb -> Qm | VMAXNM.F16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vmaxnmq[_f32]( | a -> Qnb -> Qm | VMAXNM.F32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vmaxnmq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vmaxnmq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vmaxnmq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vmaxnmq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vmaxnmaq[_f16]( | a -> Qdab -> Qm | VMAXNMA.F16 Qda, Qm | Qda -> result | MVE | 
| float32x4_t [__arm_]vmaxnmaq[_f32]( | a -> Qdab -> Qm | VMAXNMA.F32 Qda, Qm | Qda -> result | MVE | 
| float16x8_t [__arm_]vmaxnmaq_m[_f16]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMAT.F16 Qda, Qm | Qda -> result | MVE | 
| float32x4_t [__arm_]vmaxnmaq_m[_f32]( | a -> Qdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMAT.F32 Qda, Qm | Qda -> result | MVE | 
| float16_t [__arm_]vmaxnmvq[_f16]( | a -> Rdab -> Qm | VMAXNMV.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vmaxnmvq[_f32]( | a -> Rdab -> Qm | VMAXNMV.F32 Rda, Qm | Rda -> result | MVE | 
| float16_t [__arm_]vmaxnmvq_p[_f16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMVT.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vmaxnmvq_p[_f32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMVT.F32 Rda, Qm | Rda -> result | MVE | 
| float16_t [__arm_]vmaxnmavq[_f16]( | a -> Rdab -> Qm | VMAXNMAV.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vmaxnmavq[_f32]( | a -> Rdab -> Qm | VMAXNMAV.F32 Rda, Qm | Rda -> result | MVE | 
| float16_t [__arm_]vmaxnmavq_p[_f16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMAVT.F16 Rda, Qm | Rda -> result | MVE | 
| float32_t [__arm_]vmaxnmavq_p[_f32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVMAXNMAVT.F32 Rda, Qm | Rda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| uint32_t [__arm_]vabavq[_s8]( | a -> Rdab -> Qnc -> Qm | VABAV.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq[_s16]( | a -> Rdab -> Qnc -> Qm | VABAV.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq[_s32]( | a -> Rdab -> Qnc -> Qm | VABAV.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq[_u8]( | a -> Rdab -> Qnc -> Qm | VABAV.U8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq[_u16]( | a -> Rdab -> Qnc -> Qm | VABAV.U16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq[_u32]( | a -> Rdab -> Qnc -> Qm | VABAV.U32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq_p[_s8]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVABAVT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq_p[_s16]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVABAVT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq_p[_s32]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVABAVT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq_p[_u8]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVABAVT.U8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq_p[_u16]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVABAVT.U16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vabavq_p[_u32]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVABAVT.U32 Rda, Qn, Qm | Rda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vabdq[_s8]( | a -> Qnb -> Qm | VABD.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vabdq[_s16]( | a -> Qnb -> Qm | VABD.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vabdq[_s32]( | a -> Qnb -> Qm | VABD.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vabdq[_u8]( | a -> Qnb -> Qm | VABD.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vabdq[_u16]( | a -> Qnb -> Qm | VABD.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vabdq[_u32]( | a -> Qnb -> Qm | VABD.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vabdq[_f16]( | a -> Qnb -> Qm | VABD.F16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vabdq[_f32]( | a -> Qnb -> Qm | VABD.F32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vabdq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vabdq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vabdq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vabdq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vabdq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vabdq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vabdq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vabdq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vabdq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vabdq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vabdq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vabdq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vabdq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vabdq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vabdq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vabdq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVABDT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vabsq[_f16](float16x8_t a) | a -> Qm | VABS.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vabsq[_f32](float32x4_t a) | a -> Qm | VABS.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vabsq[_s8](int8x16_t a) | a -> Qm | VABS.S8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vabsq[_s16](int16x8_t a) | a -> Qm | VABS.S16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vabsq[_s32](int32x4_t a) | a -> Qm | VABS.S32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vabsq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVABST.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vabsq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVABST.F32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vabsq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVABST.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vabsq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVABST.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vabsq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVABST.S32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vabsq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVABST.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vabsq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVABST.F32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vabsq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVABST.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vabsq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVABST.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vabsq_x[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVABST.S32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqabsq[_s8](int8x16_t a) | a -> Qm | VQABS.S8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqabsq[_s16](int16x8_t a) | a -> Qm | VQABS.S16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqabsq[_s32](int32x4_t a) | a -> Qm | VQABS.S32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqabsq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVQABST.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqabsq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVQABST.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqabsq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVQABST.S32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int32x4_t [__arm_]vadciq[_s32]( | a -> Qnb -> Qm | VADCI.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| uint32x4_t [__arm_]vadciq[_u32]( | a -> Qnb -> Qm | VADCI.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| int32x4_t [__arm_]vadciq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADCIT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| uint32x4_t [__arm_]vadciq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADCIT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| int32x4_t [__arm_]vadcq[_s32]( | a -> Qnb -> Qm*carry -> Rt | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVADC.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| uint32x4_t [__arm_]vadcq[_u32]( | a -> Qnb -> Qm*carry -> Rt | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVADC.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| int32x4_t [__arm_]vadcq_m[_s32]( | inactive -> Qda -> Qnb -> Qm*carry -> Rtp -> Rp | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVMSR P0, RpVPSTVADCT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| uint32x4_t [__arm_]vadcq_m[_u32]( | inactive -> Qda -> Qnb -> Qm*carry -> Rtp -> Rp | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVMSR P0, RpVPSTVADCT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| float16x8_t [__arm_]vaddq[_f16]( | a -> Qnb -> Qm | VADD.F16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vaddq[_f32]( | a -> Qnb -> Qm | VADD.F32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vaddq[_n_f16]( | a -> Qnb -> Rm | VADD.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vaddq[_n_f32]( | a -> Qnb -> Rm | VADD.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vaddq[_s8]( | a -> Qnb -> Qm | VADD.I8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vaddq[_s16]( | a -> Qnb -> Qm | VADD.I16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vaddq[_s32]( | a -> Qnb -> Qm | VADD.I32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vaddq[_n_s8]( | a -> Qnb -> Rm | VADD.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vaddq[_n_s16]( | a -> Qnb -> Rm | VADD.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vaddq[_n_s32]( | a -> Qnb -> Rm | VADD.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vaddq[_u8]( | a -> Qnb -> Qm | VADD.I8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vaddq[_u16]( | a -> Qnb -> Qm | VADD.I16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vaddq[_u32]( | a -> Qnb -> Qm | VADD.I32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vaddq[_n_u8]( | a -> Qnb -> Rm | VADD.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vaddq[_n_u16]( | a -> Qnb -> Rm | VADD.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vaddq[_n_u32]( | a -> Qnb -> Rm | VADD.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vaddq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vaddq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vaddq_m[_n_f16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vaddq_m[_n_f32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vaddq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vaddq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vaddq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vaddq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vaddq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vaddq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vaddq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vaddq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vaddq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vaddq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vaddq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vaddq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vaddq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vaddq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vaddq_x[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vaddq_x[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vaddq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vaddq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vaddq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vaddq_x[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vaddq_x[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vaddq_x[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vaddq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vaddq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vaddq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vaddq_x[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vaddq_x[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vaddq_x[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVADDT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| int64_t [__arm_]vaddlvaq[_s32]( | a -> [RdaHi,RdaLo]b -> Qm | VADDLVA.S32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vaddlvaq[_u32]( | a -> [RdaHi,RdaLo]b -> Qm | VADDLVA.U32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vaddlvaq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qmp -> Rp | VMSR P0, RpVPSTVADDLVAT.S32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vaddlvaq_p[_u32]( | a -> [RdaHi,RdaLo]b -> Qmp -> Rp | VMSR P0, RpVPSTVADDLVAT.U32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vaddlvq[_s32](int32x4_t a) | a -> Qm | VADDLV.S32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vaddlvq[_u32](uint32x4_t a) | a -> Qm | VADDLV.U32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vaddlvq_p[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDLVT.S32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vaddlvq_p[_u32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDLVT.U32 RdaLo, RdaHi, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int32_t [__arm_]vaddvaq[_s8]( | a -> Rdab -> Qm | VADDVA.S8 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvaq[_s16]( | a -> Rdab -> Qm | VADDVA.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvaq[_s32]( | a -> Rdab -> Qm | VADDVA.S32 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvaq[_u8]( | a -> Rdab -> Qm | VADDVA.U8 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvaq[_u16]( | a -> Rdab -> Qm | VADDVA.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvaq[_u32]( | a -> Rdab -> Qm | VADDVA.U32 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvaq_p[_s8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVADDVAT.S8 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvaq_p[_s16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVADDVAT.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvaq_p[_s32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVADDVAT.S32 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvaq_p[_u8]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVADDVAT.U8 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvaq_p[_u16]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVADDVAT.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvaq_p[_u32]( | a -> Rdab -> Qmp -> Rp | VMSR P0, RpVPSTVADDVAT.U32 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvq[_s8](int8x16_t a) | a -> Qm | VADDV.S8 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvq[_s16](int16x8_t a) | a -> Qm | VADDV.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvq[_s32](int32x4_t a) | a -> Qm | VADDV.S32 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvq[_u8](uint8x16_t a) | a -> Qm | VADDV.U8 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvq[_u16](uint16x8_t a) | a -> Qm | VADDV.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvq[_u32](uint32x4_t a) | a -> Qm | VADDV.U32 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvq_p[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDVT.S8 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvq_p[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDVT.S16 Rda, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vaddvq_p[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDVT.S32 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvq_p[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDVT.U8 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvq_p[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDVT.U16 Rda, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vaddvq_p[_u32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVADDVT.U32 Rda, Qm | Rda -> result | MVE | 
| int8x16_t [__arm_]vhaddq[_n_s8]( | a -> Qnb -> Rm | VHADD.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhaddq[_n_s16]( | a -> Qnb -> Rm | VHADD.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhaddq[_n_s32]( | a -> Qnb -> Rm | VHADD.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhaddq[_n_u8]( | a -> Qnb -> Rm | VHADD.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhaddq[_n_u16]( | a -> Qnb -> Rm | VHADD.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhaddq[_n_u32]( | a -> Qnb -> Rm | VHADD.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhaddq[_s8]( | a -> Qnb -> Qm | VHADD.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vhaddq[_s16]( | a -> Qnb -> Qm | VHADD.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vhaddq[_s32]( | a -> Qnb -> Qm | VHADD.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vhaddq[_u8]( | a -> Qnb -> Qm | VHADD.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vhaddq[_u16]( | a -> Qnb -> Qm | VHADD.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vhaddq[_u32]( | a -> Qnb -> Qm | VHADD.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vhaddq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhaddq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhaddq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhaddq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhaddq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhaddq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhaddq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhaddq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhaddq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhaddq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhaddq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhaddq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhaddq_x[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhaddq_x[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhaddq_x[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhaddq_x[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhaddq_x[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhaddq_x[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHADDT.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhaddq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhaddq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhaddq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhaddq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhaddq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhaddq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHADDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrhaddq[_s8]( | a -> Qnb -> Qm | VRHADD.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vrhaddq[_s16]( | a -> Qnb -> Qm | VRHADD.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vrhaddq[_s32]( | a -> Qnb -> Qm | VRHADD.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vrhaddq[_u8]( | a -> Qnb -> Qm | VRHADD.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vrhaddq[_u16]( | a -> Qnb -> Qm | VRHADD.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vrhaddq[_u32]( | a -> Qnb -> Qm | VRHADD.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vrhaddq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrhaddq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrhaddq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrhaddq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrhaddq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrhaddq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrhaddq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrhaddq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrhaddq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrhaddq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrhaddq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrhaddq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRHADDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqaddq[_n_s8]( | a -> Qnb -> Rm | VQADD.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqaddq[_n_s16]( | a -> Qnb -> Rm | VQADD.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqaddq[_n_s32]( | a -> Qnb -> Rm | VQADD.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqaddq[_n_u8]( | a -> Qnb -> Rm | VQADD.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqaddq[_n_u16]( | a -> Qnb -> Rm | VQADD.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqaddq[_n_u32]( | a -> Qnb -> Rm | VQADD.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqaddq[_s8]( | a -> Qnb -> Qm | VQADD.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqaddq[_s16]( | a -> Qnb -> Qm | VQADD.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqaddq[_s32]( | a -> Qnb -> Qm | VQADD.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vqaddq[_u8]( | a -> Qnb -> Qm | VQADD.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vqaddq[_u16]( | a -> Qnb -> Qm | VQADD.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vqaddq[_u32]( | a -> Qnb -> Qm | VQADD.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqaddq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQADDT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqaddq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQADDT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqaddq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQADDT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqaddq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQADDT.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqaddq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQADDT.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqaddq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQADDT.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqaddq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQADDT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqaddq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQADDT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqaddq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQADDT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqaddq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQADDT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqaddq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQADDT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqaddq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQADDT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vmulhq[_s8]( | a -> Qnb -> Qm | VMULH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulhq[_s16]( | a -> Qnb -> Qm | VMULH.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulhq[_s32]( | a -> Qnb -> Qm | VMULH.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulhq[_u8]( | a -> Qnb -> Qm | VMULH.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulhq[_u16]( | a -> Qnb -> Qm | VMULH.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulhq[_u32]( | a -> Qnb -> Qm | VMULH.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmulhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulhq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulhq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulhq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmulhq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulhq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulhq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulhq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulhq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulhq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULHT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmullbq_poly[_p8]( | a -> Qnb -> Qm | VMULLB.P8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmullbq_poly[_p16]( | a -> Qnb -> Qm | VMULLB.P16 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmullbq_int[_s8]( | a -> Qnb -> Qm | VMULLB.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmullbq_int[_s16]( | a -> Qnb -> Qm | VMULLB.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vmullbq_int[_s32]( | a -> Qnb -> Qm | VMULLB.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmullbq_int[_u8]( | a -> Qnb -> Qm | VMULLB.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmullbq_int[_u16]( | a -> Qnb -> Qm | VMULLB.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vmullbq_int[_u32]( | a -> Qnb -> Qm | VMULLB.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmullbq_poly_m[_p8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.P8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmullbq_poly_m[_p16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.P16 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmullbq_int_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmullbq_int_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vmullbq_int_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmullbq_int_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmullbq_int_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vmullbq_int_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmullbq_poly_x[_p8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.P8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmullbq_poly_x[_p16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.P16 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmullbq_int_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmullbq_int_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vmullbq_int_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmullbq_int_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmullbq_int_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vmullbq_int_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLBT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulltq_poly[_p8]( | a -> Qnb -> Qm | VMULLT.P8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulltq_poly[_p16]( | a -> Qnb -> Qm | VMULLT.P16 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulltq_int[_s8]( | a -> Qnb -> Qm | VMULLT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulltq_int[_s16]( | a -> Qnb -> Qm | VMULLT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vmulltq_int[_s32]( | a -> Qnb -> Qm | VMULLT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulltq_int[_u8]( | a -> Qnb -> Qm | VMULLT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulltq_int[_u16]( | a -> Qnb -> Qm | VMULLT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vmulltq_int[_u32]( | a -> Qnb -> Qm | VMULLT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulltq_poly_m[_p8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.P8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulltq_poly_m[_p16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.P16 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulltq_int_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulltq_int_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vmulltq_int_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulltq_int_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulltq_int_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vmulltq_int_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulltq_poly_x[_p8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.P8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulltq_poly_x[_p16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.P16 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulltq_int_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulltq_int_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vmulltq_int_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulltq_int_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulltq_int_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vmulltq_int_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULLTT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vmulq[_f16]( | a -> Qnb -> Qm | VMUL.F16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vmulq[_f32]( | a -> Qnb -> Qm | VMUL.F32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vmulq[_n_f16]( | a -> Qnb -> Rm | VMUL.F16 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vmulq[_n_f32]( | a -> Qnb -> Rm | VMUL.F32 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vmulq[_s8]( | a -> Qnb -> Qm | VMUL.I8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vmulq[_s16]( | a -> Qnb -> Qm | VMUL.I16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vmulq[_s32]( | a -> Qnb -> Qm | VMUL.I32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vmulq[_n_s8]( | a -> Qnb -> Rm | VMUL.I8 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vmulq[_n_s16]( | a -> Qnb -> Rm | VMUL.I16 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vmulq[_n_s32]( | a -> Qnb -> Rm | VMUL.I32 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vmulq[_u8]( | a -> Qnb -> Qm | VMUL.I8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vmulq[_u16]( | a -> Qnb -> Qm | VMUL.I16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vmulq[_u32]( | a -> Qnb -> Qm | VMUL.I32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vmulq[_n_u8]( | a -> Qnb -> Rm | VMUL.I8 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vmulq[_n_u16]( | a -> Qnb -> Rm | VMUL.I16 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vmulq[_n_u32]( | a -> Qnb -> Rm | VMUL.I32 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vmulq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vmulq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vmulq_m[_n_f16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vmulq_m[_n_f32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmulq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmulq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vmulq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vmulq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vmulq_x[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vmulq_x[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmulq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmulq_x[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmulq_x[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmulq_x[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmulq_x[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmulq_x[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmulq_x[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVMULT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrmulhq[_s8]( | a -> Qnb -> Qm | VRMULH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrmulhq[_s16]( | a -> Qnb -> Qm | VRMULH.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrmulhq[_s32]( | a -> Qnb -> Qm | VRMULH.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrmulhq[_u8]( | a -> Qnb -> Qm | VRMULH.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrmulhq[_u16]( | a -> Qnb -> Qm | VRMULH.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrmulhq[_u32]( | a -> Qnb -> Qm | VRMULH.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrmulhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrmulhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrmulhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrmulhq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrmulhq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrmulhq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrmulhq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrmulhq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrmulhq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrmulhq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrmulhq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrmulhq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMULHT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqdmladhq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQDMLADH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmladhq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQDMLADH.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmladhq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQDMLADH.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmladhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLADHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmladhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLADHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmladhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLADHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmladhxq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQDMLADHX.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmladhxq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQDMLADHX.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmladhxq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQDMLADHX.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmladhxq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLADHXT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmladhxq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLADHXT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmladhxq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLADHXT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmladhq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQRDMLADH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmladhq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQRDMLADH.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmladhq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQRDMLADH.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmladhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLADHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmladhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLADHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmladhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLADHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmladhxq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQRDMLADHX.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmladhxq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQRDMLADHX.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmladhxq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQRDMLADHX.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmladhxq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLADHXT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmladhxq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLADHXT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmladhxq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLADHXT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmlahq[_n_s8]( | add -> Qdam1 -> Qnm2 -> Rm | VQDMLAH.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqdmlahq[_n_s16]( | add -> Qdam1 -> Qnm2 -> Rm | VQDMLAH.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqdmlahq[_n_s32]( | add -> Qdam1 -> Qnm2 -> Rm | VQDMLAH.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqdmlahq_m[_n_s8]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVQDMLAHT.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqdmlahq_m[_n_s16]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVQDMLAHT.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqdmlahq_m[_n_s32]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVQDMLAHT.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqrdmlahq[_n_s8]( | add -> Qdam1 -> Qnm2 -> Rm | VQRDMLAH.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqrdmlahq[_n_s16]( | add -> Qdam1 -> Qnm2 -> Rm | VQRDMLAH.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqrdmlahq[_n_s32]( | add -> Qdam1 -> Qnm2 -> Rm | VQRDMLAH.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqrdmlahq_m[_n_s8]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMLAHT.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqrdmlahq_m[_n_s16]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMLAHT.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqrdmlahq_m[_n_s32]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMLAHT.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqdmlashq[_n_s8]( | m1 -> Qdam2 -> Qnadd -> Rm | VQDMLASH.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqdmlashq[_n_s16]( | m1 -> Qdam2 -> Qnadd -> Rm | VQDMLASH.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqdmlashq[_n_s32]( | m1 -> Qdam2 -> Qnadd -> Rm | VQDMLASH.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqdmlashq_m[_n_s8]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVQDMLASHT.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqdmlashq_m[_n_s16]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVQDMLASHT.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqdmlashq_m[_n_s32]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVQDMLASHT.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqrdmlashq[_n_s8]( | m1 -> Qdam2 -> Qnadd -> Rm | VQRDMLASH.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqrdmlashq[_n_s16]( | m1 -> Qdam2 -> Qnadd -> Rm | VQRDMLASH.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqrdmlashq[_n_s32]( | m1 -> Qdam2 -> Qnadd -> Rm | VQRDMLASH.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqrdmlashq_m[_n_s8]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMLASHT.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqrdmlashq_m[_n_s16]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMLASHT.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqrdmlashq_m[_n_s32]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMLASHT.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqdmlsdhq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQDMLSDH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmlsdhq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQDMLSDH.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmlsdhq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQDMLSDH.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmlsdhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLSDHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmlsdhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLSDHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmlsdhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLSDHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmlsdhxq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQDMLSDHX.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmlsdhxq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQDMLSDHX.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmlsdhxq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQDMLSDHX.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmlsdhxq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLSDHXT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmlsdhxq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLSDHXT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmlsdhxq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMLSDHXT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmlsdhq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQRDMLSDH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmlsdhq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQRDMLSDH.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmlsdhq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQRDMLSDH.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmlsdhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLSDHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmlsdhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLSDHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmlsdhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLSDHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmlsdhxq[_s8]( | inactive -> Qda -> Qnb -> Qm | VQRDMLSDHX.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmlsdhxq[_s16]( | inactive -> Qda -> Qnb -> Qm | VQRDMLSDHX.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmlsdhxq[_s32]( | inactive -> Qda -> Qnb -> Qm | VQRDMLSDHX.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmlsdhxq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLSDHXT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmlsdhxq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLSDHXT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmlsdhxq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMLSDHXT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqdmulhq[_n_s8]( | a -> Qnb -> Rm | VQDMULH.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmulhq[_n_s16]( | a -> Qnb -> Rm | VQDMULH.S16 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqdmulhq[_n_s32]( | a -> Qnb -> Rm | VQDMULH.S32 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqdmulhq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULHT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmulhq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULHT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmulhq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULHT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqdmulhq[_s8]( | a -> Qnb -> Qm | VQDMULH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmulhq[_s16]( | a -> Qnb -> Qm | VQDMULH.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqdmulhq[_s32]( | a -> Qnb -> Qm | VQDMULH.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqdmulhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqdmulhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmulhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmulhq[_n_s8]( | a -> Qnb -> Rm | VQRDMULH.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmulhq[_n_s16]( | a -> Qnb -> Rm | VQRDMULH.S16 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqrdmulhq[_n_s32]( | a -> Qnb -> Rm | VQRDMULH.S32 Qd, Qn, Rm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqrdmulhq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMULHT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmulhq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMULHT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmulhq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQRDMULHT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrdmulhq[_s8]( | a -> Qnb -> Qm | VQRDMULH.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmulhq[_s16]( | a -> Qnb -> Qm | VQRDMULH.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqrdmulhq[_s32]( | a -> Qnb -> Qm | VQRDMULH.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqrdmulhq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMULHT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrdmulhq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMULHT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrdmulhq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQRDMULHT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmullbq[_n_s16]( | a -> Qnb -> Rm | VQDMULLB.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmullbq[_n_s32]( | a -> Qnb -> Rm | VQDMULLB.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmullbq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULLBT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmullbq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULLBT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmullbq[_s16]( | a -> Qnb -> Qm | VQDMULLB.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmullbq[_s32]( | a -> Qnb -> Qm | VQDMULLB.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmullbq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULLBT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmullbq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULLBT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmulltq[_n_s16]( | a -> Qnb -> Rm | VQDMULLT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmulltq[_n_s32]( | a -> Qnb -> Rm | VQDMULLT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmulltq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULLTT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmulltq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQDMULLTT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmulltq[_s16]( | a -> Qnb -> Qm | VQDMULLT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmulltq[_s32]( | a -> Qnb -> Qm | VQDMULLT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqdmulltq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULLTT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vqdmulltq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQDMULLTT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int32_t [__arm_]vmladavaq[_s8]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVA.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaq[_s16]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVA.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaq[_s32]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVA.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavaq[_u8]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVA.U8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavaq[_u16]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVA.U16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavaq[_u32]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVA.U32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaq_p[_s8]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaq_p[_s16]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaq_p[_s32]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavaq_p[_u8]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAT.U8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavaq_p[_u16]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAT.U16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavaq_p[_u32]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAT.U32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavq[_s8]( | m1 -> Qnm2 -> Qm | VMLADAV.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavq[_s16]( | m1 -> Qnm2 -> Qm | VMLADAV.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavq[_s32]( | m1 -> Qnm2 -> Qm | VMLADAV.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavq[_u8]( | m1 -> Qnm2 -> Qm | VMLADAV.U8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavq[_u16]( | m1 -> Qnm2 -> Qm | VMLADAV.U16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavq[_u32]( | m1 -> Qnm2 -> Qm | VMLADAV.U32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavq_p[_s8]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavq_p[_s16]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavq_p[_s32]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavq_p[_u8]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVT.U8 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavq_p[_u16]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVT.U16 Rda, Qn, Qm | Rda -> result | MVE | 
| uint32_t [__arm_]vmladavq_p[_u32]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVT.U32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaxq[_s8]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVAX.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaxq[_s16]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVAX.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaxq[_s32]( | add -> Rdam1 -> Qnm2 -> Qm | VMLADAVAX.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaxq_p[_s8]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAXT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaxq_p[_s16]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAXT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavaxq_p[_s32]( | add -> Rdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVAXT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavxq[_s8]( | m1 -> Qnm2 -> Qm | VMLADAVX.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavxq[_s16]( | m1 -> Qnm2 -> Qm | VMLADAVX.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavxq[_s32]( | m1 -> Qnm2 -> Qm | VMLADAVX.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavxq_p[_s8]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVXT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavxq_p[_s16]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVXT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmladavxq_p[_s32]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLADAVXT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int64_t [__arm_]vmlaldavaq[_s16]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qm | VMLALDAVA.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaq[_s32]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qm | VMLALDAVA.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavaq[_u16]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qm | VMLALDAVA.U16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavaq[_u32]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qm | VMLALDAVA.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaq_p[_s16]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVAT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaq_p[_s32]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVAT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavaq_p[_u16]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVAT.U16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavaq_p[_u32]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVAT.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavq[_s16]( | m1 -> Qnm2 -> Qm | VMLALDAV.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavq[_s32]( | m1 -> Qnm2 -> Qm | VMLALDAV.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavq[_u16]( | m1 -> Qnm2 -> Qm | VMLALDAV.U16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavq[_u32]( | m1 -> Qnm2 -> Qm | VMLALDAV.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavq_p[_s16]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavq_p[_s32]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavq_p[_u16]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVT.U16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vmlaldavq_p[_u32]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVT.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaxq[_s16]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qm | VMLALDAVAX.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaxq[_s32]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qm | VMLALDAVAX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaxq_p[_s16]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVAXT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavaxq_p[_s32]( | add -> [RdaHi,RdaLo]m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVAXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavxq[_s16]( | m1 -> Qnm2 -> Qm | VMLALDAVX.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavxq[_s32]( | m1 -> Qnm2 -> Qm | VMLALDAVX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavxq_p[_s16]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVXT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlaldavxq_p[_s32]( | m1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVMLALDAVXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int8x16_t [__arm_]vmlaq[_n_s8]( | add -> Qdam1 -> Qnm2 -> Rm | VMLA.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vmlaq[_n_s16]( | add -> Qdam1 -> Qnm2 -> Rm | VMLA.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vmlaq[_n_s32]( | add -> Qdam1 -> Qnm2 -> Rm | VMLA.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vmlaq[_n_u8]( | add -> Qdam1 -> Qnm2 -> Rm | VMLA.U8 Qda, Qn, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vmlaq[_n_u16]( | add -> Qdam1 -> Qnm2 -> Rm | VMLA.U16 Qda, Qn, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vmlaq[_n_u32]( | add -> Qdam1 -> Qnm2 -> Rm | VMLA.U32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vmlaq_m[_n_s8]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVMLAT.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vmlaq_m[_n_s16]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVMLAT.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vmlaq_m[_n_s32]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVMLAT.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vmlaq_m[_n_u8]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVMLAT.U8 Qda, Qn, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vmlaq_m[_n_u16]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVMLAT.U16 Qda, Qn, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vmlaq_m[_n_u32]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVMLAT.U32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vmlasq[_n_s8]( | m1 -> Qdam2 -> Qnadd -> Rm | VMLAS.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vmlasq[_n_s16]( | m1 -> Qdam2 -> Qnadd -> Rm | VMLAS.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vmlasq[_n_s32]( | m1 -> Qdam2 -> Qnadd -> Rm | VMLAS.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vmlasq[_n_u8]( | m1 -> Qdam2 -> Qnadd -> Rm | VMLAS.U8 Qda, Qn, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vmlasq[_n_u16]( | m1 -> Qdam2 -> Qnadd -> Rm | VMLAS.U16 Qda, Qn, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vmlasq[_n_u32]( | m1 -> Qdam2 -> Qnadd -> Rm | VMLAS.U32 Qda, Qn, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vmlasq_m[_n_s8]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVMLAST.S8 Qda, Qn, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vmlasq_m[_n_s16]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVMLAST.S16 Qda, Qn, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vmlasq_m[_n_s32]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVMLAST.S32 Qda, Qn, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vmlasq_m[_n_u8]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVMLAST.U8 Qda, Qn, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vmlasq_m[_n_u16]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVMLAST.U16 Qda, Qn, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vmlasq_m[_n_u32]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVMLAST.U32 Qda, Qn, Rm | Qda -> result | MVE | 
| int32_t [__arm_]vmlsdavaq[_s8]( | a -> Rdab -> Qnc -> Qm | VMLSDAVA.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaq[_s16]( | a -> Rdab -> Qnc -> Qm | VMLSDAVA.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaq[_s32]( | a -> Rdab -> Qnc -> Qm | VMLSDAVA.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaq_p[_s8]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVAT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaq_p[_s16]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVAT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaq_p[_s32]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVAT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavq[_s8]( | a -> Qnb -> Qm | VMLSDAV.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavq[_s16]( | a -> Qnb -> Qm | VMLSDAV.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavq[_s32]( | a -> Qnb -> Qm | VMLSDAV.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavq_p[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavq_p[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaxq[_s8]( | a -> Rdab -> Qnc -> Qm | VMLSDAVAX.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaxq[_s16]( | a -> Rdab -> Qnc -> Qm | VMLSDAVAX.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaxq[_s32]( | a -> Rdab -> Qnc -> Qm | VMLSDAVAX.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaxq_p[_s8]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVAXT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaxq_p[_s16]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVAXT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavaxq_p[_s32]( | a -> Rdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVAXT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavxq[_s8]( | a -> Qnb -> Qm | VMLSDAVX.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavxq[_s16]( | a -> Qnb -> Qm | VMLSDAVX.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavxq[_s32]( | a -> Qnb -> Qm | VMLSDAVX.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavxq_p[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVXT.S8 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavxq_p[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVXT.S16 Rda, Qn, Qm | Rda -> result | MVE | 
| int32_t [__arm_]vmlsdavxq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSDAVXT.S32 Rda, Qn, Qm | Rda -> result | MVE | 
| int64_t [__arm_]vmlsldavaq[_s16]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VMLSLDAVA.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaq[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VMLSLDAVA.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaq_p[_s16]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVAT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVAT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavq[_s16]( | a -> Qnb -> Qm | VMLSLDAV.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavq[_s32]( | a -> Qnb -> Qm | VMLSLDAV.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavq_p[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaxq[_s16]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VMLSLDAVAX.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaxq[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VMLSLDAVAX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaxq_p[_s16]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVAXT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavaxq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVAXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavxq[_s16]( | a -> Qnb -> Qm | VMLSLDAVX.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavxq[_s32]( | a -> Qnb -> Qm | VMLSLDAVX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavxq_p[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVXT.S16 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vmlsldavxq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVMLSLDAVXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhaq[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VRMLALDAVHA.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vrmlaldavhaq[_u32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VRMLALDAVHA.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhaq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVRMLALDAVHAT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vrmlaldavhaq_p[_u32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVRMLALDAVHAT.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhq[_s32]( | a -> Qnb -> Qm | VRMLALDAVH.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vrmlaldavhq[_u32]( | a -> Qnb -> Qm | VRMLALDAVH.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMLALDAVHT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]vrmlaldavhq_p[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMLALDAVHT.U32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhaxq[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VRMLALDAVHAX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhaxq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVRMLALDAVHAXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhxq[_s32]( | a -> Qnb -> Qm | VRMLALDAVHX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlaldavhxq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMLALDAVHXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhaq[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VRMLSLDAVHA.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhaq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVRMLSLDAVHAT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhq[_s32]( | a -> Qnb -> Qm | VRMLSLDAVH.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMLSLDAVHT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhaxq[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qm | VRMLSLDAVHAX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhaxq_p[_s32]( | a -> [RdaHi,RdaLo]b -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVRMLSLDAVHAXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhxq[_s32]( | a -> Qnb -> Qm | VRMLSLDAVHX.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]vrmlsldavhxq_p[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVRMLSLDAVHXT.S32 RdaLo, RdaHi, Qn, Qm | [RdaHi,RdaLo] -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vfmaq[_n_f16]( | add -> Qdam1 -> Qnm2 -> Rm | VFMA.F16 Qda, Qn, Rm | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vfmaq[_n_f32]( | add -> Qdam1 -> Qnm2 -> Rm | VFMA.F32 Qda, Qn, Rm | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vfmaq_m[_n_f16]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVFMAT.F16 Qda, Qn, Rm | Qda -> result | MVE | 
| float32x4_t [__arm_]vfmaq_m[_n_f32]( | add -> Qdam1 -> Qnm2 -> Rmp -> Rp | VMSR P0, RpVPSTVFMAT.F32 Qda, Qn, Rm | Qda -> result | MVE | 
| float16x8_t [__arm_]vfmaq[_f16]( | add -> Qdam1 -> Qnm2 -> Qm | VFMA.F16 Qda, Qn, Qm | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vfmaq[_f32]( | add -> Qdam1 -> Qnm2 -> Qm | VFMA.F32 Qda, Qn, Qm | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vfmaq_m[_f16]( | add -> Qdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVFMAT.F16 Qda, Qn, Qm | Qda -> result | MVE | 
| float32x4_t [__arm_]vfmaq_m[_f32]( | add -> Qdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVFMAT.F32 Qda, Qn, Qm | Qda -> result | MVE | 
| float16x8_t [__arm_]vfmasq[_n_f16]( | m1 -> Qdam2 -> Qnadd -> Rm | VFMAS.F16 Qda, Qn, Rm | Qda -> result | MVE | 
| float32x4_t [__arm_]vfmasq[_n_f32]( | m1 -> Qdam2 -> Qnadd -> Rm | VFMAS.F32 Qda, Qn, Rm | Qda -> result | MVE | 
| float16x8_t [__arm_]vfmasq_m[_n_f16]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVFMAST.F16 Qda, Qn, Rm | Qda -> result | MVE | 
| float32x4_t [__arm_]vfmasq_m[_n_f32]( | m1 -> Qdam2 -> Qnadd -> Rmp -> Rp | VMSR P0, RpVPSTVFMAST.F32 Qda, Qn, Rm | Qda -> result | MVE | 
| float16x8_t [__arm_]vfmsq[_f16]( | add -> Qdam1 -> Qnm2 -> Qm | VFMS.F16 Qda, Qn, Qm | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vfmsq[_f32]( | add -> Qdam1 -> Qnm2 -> Qm | VFMS.F32 Qda, Qn, Qm | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vfmsq_m[_f16]( | add -> Qdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVFMST.F16 Qda, Qn, Qm | Qda -> result | MVE | 
| float32x4_t [__arm_]vfmsq_m[_f32]( | add -> Qdam1 -> Qnm2 -> Qmp -> Rp | VMSR P0, RpVPSTVFMST.F32 Qda, Qn, Qm | Qda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int32x4_t [__arm_]vsbciq[_s32]( | a -> Qnb -> Qm | VSBCI.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| uint32x4_t [__arm_]vsbciq[_u32]( | a -> Qnb -> Qm | VSBCI.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| int32x4_t [__arm_]vsbciq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSBCIT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| uint32x4_t [__arm_]vsbciq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSBCIT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry_out | MVE | 
| int32x4_t [__arm_]vsbcq[_s32]( | a -> Qnb -> Qm*carry -> Rt | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVSBC.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| uint32x4_t [__arm_]vsbcq[_u32]( | a -> Qnb -> Qm*carry -> Rt | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVSBC.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| int32x4_t [__arm_]vsbcq_m[_s32]( | inactive -> Qda -> Qnb -> Qm*carry -> Rtp -> Rp | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVMSR P0, RpVPSTVSBCT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| uint32x4_t [__arm_]vsbcq_m[_u32]( | inactive -> Qda -> Qnb -> Qm*carry -> Rtp -> Rp | VMRS Rs, FPSCR_nzcvqcBFI Rs, Rt, #29, #1VMSR FPSCR_nzcvqc, RsVMSR P0, RpVPSTVSBCT.I32 Qd, Qn, QmVMRS Rt, FPSCR_nzcvqcLSR Rt, #29AND Rt, #1 | Qd -> resultRt -> *carry | MVE | 
| int8x16_t [__arm_]vsubq[_s8]( | a -> Qnb -> Qm | VSUB.I8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vsubq[_s16]( | a -> Qnb -> Qm | VSUB.I16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vsubq[_s32]( | a -> Qnb -> Qm | VSUB.I32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vsubq[_n_s8]( | a -> Qnb -> Rm | VSUB.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsubq[_n_s16]( | a -> Qnb -> Rm | VSUB.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsubq[_n_s32]( | a -> Qnb -> Rm | VSUB.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsubq[_u8]( | a -> Qnb -> Qm | VSUB.I8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vsubq[_u16]( | a -> Qnb -> Qm | VSUB.I16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vsubq[_u32]( | a -> Qnb -> Qm | VSUB.I32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vsubq[_n_u8]( | a -> Qnb -> Rm | VSUB.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsubq[_n_u16]( | a -> Qnb -> Rm | VSUB.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsubq[_n_u32]( | a -> Qnb -> Rm | VSUB.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vsubq[_f16]( | a -> Qnb -> Qm | VSUB.F16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vsubq[_f32]( | a -> Qnb -> Qm | VSUB.F32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vsubq[_n_f16]( | a -> Qnb -> Rm | VSUB.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vsubq[_n_f32]( | a -> Qnb -> Rm | VSUB.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vsubq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsubq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsubq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vsubq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsubq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsubq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsubq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsubq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsubq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsubq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsubq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsubq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vsubq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vsubq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vsubq_m[_n_f16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vsubq_m[_n_f32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vsubq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsubq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsubq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vsubq_x[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsubq_x[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsubq_x[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsubq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsubq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsubq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsubq_x[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsubq_x[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsubq_x[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.I32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vsubq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.F16 Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vsubq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVSUBT.F32 Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vsubq_x[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.F16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vsubq_x[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVSUBT.F32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhsubq[_n_s8]( | a -> Qnb -> Rm | VHSUB.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhsubq[_n_s16]( | a -> Qnb -> Rm | VHSUB.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhsubq[_n_s32]( | a -> Qnb -> Rm | VHSUB.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhsubq[_n_u8]( | a -> Qnb -> Rm | VHSUB.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhsubq[_n_u16]( | a -> Qnb -> Rm | VHSUB.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhsubq[_n_u32]( | a -> Qnb -> Rm | VHSUB.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhsubq[_s8]( | a -> Qnb -> Qm | VHSUB.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vhsubq[_s16]( | a -> Qnb -> Qm | VHSUB.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vhsubq[_s32]( | a -> Qnb -> Qm | VHSUB.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vhsubq[_u8]( | a -> Qnb -> Qm | VHSUB.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vhsubq[_u16]( | a -> Qnb -> Qm | VHSUB.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vhsubq[_u32]( | a -> Qnb -> Qm | VHSUB.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vhsubq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhsubq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhsubq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhsubq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhsubq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhsubq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhsubq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhsubq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhsubq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhsubq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhsubq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhsubq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhsubq_x[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhsubq_x[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhsubq_x[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhsubq_x[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhsubq_x[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhsubq_x[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVHSUBT.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vhsubq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vhsubq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vhsubq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vhsubq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vhsubq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vhsubq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHSUBT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqsubq[_n_s8]( | a -> Qnb -> Rm | VQSUB.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqsubq[_n_s16]( | a -> Qnb -> Rm | VQSUB.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqsubq[_n_s32]( | a -> Qnb -> Rm | VQSUB.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqsubq[_n_u8]( | a -> Qnb -> Rm | VQSUB.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqsubq[_n_u16]( | a -> Qnb -> Rm | VQSUB.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqsubq[_n_u32]( | a -> Qnb -> Rm | VQSUB.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqsubq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQSUBT.S8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqsubq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQSUBT.S16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqsubq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQSUBT.S32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqsubq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQSUBT.U8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqsubq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQSUBT.U16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqsubq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVQSUBT.U32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqsubq[_s8]( | a -> Qnb -> Qm | VQSUB.S8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqsubq[_s16]( | a -> Qnb -> Qm | VQSUB.S16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqsubq[_s32]( | a -> Qnb -> Qm | VQSUB.S32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vqsubq[_u8]( | a -> Qnb -> Qm | VQSUB.U8 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vqsubq[_u16]( | a -> Qnb -> Qm | VQSUB.U16 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vqsubq[_u32]( | a -> Qnb -> Qm | VQSUB.U32 Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqsubq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQSUBT.S8 Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqsubq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQSUBT.S16 Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqsubq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQSUBT.S32 Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqsubq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQSUBT.U8 Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqsubq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQSUBT.U16 Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqsubq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVQSUBT.U32 Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vrndq[_f16](float16x8_t a) | a -> Qm | VRINTZ.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndq[_f32](float32x4_t a) | a -> Qm | VRINTZ.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrndq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTZT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTZT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTZT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTZT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndnq[_f16](float16x8_t a) | a -> Qm | VRINTN.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndnq[_f32](float32x4_t a) | a -> Qm | VRINTN.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrndnq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTNT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndnq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTNT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndnq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTNT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndnq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTNT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndmq[_f16](float16x8_t a) | a -> Qm | VRINTM.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndmq[_f32](float32x4_t a) | a -> Qm | VRINTM.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrndmq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTMT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndmq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTMT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndmq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTMT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndmq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTMT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndpq[_f16](float16x8_t a) | a -> Qm | VRINTP.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndpq[_f32](float32x4_t a) | a -> Qm | VRINTP.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrndpq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTPT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndpq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTPT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndpq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTPT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndpq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTPT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndaq[_f16](float16x8_t a) | a -> Qm | VRINTA.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndaq[_f32](float32x4_t a) | a -> Qm | VRINTA.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrndaq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTAT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndaq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTAT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndaq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTAT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndaq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTAT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndxq[_f16](float16x8_t a) | a -> Qm | VRINTX.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndxq[_f32](float32x4_t a) | a -> Qm | VRINTX.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vrndxq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTXT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndxq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVRINTXT.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vrndxq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTXT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vrndxq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVRINTXT.F32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vclsq[_s8](int8x16_t a) | a -> Qm | VCLS.S8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vclsq[_s16](int16x8_t a) | a -> Qm | VCLS.S16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vclsq[_s32](int32x4_t a) | a -> Qm | VCLS.S32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vclsq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLST.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vclsq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLST.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vclsq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLST.S32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vclsq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLST.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vclsq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLST.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vclsq_x[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLST.S32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vclzq[_s8](int8x16_t a) | a -> Qm | VCLZ.I8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vclzq[_s16](int16x8_t a) | a -> Qm | VCLZ.I16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vclzq[_s32](int32x4_t a) | a -> Qm | VCLZ.I32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vclzq[_u8](uint8x16_t a) | a -> Qm | VCLZ.I8 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vclzq[_u16](uint16x8_t a) | a -> Qm | VCLZ.I16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vclzq[_u32](uint32x4_t a) | a -> Qm | VCLZ.I32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vclzq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vclzq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vclzq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vclzq_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I8 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vclzq_m[_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vclzq_m[_u32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vclzq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vclzq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vclzq_x[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vclzq_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I8 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vclzq_x[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vclzq_x[_u32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCLZT.I32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vbicq[_s8]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vbicq[_s16]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vbicq[_s32]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vbicq[_u8]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vbicq[_u16]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vbicq[_u32]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vbicq[_f16]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vbicq[_f32]( | a -> Qnb -> Qm | VBIC Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vbicq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vbicq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vbicq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vbicq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vbicq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vbicq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vbicq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vbicq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vbicq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vbicq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vbicq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vbicq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vbicq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vbicq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vbicq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vbicq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVBICT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vbicq[_n_s16]( | a -> Qdaimm in AdvSIMDExpandImm | VBIC.I16 Qda, #imm | Qda -> result | MVE | 
| int32x4_t [__arm_]vbicq[_n_s32]( | a -> Qdaimm in AdvSIMDExpandImm | VBIC.I32 Qda, #imm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vbicq[_n_u16]( | a -> Qdaimm in AdvSIMDExpandImm | VBIC.I16 Qda, #imm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vbicq[_n_u32]( | a -> Qdaimm in AdvSIMDExpandImm | VBIC.I32 Qda, #imm | Qda -> result | MVE | 
| int16x8_t [__arm_]vbicq_m_n[_s16]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVBICT.I16 Qda, #imm | Qda -> result | MVE | 
| int32x4_t [__arm_]vbicq_m_n[_s32]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVBICT.I32 Qda, #imm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vbicq_m_n[_u16]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVBICT.I16 Qda, #imm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vbicq_m_n[_u32]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVBICT.I32 Qda, #imm | Qda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vnegq[_f16](float16x8_t a) | a -> Qm | VNEG.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vnegq[_f32](float32x4_t a) | a -> Qm | VNEG.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vnegq[_s8](int8x16_t a) | a -> Qm | VNEG.S8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vnegq[_s16](int16x8_t a) | a -> Qm | VNEG.S16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vnegq[_s32](int32x4_t a) | a -> Qm | VNEG.S32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vnegq_m[_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vnegq_m[_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.F32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vnegq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vnegq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vnegq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.S32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vnegq_x[_f16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vnegq_x[_f32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.F32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vnegq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vnegq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vnegq_x[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVNEGT.S32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqnegq[_s8](int8x16_t a) | a -> Qm | VQNEG.S8 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqnegq[_s16](int16x8_t a) | a -> Qm | VQNEG.S16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqnegq[_s32](int32x4_t a) | a -> Qm | VQNEG.S32 Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqnegq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVQNEGT.S8 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqnegq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVQNEGT.S16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqnegq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVQNEGT.S32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vandq[_s8]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vandq[_s16]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vandq[_s32]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vandq[_u8]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vandq[_u16]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vandq[_u32]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vandq[_f16]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vandq[_f32]( | a -> Qnb -> Qm | VAND Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vandq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vandq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vandq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vandq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vandq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vandq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vandq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vandq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vandq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vandq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vandq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vandq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vandq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vandq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vandq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vandq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVANDT Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]veorq[_s8]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]veorq[_s16]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]veorq[_s32]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]veorq[_u8]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]veorq[_u16]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]veorq[_u32]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]veorq[_f16]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]veorq[_f32]( | a -> Qnb -> Qm | VEOR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]veorq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]veorq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]veorq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]veorq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]veorq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]veorq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]veorq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]veorq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]veorq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]veorq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]veorq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]veorq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]veorq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]veorq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]veorq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]veorq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVEORT Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vmvnq[_s8](int8x16_t a) | a -> Qm | VMVN Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vmvnq[_s16](int16x8_t a) | a -> Qm | VMVN Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vmvnq[_s32](int32x4_t a) | a -> Qm | VMVN Qd, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vmvnq[_u8](uint8x16_t a) | a -> Qm | VMVN Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vmvnq[_u16](uint16x8_t a) | a -> Qm | VMVN Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vmvnq[_u32](uint32x4_t a) | a -> Qm | VMVN Qd, Qm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vmvnq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmvnq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmvnq_m[_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmvnq_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmvnq_m[_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmvnq_m[_u32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmvnq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmvnq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmvnq_x[_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmvnq_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmvnq_x[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmvnq_x[_u32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMVNT Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmvnq_n_s16(const int16_t imm) | imm in AdvSIMDExpandImm | VMVN.I16 Qd, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmvnq_n_s32(const int32_t imm) | imm in AdvSIMDExpandImm | VMVN.I32 Qd, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmvnq_n_u16(const uint16_t imm) | imm in AdvSIMDExpandImm | VMVN.I16 Qd, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmvnq_n_u32(const uint32_t imm) | imm in AdvSIMDExpandImm | VMVN.I32 Qd, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmvnq_m[_n_s16]( | inactive -> Qdimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I16 Qd, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmvnq_m[_n_s32]( | inactive -> Qdimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I32 Qd, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmvnq_m[_n_u16]( | inactive -> Qdimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I16 Qd, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmvnq_m[_n_u32]( | inactive -> Qdimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I32 Qd, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmvnq_x_n_s16( | imm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I16 Qd, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmvnq_x_n_s32( | imm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I32 Qd, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmvnq_x_n_u16( | imm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I16 Qd, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmvnq_x_n_u32( | imm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVMVNT.I32 Qd, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vornq[_f16]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vornq[_f32]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vornq[_s8]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vornq[_s16]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vornq[_s32]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vornq[_u8]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vornq[_u16]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vornq[_u32]( | a -> Qnb -> Qm | VORN Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vornq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vornq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vornq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vornq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vornq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vornq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vornq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vornq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vornq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vornq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vornq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vornq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vornq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vornq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vornq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vornq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORNT Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vorrq[_f16]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vorrq[_f32]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vorrq[_s8]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vorrq[_s16]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vorrq[_s32]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vorrq[_u8]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vorrq[_u16]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vorrq[_u32]( | a -> Qnb -> Qm | VORR Qd, Qn, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vorrq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vorrq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vorrq_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vorrq_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vorrq_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vorrq_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vorrq_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vorrq_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vorrq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vorrq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vorrq_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vorrq_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vorrq_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vorrq_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vorrq_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vorrq_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVORRT Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vorrq[_n_s16]( | a -> Qdaimm in AdvSIMDExpandImm | VORR.I16 Qda, #imm | Qda -> result | MVE | 
| int32x4_t [__arm_]vorrq[_n_s32]( | a -> Qdaimm in AdvSIMDExpandImm | VORR.I32 Qda, #imm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vorrq[_n_u16]( | a -> Qdaimm in AdvSIMDExpandImm | VORR.I16 Qda, #imm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vorrq[_n_u32]( | a -> Qdaimm in AdvSIMDExpandImm | VORR.I32 Qda, #imm | Qda -> result | MVE | 
| int16x8_t [__arm_]vorrq_m_n[_s16]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVORRT.I16 Qda, #imm | Qda -> result | MVE | 
| int32x4_t [__arm_]vorrq_m_n[_s32]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVORRT.I32 Qda, #imm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vorrq_m_n[_u16]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVORRT.I16 Qda, #imm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vorrq_m_n[_u32]( | a -> Qdaimm in AdvSIMDExpandImmp -> Rp | VMSR P0, RpVPSTVORRT.I32 Qda, #imm | Qda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vcaddq_rot90[_f16]( | a -> Qnb -> Qm | VCADD.F16 Qd, Qn, Qm, #90 | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vcaddq_rot90[_f32]( | a -> Qnb -> Qm | VCADD.F32 Qd, Qn, Qm, #90 | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vcaddq_rot90[_s8]( | a -> Qnb -> Qm | VCADD.I8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcaddq_rot90[_s16]( | a -> Qnb -> Qm | VCADD.I16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcaddq_rot90[_s32]( | a -> Qnb -> Qm | VCADD.I32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcaddq_rot90[_u8]( | a -> Qnb -> Qm | VCADD.I8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcaddq_rot90[_u16]( | a -> Qnb -> Qm | VCADD.I16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcaddq_rot90[_u32]( | a -> Qnb -> Qm | VCADD.I32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcaddq_rot270[_f16]( | a -> Qnb -> Qm | VCADD.F16 Qd, Qn, Qm, #270 | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vcaddq_rot270[_f32]( | a -> Qnb -> Qm | VCADD.F32 Qd, Qn, Qm, #270 | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vcaddq_rot270[_s8]( | a -> Qnb -> Qm | VCADD.I8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcaddq_rot270[_s16]( | a -> Qnb -> Qm | VCADD.I16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcaddq_rot270[_s32]( | a -> Qnb -> Qm | VCADD.I32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcaddq_rot270[_u8]( | a -> Qnb -> Qm | VCADD.I8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcaddq_rot270[_u16]( | a -> Qnb -> Qm | VCADD.I16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcaddq_rot270[_u32]( | a -> Qnb -> Qm | VCADD.I32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcaddq_rot90_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcaddq_rot90_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int8x16_t [__arm_]vcaddq_rot90_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcaddq_rot90_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcaddq_rot90_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcaddq_rot90_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcaddq_rot90_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcaddq_rot90_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcaddq_rot270_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcaddq_rot270_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int8x16_t [__arm_]vcaddq_rot270_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcaddq_rot270_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcaddq_rot270_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcaddq_rot270_m[_u8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcaddq_rot270_m[_u16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcaddq_rot270_m[_u32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcaddq_rot90_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcaddq_rot90_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int8x16_t [__arm_]vcaddq_rot90_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcaddq_rot90_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcaddq_rot90_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcaddq_rot90_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcaddq_rot90_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcaddq_rot90_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcaddq_rot270_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcaddq_rot270_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.F32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int8x16_t [__arm_]vcaddq_rot270_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcaddq_rot270_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcaddq_rot270_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint8x16_t [__arm_]vcaddq_rot270_x[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcaddq_rot270_x[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcaddq_rot270_x[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCADDT.I32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int8x16_t [__arm_]vhcaddq_rot90[_s8]( | a -> Qnb -> Qm | VHCADD.S8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int16x8_t [__arm_]vhcaddq_rot90[_s16]( | a -> Qnb -> Qm | VHCADD.S16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int32x4_t [__arm_]vhcaddq_rot90[_s32]( | a -> Qnb -> Qm | VHCADD.S32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int8x16_t [__arm_]vhcaddq_rot90_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int16x8_t [__arm_]vhcaddq_rot90_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int32x4_t [__arm_]vhcaddq_rot90_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int8x16_t [__arm_]vhcaddq_rot90_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S8 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int16x8_t [__arm_]vhcaddq_rot90_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int32x4_t [__arm_]vhcaddq_rot90_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| int8x16_t [__arm_]vhcaddq_rot270[_s8]( | a -> Qnb -> Qm | VHCADD.S8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int16x8_t [__arm_]vhcaddq_rot270[_s16]( | a -> Qnb -> Qm | VHCADD.S16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int32x4_t [__arm_]vhcaddq_rot270[_s32]( | a -> Qnb -> Qm | VHCADD.S32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int8x16_t [__arm_]vhcaddq_rot270_m[_s8]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int16x8_t [__arm_]vhcaddq_rot270_m[_s16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int32x4_t [__arm_]vhcaddq_rot270_m[_s32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int8x16_t [__arm_]vhcaddq_rot270_x[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S8 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int16x8_t [__arm_]vhcaddq_rot270_x[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| int32x4_t [__arm_]vhcaddq_rot270_x[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVHCADDT.S32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vcmlaq[_f16]( | a -> Qdab -> Qnc -> Qm | VCMLA.F16 Qda, Qn, Qm, #0 | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vcmlaq[_f32]( | a -> Qdab -> Qnc -> Qm | VCMLA.F32 Qda, Qn, Qm, #0 | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vcmlaq_rot90[_f16]( | a -> Qdab -> Qnc -> Qm | VCMLA.F16 Qda, Qn, Qm, #90 | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vcmlaq_rot90[_f32]( | a -> Qdab -> Qnc -> Qm | VCMLA.F32 Qda, Qn, Qm, #90 | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vcmlaq_rot180[_f16]( | a -> Qdab -> Qnc -> Qm | VCMLA.F16 Qda, Qn, Qm, #180 | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vcmlaq_rot180[_f32]( | a -> Qdab -> Qnc -> Qm | VCMLA.F32 Qda, Qn, Qm, #180 | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vcmlaq_rot270[_f16]( | a -> Qdab -> Qnc -> Qm | VCMLA.F16 Qda, Qn, Qm, #270 | Qda -> result | MVE/NEON | 
| float32x4_t [__arm_]vcmlaq_rot270[_f32]( | a -> Qdab -> Qnc -> Qm | VCMLA.F32 Qda, Qn, Qm, #270 | Qda -> result | MVE/NEON | 
| float16x8_t [__arm_]vcmlaq_m[_f16]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F16 Qda, Qn, Qm, #0 | Qda -> result | MVE | 
| float32x4_t [__arm_]vcmlaq_m[_f32]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F32 Qda, Qn, Qm, #0 | Qda -> result | MVE | 
| float16x8_t [__arm_]vcmlaq_rot90_m[_f16]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F16 Qda, Qn, Qm, #90 | Qda -> result | MVE | 
| float32x4_t [__arm_]vcmlaq_rot90_m[_f32]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F32 Qda, Qn, Qm, #90 | Qda -> result | MVE | 
| float16x8_t [__arm_]vcmlaq_rot180_m[_f16]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F16 Qda, Qn, Qm, #180 | Qda -> result | MVE | 
| float32x4_t [__arm_]vcmlaq_rot180_m[_f32]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F32 Qda, Qn, Qm, #180 | Qda -> result | MVE | 
| float16x8_t [__arm_]vcmlaq_rot270_m[_f16]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F16 Qda, Qn, Qm, #270 | Qda -> result | MVE | 
| float32x4_t [__arm_]vcmlaq_rot270_m[_f32]( | a -> Qdab -> Qnc -> Qmp -> Rp | VMSR P0, RpVPSTVCMLAT.F32 Qda, Qn, Qm, #270 | Qda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| float16x8_t [__arm_]vcmulq[_f16]( | a -> Qnb -> Qm | VCMUL.F16 Qd, Qn, Qm, #0 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq[_f32]( | a -> Qnb -> Qm | VCMUL.F32 Qd, Qn, Qm, #0 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot90[_f16]( | a -> Qnb -> Qm | VCMUL.F16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot90[_f32]( | a -> Qnb -> Qm | VCMUL.F32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot180[_f16]( | a -> Qnb -> Qm | VCMUL.F16 Qd, Qn, Qm, #180 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot180[_f32]( | a -> Qnb -> Qm | VCMUL.F32 Qd, Qn, Qm, #180 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot270[_f16]( | a -> Qnb -> Qm | VCMUL.F16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot270[_f32]( | a -> Qnb -> Qm | VCMUL.F32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #0 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #0 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot90_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot90_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot180_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #180 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot180_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #180 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot270_m[_f16]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot270_m[_f32]( | inactive -> Qda -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #0 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #0 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot90_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot90_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #90 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot180_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #180 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot180_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #180 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcmulq_rot270_x[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F16 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcmulq_rot270_x[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSTVCMULT.F32 Qd, Qn, Qm, #270 | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16x2_t [__arm_]vld2q[_s8](int8_t const *addr) | addr -> Rn | VLD20.8 {Qd - Qd2}, [Rn]VLD21.8 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| int16x8x2_t [__arm_]vld2q[_s16](int16_t const *addr) | addr -> Rn | VLD20.16 {Qd - Qd2}, [Rn]VLD21.16 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| int32x4x2_t [__arm_]vld2q[_s32](int32_t const *addr) | addr -> Rn | VLD20.32 {Qd - Qd2}, [Rn]VLD21.32 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| uint8x16x2_t [__arm_]vld2q[_u8](uint8_t const *addr) | addr -> Rn | VLD20.8 {Qd - Qd2}, [Rn]VLD21.8 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| uint16x8x2_t [__arm_]vld2q[_u16](uint16_t const *addr) | addr -> Rn | VLD20.16 {Qd - Qd2}, [Rn]VLD21.16 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| uint32x4x2_t [__arm_]vld2q[_u32](uint32_t const *addr) | addr -> Rn | VLD20.32 {Qd - Qd2}, [Rn]VLD21.32 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| float16x8x2_t [__arm_]vld2q[_f16](float16_t const *addr) | addr -> Rn | VLD20.16 {Qd - Qd2}, [Rn]VLD21.16 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| float32x4x2_t [__arm_]vld2q[_f32](float32_t const *addr) | addr -> Rn | VLD20.32 {Qd - Qd2}, [Rn]VLD21.32 {Qd - Qd2}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1] | MVE | 
| int8x16x4_t [__arm_]vld4q[_s8](int8_t const *addr) | addr -> Rn | VLD40.8 {Qd - Qd4}, [Rn]VLD41.8 {Qd - Qd4}, [Rn]VLD42.8 {Qd - Qd4}, [Rn]VLD43.8 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| int16x8x4_t [__arm_]vld4q[_s16](int16_t const *addr) | addr -> Rn | VLD40.16 {Qd - Qd4}, [Rn]VLD41.16 {Qd - Qd4}, [Rn]VLD42.16 {Qd - Qd4}, [Rn]VLD43.16 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| int32x4x4_t [__arm_]vld4q[_s32](int32_t const *addr) | addr -> Rn | VLD40.32 {Qd - Qd4}, [Rn]VLD41.32 {Qd - Qd4}, [Rn]VLD42.32 {Qd - Qd4}, [Rn]VLD43.32 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| uint8x16x4_t [__arm_]vld4q[_u8](uint8_t const *addr) | addr -> Rn | VLD40.8 {Qd - Qd4}, [Rn]VLD41.8 {Qd - Qd4}, [Rn]VLD42.8 {Qd - Qd4}, [Rn]VLD43.8 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| uint16x8x4_t [__arm_]vld4q[_u16](uint16_t const *addr) | addr -> Rn | VLD40.16 {Qd - Qd4}, [Rn]VLD41.16 {Qd - Qd4}, [Rn]VLD42.16 {Qd - Qd4}, [Rn]VLD43.16 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| uint32x4x4_t [__arm_]vld4q[_u32](uint32_t const *addr) | addr -> Rn | VLD40.32 {Qd - Qd4}, [Rn]VLD41.32 {Qd - Qd4}, [Rn]VLD42.32 {Qd - Qd4}, [Rn]VLD43.32 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| float16x8x4_t [__arm_]vld4q[_f16](float16_t const *addr) | addr -> Rn | VLD40.16 {Qd - Qd4}, [Rn]VLD41.16 {Qd - Qd4}, [Rn]VLD42.16 {Qd - Qd4}, [Rn]VLD43.16 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| float32x4x4_t [__arm_]vld4q[_f32](float32_t const *addr) | addr -> Rn | VLD40.32 {Qd - Qd4}, [Rn]VLD41.32 {Qd - Qd4}, [Rn]VLD42.32 {Qd - Qd4}, [Rn]VLD43.32 {Qd - Qd4}, [Rn] | Qd -> result.val[0]Qd2 -> result.val[1]Qd3 -> result.val[2]Qd4 -> result.val[3] | MVE | 
| int8x16_t [__arm_]vld1q[_s8](int8_t const *base) | base -> Rn | VLDRB.8 Qd, [Rn] | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vld1q[_s16](int16_t const *base) | base -> Rn | VLDRH.16 Qd, [Rn] | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vld1q[_s32](int32_t const *base) | base -> Rn | VLDRW.32 Qd, [Rn] | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vld1q[_u8](uint8_t const *base) | base -> Rn | VLDRB.8 Qd, [Rn] | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vld1q[_u16](uint16_t const *base) | base -> Rn | VLDRH.16 Qd, [Rn] | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vld1q[_u32](uint32_t const *base) | base -> Rn | VLDRW.32 Qd, [Rn] | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vld1q[_f16](float16_t const *base) | base -> Rn | VLDRH.16 Qd, [Rn] | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vld1q[_f32](float32_t const *base) | base -> Rn | VLDRW.32 Qd, [Rn] | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vld1q_z[_s8]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.8 Qd, [Rn] | Qd -> result | MVE | 
| int16x8_t [__arm_]vld1q_z[_s16]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.16 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vld1q_z[_s32]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRWT.32 Qd, [Rn] | Qd -> result | MVE | 
| uint8x16_t [__arm_]vld1q_z[_u8]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.8 Qd, [Rn] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vld1q_z[_u16]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.16 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vld1q_z[_u32]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRWT.32 Qd, [Rn] | Qd -> result | MVE | 
| float16x8_t [__arm_]vld1q_z[_f16]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.16 Qd, [Rn] | Qd -> result | MVE | 
| float32x4_t [__arm_]vld1q_z[_f32]( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRWT.32 Qd, [Rn] | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vldrbq_s8(int8_t const *base) | base -> Rn | VLDRB.8 Qd, [Rn] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrbq_s16(int8_t const *base) | base -> Rn | VLDRB.S16 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrbq_s32(int8_t const *base) | base -> Rn | VLDRB.S32 Qd, [Rn] | Qd -> result | MVE | 
| uint8x16_t [__arm_]vldrbq_u8(uint8_t const *base) | base -> Rn | VLDRB.8 Qd, [Rn] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrbq_u16(uint8_t const *base) | base -> Rn | VLDRB.U16 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrbq_u32(uint8_t const *base) | base -> Rn | VLDRB.U32 Qd, [Rn] | Qd -> result | MVE | 
| int8x16_t [__arm_]vldrbq_z_s8( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.8 Qd, [Rn] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrbq_z_s16( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.S16 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrbq_z_s32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.S32 Qd, [Rn] | Qd -> result | MVE | 
| uint8x16_t [__arm_]vldrbq_z_u8( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.8 Qd, [Rn] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrbq_z_u16( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.U16 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrbq_z_u32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRBT.U32 Qd, [Rn] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrhq_s16(int16_t const *base) | base -> Rn | VLDRH.16 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrhq_s32(int16_t const *base) | base -> Rn | VLDRH.S32 Qd, [Rn] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrhq_u16(uint16_t const *base) | base -> Rn | VLDRH.16 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrhq_u32(uint16_t const *base) | base -> Rn | VLDRH.U32 Qd, [Rn] | Qd -> result | MVE | 
| float16x8_t [__arm_]vldrhq_f16(float16_t const *base) | base -> Rn | VLDRH.16 Qd, [Rn] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrhq_z_s16( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.S16 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrhq_z_s32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.S32 Qd, [Rn] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrhq_z_u16( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.U16 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrhq_z_u32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.U32 Qd, [Rn] | Qd -> result | MVE | 
| float16x8_t [__arm_]vldrhq_z_f16( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRHT.F16 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_s32(int32_t const *base) | base -> Rn | VLDRW.32 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_u32(uint32_t const *base) | base -> Rn | VLDRW.32 Qd, [Rn] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_f32(float32_t const *base) | base -> Rn | VLDRW.32 Qd, [Rn] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_z_s32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRWT.32 Qd, [Rn] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_z_u32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRWT.32 Qd, [Rn] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_z_f32( | base -> Rnp -> Rp | VMSR P0, RpVPSTVLDRWT.32 Qd, [Rn] | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int16x8_t [__arm_]vldrhq_gather_offset[_s16]( | base -> Rnoffset -> Qm | VLDRH.U16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrhq_gather_offset[_s32]( | base -> Rnoffset -> Qm | VLDRH.S32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrhq_gather_offset[_u16]( | base -> Rnoffset -> Qm | VLDRH.U16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrhq_gather_offset[_u32]( | base -> Rnoffset -> Qm | VLDRH.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| float16x8_t [__arm_]vldrhq_gather_offset[_f16]( | base -> Rnoffset -> Qm | VLDRH.F16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrhq_gather_offset_z[_s16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.U16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrhq_gather_offset_z[_s32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.S32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrhq_gather_offset_z[_u16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.U16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrhq_gather_offset_z[_u32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| float16x8_t [__arm_]vldrhq_gather_offset_z[_f16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.F16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrhq_gather_shifted_offset[_s16]( | base -> Rnoffset -> Qm | VLDRH.U16 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrhq_gather_shifted_offset[_s32]( | base -> Rnoffset -> Qm | VLDRH.S32 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrhq_gather_shifted_offset[_u16]( | base -> Rnoffset -> Qm | VLDRH.U16 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrhq_gather_shifted_offset[_u32]( | base -> Rnoffset -> Qm | VLDRH.U32 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| float16x8_t [__arm_]vldrhq_gather_shifted_offset[_f16]( | base -> Rnoffset -> Qm | VLDRH.F16 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrhq_gather_shifted_offset_z[_s16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.U16 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrhq_gather_shifted_offset_z[_s32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.S32 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrhq_gather_shifted_offset_z[_u16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.U16 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrhq_gather_shifted_offset_z[_u32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.U32 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| float16x8_t [__arm_]vldrhq_gather_shifted_offset_z[_f16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRHT.F16 Qd, [Rn, Qm, UXTW #1] | Qd -> result | MVE | 
| int8x16_t [__arm_]vldrbq_gather_offset[_s8]( | base -> Rnoffset -> Qm | VLDRB.U8 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrbq_gather_offset[_s16]( | base -> Rnoffset -> Qm | VLDRB.S16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrbq_gather_offset[_s32]( | base -> Rnoffset -> Qm | VLDRB.S32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint8x16_t [__arm_]vldrbq_gather_offset[_u8]( | base -> Rnoffset -> Qm | VLDRB.U8 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrbq_gather_offset[_u16]( | base -> Rnoffset -> Qm | VLDRB.U16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrbq_gather_offset[_u32]( | base -> Rnoffset -> Qm | VLDRB.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int8x16_t [__arm_]vldrbq_gather_offset_z[_s8]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRBT.U8 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int16x8_t [__arm_]vldrbq_gather_offset_z[_s16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRBT.S16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrbq_gather_offset_z[_s32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRBT.S32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint8x16_t [__arm_]vldrbq_gather_offset_z[_u8]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRBT.U8 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint16x8_t [__arm_]vldrbq_gather_offset_z[_u16]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRBT.U16 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrbq_gather_offset_z[_u32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRBT.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_offset[_s32]( | base -> Rnoffset -> Qm | VLDRW.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_offset[_u32]( | base -> Rnoffset -> Qm | VLDRW.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_gather_offset[_f32]( | base -> Rnoffset -> Qm | VLDRW.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_offset_z[_s32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_offset_z[_u32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_gather_offset_z[_f32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_shifted_offset[_s32]( | base -> Rnoffset -> Qm | VLDRW.U32 Qd, [Rn, Qm, UXTW #2] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_shifted_offset[_u32]( | base -> Rnoffset -> Qm | VLDRW.U32 Qd, [Rn, Qm, UXTW #2] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_gather_shifted_offset[_f32]( | base -> Rnoffset -> Qm | VLDRW.U32 Qd, [Rn, Qm, UXTW #2] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_shifted_offset_z[_s32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Rn, Qm, UXTW #2] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_shifted_offset_z[_u32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Rn, Qm, UXTW #2] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_gather_shifted_offset_z[_f32]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Rn, Qm, UXTW #2] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_base_s32( | addr -> Qnoffset in +/-4*[0..127] | VLDRW.U32 Qd, [Qn, #offset] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_base_u32( | addr -> Qnoffset in +/-4*[0..127] | VLDRW.U32 Qd, [Qn, #offset] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_gather_base_f32( | addr -> Qnoffset in +/-4*[0..127] | VLDRW.U32 Qd, [Qn, #offset] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_base_z_s32( | addr -> Qnoffset in +/-4*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Qn, #offset] | Qd -> result | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_base_z_u32( | addr -> Qnoffset in +/-4*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Qn, #offset] | Qd -> result | MVE | 
| float32x4_t [__arm_]vldrwq_gather_base_z_f32( | addr -> Qnoffset in +/-4*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Qn, #offset] | Qd -> result | MVE | 
| int32x4_t [__arm_]vldrwq_gather_base_wb_s32( | *addr -> Qnoffset in +/-4*[0..127] | VLDRW.U32 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_base_wb_u32( | *addr -> Qnoffset in +/-4*[0..127] | VLDRW.U32 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| float32x4_t [__arm_]vldrwq_gather_base_wb_f32( | *addr -> Qnoffset in +/-4*[0..127] | VLDRW.U32 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| int32x4_t [__arm_]vldrwq_gather_base_wb_z_s32( | *addr -> Qnoffset in +/-4*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| uint32x4_t [__arm_]vldrwq_gather_base_wb_z_u32( | *addr -> Qnoffset in +/-4*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| float32x4_t [__arm_]vldrwq_gather_base_wb_z_f32( | *addr -> Qnoffset in +/-4*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRWT.U32 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| int64x2_t [__arm_]vldrdq_gather_offset[_s64]( | base -> Rnoffset -> Qm | VLDRD.U64 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_offset[_u64]( | base -> Rnoffset -> Qm | VLDRD.U64 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int64x2_t [__arm_]vldrdq_gather_offset_z[_s64]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Rn, Qm] | Qd -> result | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_offset_z[_u64]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Rn, Qm] | Qd -> result | MVE | 
| int64x2_t [__arm_]vldrdq_gather_shifted_offset[_s64]( | base -> Rnoffset -> Qm | VLDRD.U64 Qd, [Rn, Qm, UXTW #3] | Qd -> result | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_shifted_offset[_u64]( | base -> Rnoffset -> Qm | VLDRD.U64 Qd, [Rn, Qm, UXTW #3] | Qd -> result | MVE | 
| int64x2_t [__arm_]vldrdq_gather_shifted_offset_z[_s64]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Rn, Qm, UXTW #3] | Qd -> result | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_shifted_offset_z[_u64]( | base -> Rnoffset -> Qmp -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Rn, Qm, UXTW #3] | Qd -> result | MVE | 
| int64x2_t [__arm_]vldrdq_gather_base_s64( | addr -> Qnoffset in +/-8*[0..127] | VLDRD.64 Qd, [Qn, #offset] | Qd -> result | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_base_u64( | addr -> Qnoffset in +/-8*[0..127] | VLDRD.64 Qd, [Qn, #offset] | Qd -> result | MVE | 
| int64x2_t [__arm_]vldrdq_gather_base_z_s64( | addr -> Qnoffset in +/-8*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Qn, #offset] | Qd -> result | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_base_z_u64( | addr -> Qnoffset in +/-8*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Qn, #offset] | Qd -> result | MVE | 
| int64x2_t [__arm_]vldrdq_gather_base_wb_s64( | *addr -> Qnoffset in +/-8*[0..127] | VLDRD.64 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_base_wb_u64( | *addr -> Qnoffset in +/-8*[0..127] | VLDRD.64 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| int64x2_t [__arm_]vldrdq_gather_base_wb_z_s64( | *addr -> Qnoffset in +/-8*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| uint64x2_t [__arm_]vldrdq_gather_base_wb_z_u64( | *addr -> Qnoffset in +/-8*[0..127]p -> Rp | VMSR P0, RpVPSTVLDRDT.U64 Qd, [Qn, #offset]! | Qd -> resultQn -> *addr | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| void [__arm_]vst2q[_s8]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.8 {Qd - Qd2}, [Rn]VST21.8 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_s16]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.16 {Qd - Qd2}, [Rn]VST21.16 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_s32]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.32 {Qd - Qd2}, [Rn]VST21.32 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_u8]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.8 {Qd - Qd2}, [Rn]VST21.8 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_u16]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.16 {Qd - Qd2}, [Rn]VST21.16 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_u32]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.32 {Qd - Qd2}, [Rn]VST21.32 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_f16]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.16 {Qd - Qd2}, [Rn]VST21.16 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst2q[_f32]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2 | VST20.32 {Qd - Qd2}, [Rn]VST21.32 {Qd - Qd2}, [Rn] | MVE | |
| void [__arm_]vst4q[_s8]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.8 {Qd - Qd4}, [Rn]VST41.8 {Qd - Qd4}, [Rn]VST42.8 {Qd - Qd4}, [Rn]VST43.8 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_s16]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.16 {Qd - Qd4}, [Rn]VST41.16 {Qd - Qd4}, [Rn]VST42.16 {Qd - Qd4}, [Rn]VST43.16 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_s32]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.32 {Qd - Qd4}, [Rn]VST41.32 {Qd - Qd4}, [Rn]VST42.32 {Qd - Qd4}, [Rn]VST43.32 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_u8]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.8 {Qd - Qd4}, [Rn]VST41.8 {Qd - Qd4}, [Rn]VST42.8 {Qd - Qd4}, [Rn]VST43.8 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_u16]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.16 {Qd - Qd4}, [Rn]VST41.16 {Qd - Qd4}, [Rn]VST42.16 {Qd - Qd4}, [Rn]VST43.16 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_u32]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.32 {Qd - Qd4}, [Rn]VST41.32 {Qd - Qd4}, [Rn]VST42.32 {Qd - Qd4}, [Rn]VST43.32 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_f16]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.16 {Qd - Qd4}, [Rn]VST41.16 {Qd - Qd4}, [Rn]VST42.16 {Qd - Qd4}, [Rn]VST43.16 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst4q[_f32]( | addr -> Rnvalue.val[0] -> Qdvalue.val[1] -> Qd2value.val[2] -> Qd3value.val[3] -> Qd4 | VST40.32 {Qd - Qd4}, [Rn]VST41.32 {Qd - Qd4}, [Rn]VST42.32 {Qd - Qd4}, [Rn]VST43.32 {Qd - Qd4}, [Rn] | MVE | |
| void [__arm_]vst1q[_s8]( | base -> Rnvalue -> Qd | VSTRB.8 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_s16]( | base -> Rnvalue -> Qd | VSTRH.16 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_s32]( | base -> Rnvalue -> Qd | VSTRW.32 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_u8]( | base -> Rnvalue -> Qd | VSTRB.8 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_u16]( | base -> Rnvalue -> Qd | VSTRH.16 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_u32]( | base -> Rnvalue -> Qd | VSTRW.32 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_f16]( | base -> Rnvalue -> Qd | VSTRH.16 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q[_f32]( | base -> Rnvalue -> Qd | VSTRW.32 Qd, [Rn] | MVE/NEON | |
| void [__arm_]vst1q_p[_s8]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.8 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_s16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_s32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_u8]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.8 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_u16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_u32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_f16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn] | MVE | |
| void [__arm_]vst1q_p[_f32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn] | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| void [__arm_]vstrbq[_s8]( | base -> Rnvalue -> Qd | VSTRB.8 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq[_s16]( | base -> Rnvalue -> Qd | VSTRB.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq[_s32]( | base -> Rnvalue -> Qd | VSTRB.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq[_u8]( | base -> Rnvalue -> Qd | VSTRB.8 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq[_u16]( | base -> Rnvalue -> Qd | VSTRB.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq[_u32]( | base -> Rnvalue -> Qd | VSTRB.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq_p[_s8]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.8 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq_p[_s16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq_p[_s32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq_p[_u8]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.8 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq_p[_u16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrbq_p[_u32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq[_s16]( | base -> Rnvalue -> Qd | VSTRH.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq[_s32]( | base -> Rnvalue -> Qd | VSTRH.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq[_u16]( | base -> Rnvalue -> Qd | VSTRH.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq[_u32]( | base -> Rnvalue -> Qd | VSTRH.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq[_f16]( | base -> Rnvalue -> Qd | VSTRH.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq_p[_s16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq_p[_s32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq_p[_u16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq_p[_u32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrhq_p[_f16]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn] | MVE | |
| void [__arm_]vstrwq[_s32]( | base -> Rnvalue -> Qd | VSTRW.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrwq[_u32]( | base -> Rnvalue -> Qd | VSTRW.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrwq[_f32]( | base -> Rnvalue -> Qd | VSTRW.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrwq_p[_s32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrwq_p[_u32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn] | MVE | |
| void [__arm_]vstrwq_p[_f32]( | base -> Rnvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn] | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| void [__arm_]vstrbq_scatter_offset[_s8]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRB.8 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset[_s16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRB.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset[_s32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRB.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset[_u8]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRB.8 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset[_u16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRB.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset[_u32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRB.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset_p[_s8]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.8 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset_p[_s16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset_p[_s32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset_p[_u8]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.8 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset_p[_u16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrbq_scatter_offset_p[_u32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRBT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset[_s16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset[_s32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset[_u16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset[_u32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset[_f16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset_p[_s16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset_p[_s32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset_p[_u16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset_p[_u32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_offset_p[_f16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset[_s16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.16 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset[_s32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.32 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset[_u16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.16 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset[_u32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.32 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset[_f16]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRH.16 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset_p[_s16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset_p[_s32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.32 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset_p[_u16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset_p[_u32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.32 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrhq_scatter_shifted_offset_p[_f16]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRHT.16 Qd, [Rn, Qm, UXTW #1] | MVE | |
| void [__arm_]vstrwq_scatter_base[_s32]( | addr -> Qnoffset in +/-4*[0..127]value -> Qd | VSTRW.U32 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrwq_scatter_base[_u32]( | addr -> Qnoffset in +/-4*[0..127]value -> Qd | VSTRW.U32 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrwq_scatter_base[_f32]( | addr -> Qnoffset in +/-4*[0..127]value -> Qd | VSTRW.U32 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrwq_scatter_base_p[_s32]( | addr -> Qnoffset in +/-4*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.U32 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrwq_scatter_base_p[_u32]( | addr -> Qnoffset in +/-4*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.U32 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrwq_scatter_base_p[_f32]( | addr -> Qnoffset in +/-4*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.U32 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrwq_scatter_base_wb[_s32]( | *addr -> Qnoffset in +/-4*[0..127]value -> Qd | VSTRW.U32 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrwq_scatter_base_wb[_u32]( | *addr -> Qnoffset in +/-4*[0..127]value -> Qd | VSTRW.U32 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrwq_scatter_base_wb[_f32]( | *addr -> Qnoffset in +/-4*[0..127]value -> Qd | VSTRW.U32 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrwq_scatter_base_wb_p[_s32]( | *addr -> Qnoffset in +/-4*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.U32 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrwq_scatter_base_wb_p[_u32]( | *addr -> Qnoffset in +/-4*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.U32 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrwq_scatter_base_wb_p[_f32]( | *addr -> Qnoffset in +/-4*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.U32 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrwq_scatter_offset[_s32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRW.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrwq_scatter_offset[_u32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRW.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrwq_scatter_offset[_f32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRW.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrwq_scatter_offset_p[_s32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrwq_scatter_offset_p[_u32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrwq_scatter_offset_p[_f32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrwq_scatter_shifted_offset[_s32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRW.32 Qd, [Rn, Qm, UXTW #2] | MVE | |
| void [__arm_]vstrwq_scatter_shifted_offset[_u32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRW.32 Qd, [Rn, Qm, UXTW #2] | MVE | |
| void [__arm_]vstrwq_scatter_shifted_offset[_f32]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRW.32 Qd, [Rn, Qm, UXTW #2] | MVE | |
| void [__arm_]vstrwq_scatter_shifted_offset_p[_s32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn, Qm, UXTW #2] | MVE | |
| void [__arm_]vstrwq_scatter_shifted_offset_p[_u32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn, Qm, UXTW #2] | MVE | |
| void [__arm_]vstrwq_scatter_shifted_offset_p[_f32]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRWT.32 Qd, [Rn, Qm, UXTW #2] | MVE | |
| void [__arm_]vstrdq_scatter_base[_s64]( | addr -> Qnoffset in +/-8*[0..127]value -> Qd | VSTRD.U64 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrdq_scatter_base[_u64]( | addr -> Qnoffset in +/-8*[0..127]value -> Qd | VSTRD.U64 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrdq_scatter_base_p[_s64]( | addr -> Qnoffset in +/-8*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.U64 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrdq_scatter_base_p[_u64]( | addr -> Qnoffset in +/-8*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.U64 Qd, [Qn, #offset] | MVE | |
| void [__arm_]vstrdq_scatter_base_wb[_s64]( | *addr -> Qnoffset in +/-8*[0..127]value -> Qd | VSTRD.U64 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrdq_scatter_base_wb[_u64]( | *addr -> Qnoffset in +/-8*[0..127]value -> Qd | VSTRD.U64 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrdq_scatter_base_wb_p[_s64]( | *addr -> Qnoffset in +/-8*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.U64 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrdq_scatter_base_wb_p[_u64]( | *addr -> Qnoffset in +/-8*[0..127]value -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.U64 Qd, [Qn, #offset]! | Qn -> *addr | MVE | 
| void [__arm_]vstrdq_scatter_offset[_s64]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRD.64 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrdq_scatter_offset[_u64]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRD.64 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrdq_scatter_offset_p[_s64]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.64 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrdq_scatter_offset_p[_u64]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.64 Qd, [Rn, Qm] | MVE | |
| void [__arm_]vstrdq_scatter_shifted_offset[_s64]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRD.64 Qd, [Rn, Qm, UXTW #3] | MVE | |
| void [__arm_]vstrdq_scatter_shifted_offset[_u64]( | base -> Rnoffset -> Qmvalue -> Qd | VSTRD.64 Qd, [Rn, Qm, UXTW #3] | MVE | |
| void [__arm_]vstrdq_scatter_shifted_offset_p[_s64]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.64 Qd, [Rn, Qm, UXTW #3] | MVE | |
| void [__arm_]vstrdq_scatter_shifted_offset_p[_u64]( | base -> Rnoffset -> Qmvalue -> Qdp -> Rp | VMSR P0, RpVPSTVSTRDT.64 Qd, [Rn, Qm, UXTW #3] | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int16x8_t [__arm_]vcvtaq_s16_f16(float16x8_t a) | a -> Qm | VCVTA.S16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vcvtaq_s32_f32(float32x4_t a) | a -> Qm | VCVTA.S32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vcvtaq_u16_f16(float16x8_t a) | a -> Qm | VCVTA.U16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vcvtaq_u32_f32(float32x4_t a) | a -> Qm | VCVTA.U32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vcvtaq_m[_s16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtaq_m[_s32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtaq_m[_u16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtaq_m[_u32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtaq_x_s16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtaq_x_s32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtaq_x_u16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtaq_x_u32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTAT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtnq_s16_f16(float16x8_t a) | a -> Qm | VCVTN.S16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vcvtnq_s32_f32(float32x4_t a) | a -> Qm | VCVTN.S32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vcvtnq_u16_f16(float16x8_t a) | a -> Qm | VCVTN.U16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vcvtnq_u32_f32(float32x4_t a) | a -> Qm | VCVTN.U32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vcvtnq_m[_s16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtnq_m[_s32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtnq_m[_u16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtnq_m[_u32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtnq_x_s16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtnq_x_s32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtnq_x_u16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtnq_x_u32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTNT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtpq_s16_f16(float16x8_t a) | a -> Qm | VCVTP.S16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vcvtpq_s32_f32(float32x4_t a) | a -> Qm | VCVTP.S32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vcvtpq_u16_f16(float16x8_t a) | a -> Qm | VCVTP.U16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vcvtpq_u32_f32(float32x4_t a) | a -> Qm | VCVTP.U32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vcvtpq_m[_s16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtpq_m[_s32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtpq_m[_u16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtpq_m[_u32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtpq_x_s16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtpq_x_s32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtpq_x_u16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtpq_x_u32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTPT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtmq_s16_f16(float16x8_t a) | a -> Qm | VCVTM.S16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vcvtmq_s32_f32(float32x4_t a) | a -> Qm | VCVTM.S32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vcvtmq_u16_f16(float16x8_t a) | a -> Qm | VCVTM.U16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vcvtmq_u32_f32(float32x4_t a) | a -> Qm | VCVTM.U32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vcvtmq_m[_s16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtmq_m[_s32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtmq_m[_u16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtmq_m[_u32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtmq_x_s16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtmq_x_s32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtmq_x_u16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtmq_x_u32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTMT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtbq_f16_f32( | a -> Qdb -> Qm | VCVTB.F16.F32 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtbq_f32_f16(float16x8_t a) | a -> Qm | VCVTB.F32.F16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtbq_m_f16_f32( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVCVTBT.F16.F32 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtbq_m_f32_f16( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTBT.F32.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtbq_x_f32_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTBT.F32.F16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvttq_f16_f32( | a -> Qdb -> Qm | VCVTT.F16.F32 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvttq_f32_f16(float16x8_t a) | a -> Qm | VCVTT.F32.F16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvttq_m_f16_f32( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVCVTTT.F16.F32 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvttq_m_f32_f16( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTTT.F32.F16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvttq_x_f32_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTTT.F32.F16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq[_f16_s16](int16x8_t a) | a -> Qm | VCVT.F16.S16 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vcvtq[_f16_u16](uint16x8_t a) | a -> Qm | VCVT.F16.U16 Qd, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vcvtq[_f32_s32](int32x4_t a) | a -> Qm | VCVT.F32.S32 Qd, Qm | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vcvtq[_f32_u32](uint32x4_t a) | a -> Qm | VCVT.F32.U32 Qd, Qm | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vcvtq_m[_f16_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F16.S16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_m[_f16_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F16.U16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_m[_f32_s32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F32.S32 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_m[_f32_u32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F32.U32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_x[_f16_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F16.U16 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_x[_f16_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F16.S16 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_x[_f32_s32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F32.S32 Qd, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_x[_f32_u32]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.F32.U32 Qd, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_n[_f16_s16]( | a -> Qm1 <= imm6 <= 16 | VCVT.F16.S16 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vcvtq_n[_f16_u16]( | a -> Qm1 <= imm6 <= 16 | VCVT.F16.U16 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vcvtq_n[_f32_s32]( | a -> Qm1 <= imm6 <= 32 | VCVT.F32.S32 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vcvtq_n[_f32_u32]( | a -> Qm1 <= imm6 <= 32 | VCVT.F32.U32 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vcvtq_m_n[_f16_s16]( | inactive -> Qda -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.F16.S16 Qd, Qm, imm6 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_m_n[_f16_u16]( | inactive -> Qda -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.F16.U16 Qd, Qm, imm6 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_m_n[_f32_s32]( | inactive -> Qda -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.F32.S32 Qd, Qm, imm6 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_m_n[_f32_u32]( | inactive -> Qda -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.F32.U32 Qd, Qm, imm6 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_x_n[_f16_s16]( | a -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.F16.S16 Qd, Qm, imm6 | Qd -> result | MVE | 
| float16x8_t [__arm_]vcvtq_x_n[_f16_u16]( | a -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.F16.U16 Qd, Qm, imm6 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_x_n[_f32_s32]( | a -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.F32.S32 Qd, Qm, imm6 | Qd -> result | MVE | 
| float32x4_t [__arm_]vcvtq_x_n[_f32_u32]( | a -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.F32.U32 Qd, Qm, imm6 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtq_s16_f16(float16x8_t a) | a -> Qm | VCVT.S16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vcvtq_s32_f32(float32x4_t a) | a -> Qm | VCVT.S32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vcvtq_u16_f16(float16x8_t a) | a -> Qm | VCVT.U16.F16 Qd, Qm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vcvtq_u32_f32(float32x4_t a) | a -> Qm | VCVT.U32.F32 Qd, Qm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vcvtq_m[_s16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtq_m[_s32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtq_m[_u16_f16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtq_m[_u32_f32]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtq_x_s16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.S16.F16 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtq_x_s32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.S32.F32 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtq_x_u16_f16( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.U16.F16 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtq_x_u32_f32( | a -> Qmp -> Rp | VMSR P0, RpVPSTVCVTT.U32.F32 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtq_n_s16_f16( | a -> Qm1 <= imm6 <= 16 | VCVT.S16.F16 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vcvtq_n_s32_f32( | a -> Qm1 <= imm6 <= 32 | VCVT.S32.F32 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vcvtq_n_u16_f16( | a -> Qm1 <= imm6 <= 16 | VCVT.U16.F16 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vcvtq_n_u32_f32( | a -> Qm1 <= imm6 <= 32 | VCVT.U32.F32 Qd, Qm, imm6 | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vcvtq_m_n[_s16_f16]( | inactive -> Qda -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.S16.F16 Qd, Qm, imm6 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtq_m_n[_s32_f32]( | inactive -> Qda -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.S32.F32 Qd, Qm, imm6 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtq_m_n[_u16_f16]( | inactive -> Qda -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.U16.F16 Qd, Qm, imm6 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtq_m_n[_u32_f32]( | inactive -> Qda -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.U32.F32 Qd, Qm, imm6 | Qd -> result | MVE | 
| int16x8_t [__arm_]vcvtq_x_n_s16_f16( | a -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.S16.F16 Qd, Qm, imm6 | Qd -> result | MVE | 
| int32x4_t [__arm_]vcvtq_x_n_s32_f32( | a -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.S32.F32 Qd, Qm, imm6 | Qd -> result | MVE | 
| uint16x8_t [__arm_]vcvtq_x_n_u16_f16( | a -> Qm1 <= imm6 <= 16p -> Rp | VMSR P0, RpVPSTVCVTT.U16.F16 Qd, Qm, imm6 | Qd -> result | MVE | 
| uint32x4_t [__arm_]vcvtq_x_n_u32_f32( | a -> Qm1 <= imm6 <= 32p -> Rp | VMSR P0, RpVPSTVCVTT.U32.F32 Qd, Qm, imm6 | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int16x8_t [__arm_]vreinterpretq_s16[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_s8](int8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_s16](int16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_s32](int32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_f32](float32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_u8](uint8x16_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_u16](uint16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_u32](uint32x4_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_u64](uint64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float16x8_t [__arm_]vreinterpretq_f16[_s64](int64x2_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vreinterpretq_s8[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vreinterpretq_s16[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vreinterpretq_s32[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| float32x4_t [__arm_]vreinterpretq_f32[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vreinterpretq_u8[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vreinterpretq_u16[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vreinterpretq_u32[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| uint64x2_t [__arm_]vreinterpretq_u64[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| int64x2_t [__arm_]vreinterpretq_s64[_f16](float16x8_t a) | a -> Qd | NOP | Qd -> result | MVE/NEON | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vbrsrq[_n_s8]( | a -> Qnb -> Rm | VBRSR.8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vbrsrq[_n_s16]( | a -> Qnb -> Rm | VBRSR.16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vbrsrq[_n_s32]( | a -> Qnb -> Rm | VBRSR.32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vbrsrq[_n_u8]( | a -> Qnb -> Rm | VBRSR.8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vbrsrq[_n_u16]( | a -> Qnb -> Rm | VBRSR.16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vbrsrq[_n_u32]( | a -> Qnb -> Rm | VBRSR.32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vbrsrq[_n_f16]( | a -> Qnb -> Rm | VBRSR.16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vbrsrq[_n_f32]( | a -> Qnb -> Rm | VBRSR.32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vbrsrq_m[_n_s8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vbrsrq_m[_n_s16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vbrsrq_m[_n_s32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vbrsrq_m[_n_u8]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vbrsrq_m[_n_u16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vbrsrq_m[_n_u32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vbrsrq_m[_n_f16]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vbrsrq_m[_n_f32]( | inactive -> Qda -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.32 Qd, Qn, Rm | Qd -> result | MVE | 
| int8x16_t [__arm_]vbrsrq_x[_n_s8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.8 Qd, Qn, Rm | Qd -> result | MVE | 
| int16x8_t [__arm_]vbrsrq_x[_n_s16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.16 Qd, Qn, Rm | Qd -> result | MVE | 
| int32x4_t [__arm_]vbrsrq_x[_n_s32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.32 Qd, Qn, Rm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vbrsrq_x[_n_u8]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.8 Qd, Qn, Rm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vbrsrq_x[_n_u16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.16 Qd, Qn, Rm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vbrsrq_x[_n_u32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.32 Qd, Qn, Rm | Qd -> result | MVE | 
| float16x8_t [__arm_]vbrsrq_x[_n_f16]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.16 Qd, Qn, Rm | Qd -> result | MVE | 
| float32x4_t [__arm_]vbrsrq_x[_n_f32]( | a -> Qnb -> Rmp -> Rp | VMSR P0, RpVPSTVBRSRT.32 Qd, Qn, Rm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqrshrnbq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQRSHRNB.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrshrnbq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQRSHRNB.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshrnbq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQRSHRNB.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshrnbq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQRSHRNB.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrshrnbq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQRSHRNBT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrshrnbq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQRSHRNBT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshrnbq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQRSHRNBT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshrnbq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQRSHRNBT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrshrntq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQRSHRNT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrshrntq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQRSHRNT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshrntq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQRSHRNT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshrntq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQRSHRNT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqrshrntq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQRSHRNTT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrshrntq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQRSHRNTT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshrntq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQRSHRNTT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshrntq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQRSHRNTT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshrunbq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQRSHRUNB.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshrunbq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQRSHRUNB.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshrunbq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQRSHRUNBT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshrunbq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQRSHRUNBT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshruntq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQRSHRUNT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshruntq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQRSHRUNT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshruntq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQRSHRUNTT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshruntq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQRSHRUNTT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqshrnbq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQSHRNB.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqshrnbq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQSHRNB.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshrnbq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQSHRNB.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshrnbq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQSHRNB.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqshrnbq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQSHRNBT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqshrnbq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQSHRNBT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshrnbq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQSHRNBT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshrnbq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQSHRNBT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqshrntq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQSHRNT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqshrntq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQSHRNT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshrntq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQSHRNT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshrntq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQSHRNT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqshrntq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQSHRNTT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqshrntq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQSHRNTT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshrntq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQSHRNTT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshrntq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQSHRNTT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshrunbq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQSHRUNB.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshrunbq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQSHRUNB.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshrunbq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQSHRUNBT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshrunbq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQSHRUNBT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshruntq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VQSHRUNT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshruntq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VQSHRUNT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshruntq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVQSHRUNTT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshruntq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVQSHRUNTT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vrshrnbq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VRSHRNB.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshrnbq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VRSHRNB.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshrnbq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VRSHRNB.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshrnbq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VRSHRNB.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrshrnbq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRNBT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshrnbq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRNBT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshrnbq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRNBT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshrnbq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRNBT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrshrntq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VRSHRNT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshrntq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VRSHRNT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshrntq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VRSHRNT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshrntq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VRSHRNT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrshrntq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRNTT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshrntq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRNTT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshrntq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRNTT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshrntq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRNTT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vrshrq[_n_s8]( | a -> Qm1 <= imm <= 8 | VRSHR.S8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vrshrq[_n_s16]( | a -> Qm1 <= imm <= 16 | VRSHR.S16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vrshrq[_n_s32]( | a -> Qm1 <= imm <= 32 | VRSHR.S32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vrshrq[_n_u8]( | a -> Qm1 <= imm <= 8 | VRSHR.U8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vrshrq[_n_u16]( | a -> Qm1 <= imm <= 16 | VRSHR.U16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vrshrq[_n_u32]( | a -> Qm1 <= imm <= 32 | VRSHR.U32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vrshrq_m[_n_s8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshrq_m[_n_s16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrshrq_m[_n_s32]( | inactive -> Qda -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVRSHRT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshrq_m[_n_u8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshrq_m[_n_u16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrshrq_m[_n_u32]( | inactive -> Qda -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVRSHRT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vrshrq_x[_n_s8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshrq_x[_n_s16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vrshrq_x[_n_s32]( | a -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVRSHRT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshrq_x[_n_u8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVRSHRT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshrq_x[_n_u16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVRSHRT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrshrq_x[_n_u32]( | a -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVRSHRT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vshrnbq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VSHRNB.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshrnbq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VSHRNB.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshrnbq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VSHRNB.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshrnbq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VSHRNB.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshrnbq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRNBT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshrnbq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRNBT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshrnbq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRNBT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshrnbq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRNBT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshrntq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8 | VSHRNT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshrntq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16 | VSHRNT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshrntq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8 | VSHRNT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshrntq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16 | VSHRNT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshrntq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRNTT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshrntq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRNTT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshrntq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRNTT.I16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshrntq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRNTT.I32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vshrq[_n_s8]( | a -> Qm1 <= imm <= 8 | VSHR.S8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vshrq[_n_s16]( | a -> Qm1 <= imm <= 16 | VSHR.S16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vshrq[_n_s32]( | a -> Qm1 <= imm <= 32 | VSHR.S32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vshrq[_n_u8]( | a -> Qm1 <= imm <= 8 | VSHR.U8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vshrq[_n_u16]( | a -> Qm1 <= imm <= 16 | VSHR.U16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vshrq[_n_u32]( | a -> Qm1 <= imm <= 32 | VSHR.U32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vshrq_m[_n_s8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshrq_m[_n_s16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshrq_m[_n_s32]( | inactive -> Qda -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHRT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshrq_m[_n_u8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshrq_m[_n_u16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshrq_m[_n_u32]( | inactive -> Qda -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHRT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshrq_x[_n_s8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshrq_x[_n_s16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshrq_x[_n_s32]( | a -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHRT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshrq_x[_n_u8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHRT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshrq_x[_n_u16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHRT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshrq_x[_n_u32]( | a -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHRT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vsriq[_n_s8]( | a -> Qdb -> Qm1 <= imm <= 8 | VSRI.8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vsriq[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 16 | VSRI.16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vsriq[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 32 | VSRI.32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vsriq[_n_u8]( | a -> Qdb -> Qm1 <= imm <= 8 | VSRI.8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vsriq[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 16 | VSRI.16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vsriq[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 32 | VSRI.32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vsriq_m[_n_s8]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSRIT.8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsriq_m[_n_s16]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSRIT.16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsriq_m[_n_s32]( | a -> Qdb -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSRIT.32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsriq_m[_n_u8]( | a -> Qdb -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSRIT.8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsriq_m[_n_u16]( | a -> Qdb -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSRIT.16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsriq_m[_n_u32]( | a -> Qdb -> Qm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSRIT.32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqrshlq[_n_s8]( | a -> Qdab -> Rm | VQRSHL.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqrshlq[_n_s16]( | a -> Qdab -> Rm | VQRSHL.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqrshlq[_n_s32]( | a -> Qdab -> Rm | VQRSHL.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vqrshlq[_n_u8]( | a -> Qdab -> Rm | VQRSHL.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vqrshlq[_n_u16]( | a -> Qdab -> Rm | VQRSHL.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vqrshlq[_n_u32]( | a -> Qdab -> Rm | VQRSHL.U32 Qda, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqrshlq_m_n[_s8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQRSHLT.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqrshlq_m_n[_s16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQRSHLT.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqrshlq_m_n[_s32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQRSHLT.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vqrshlq_m_n[_u8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQRSHLT.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vqrshlq_m_n[_u16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQRSHLT.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vqrshlq_m_n[_u32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQRSHLT.U32 Qda, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqrshlq[_s8]( | a -> Qmb -> Qn | VQRSHL.S8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqrshlq[_s16]( | a -> Qmb -> Qn | VQRSHL.S16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqrshlq[_s32]( | a -> Qmb -> Qn | VQRSHL.S32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vqrshlq[_u8]( | a -> Qmb -> Qn | VQRSHL.U8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vqrshlq[_u16]( | a -> Qmb -> Qn | VQRSHL.U16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vqrshlq[_u32]( | a -> Qmb -> Qn | VQRSHL.U32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqrshlq_m[_s8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQRSHLT.S8 Qd, Qm, Qn | Qd -> result | MVE | 
| int16x8_t [__arm_]vqrshlq_m[_s16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQRSHLT.S16 Qd, Qm, Qn | Qd -> result | MVE | 
| int32x4_t [__arm_]vqrshlq_m[_s32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQRSHLT.S32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqrshlq_m[_u8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQRSHLT.U8 Qd, Qm, Qn | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqrshlq_m[_u16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQRSHLT.U16 Qd, Qm, Qn | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqrshlq_m[_u32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQRSHLT.U32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint64_t [__arm_]uqrshll( | value -> [RdaHi,RdaLo]shift -> Rm | UQRSHLL RdaLo, RdaHi, #64, Rm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]uqrshll_sat48( | value -> [RdaHi,RdaLo]shift -> Rm | UQRSHLL RdaLo, RdaHi, #48, Rm | [RdaHi,RdaLo] -> result | MVE | 
| uint64_t [__arm_]uqshll( | value -> [RdaHi,RdaLo]1 <= shift <= 32 | UQSHLL RdaLo, RdaHi, #shift | [RdaHi,RdaLo] -> result | MVE | 
| uint32_t [__arm_]uqrshl( | value -> Rdashift -> Rm | UQRSHL Rda, Rm | Rda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqshlq[_s8]( | a -> Qmb -> Qn | VQSHL.S8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqshlq[_s16]( | a -> Qmb -> Qn | VQSHL.S16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqshlq[_s32]( | a -> Qmb -> Qn | VQSHL.S32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vqshlq[_u8]( | a -> Qmb -> Qn | VQSHL.U8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vqshlq[_u16]( | a -> Qmb -> Qn | VQSHL.U16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vqshlq[_u32]( | a -> Qmb -> Qn | VQSHL.U32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqshlq_m[_s8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQSHLT.S8 Qd, Qm, Qn | Qd -> result | MVE | 
| int16x8_t [__arm_]vqshlq_m[_s16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQSHLT.S16 Qd, Qm, Qn | Qd -> result | MVE | 
| int32x4_t [__arm_]vqshlq_m[_s32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQSHLT.S32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshlq_m[_u8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQSHLT.U8 Qd, Qm, Qn | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshlq_m[_u16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQSHLT.U16 Qd, Qm, Qn | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqshlq_m[_u32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVQSHLT.U32 Qd, Qm, Qn | Qd -> result | MVE | 
| int8x16_t [__arm_]vqshlq_n[_s8]( | a -> Qm0 <= imm <= 7 | VQSHL.S8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vqshlq_n[_s16]( | a -> Qm0 <= imm <= 15 | VQSHL.S16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vqshlq_n[_s32]( | a -> Qm0 <= imm <= 31 | VQSHL.S32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vqshlq_n[_u8]( | a -> Qm0 <= imm <= 7 | VQSHL.U8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vqshlq_n[_u16]( | a -> Qm0 <= imm <= 15 | VQSHL.U16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vqshlq_n[_u32]( | a -> Qm0 <= imm <= 31 | VQSHL.U32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vqshlq_m_n[_s8]( | inactive -> Qda -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVQSHLT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqshlq_m_n[_s16]( | inactive -> Qda -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVQSHLT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vqshlq_m_n[_s32]( | inactive -> Qda -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVQSHLT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshlq_m_n[_u8]( | inactive -> Qda -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVQSHLT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshlq_m_n[_u16]( | inactive -> Qda -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVQSHLT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqshlq_m_n[_u32]( | inactive -> Qda -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVQSHLT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqshlq_r[_s8]( | a -> Qdab -> Rm | VQSHL.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqshlq_r[_s16]( | a -> Qdab -> Rm | VQSHL.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqshlq_r[_s32]( | a -> Qdab -> Rm | VQSHL.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vqshlq_r[_u8]( | a -> Qdab -> Rm | VQSHL.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vqshlq_r[_u16]( | a -> Qdab -> Rm | VQSHL.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vqshlq_r[_u32]( | a -> Qdab -> Rm | VQSHL.U32 Qda, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vqshlq_m_r[_s8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQSHLT.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vqshlq_m_r[_s16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQSHLT.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vqshlq_m_r[_s32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQSHLT.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vqshlq_m_r[_u8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQSHLT.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vqshlq_m_r[_u16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQSHLT.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vqshlq_m_r[_u32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVQSHLT.U32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vqshluq[_n_s8]( | a -> Qm0 <= imm <= 7 | VQSHLU.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshluq[_n_s16]( | a -> Qm0 <= imm <= 15 | VQSHLU.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqshluq[_n_s32]( | a -> Qm0 <= imm <= 31 | VQSHLU.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqshluq_m[_n_s8]( | inactive -> Qda -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVQSHLUT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqshluq_m[_n_s16]( | inactive -> Qda -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVQSHLUT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vqshluq_m[_n_s32]( | inactive -> Qda -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVQSHLUT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| int64_t [__arm_]sqshll( | value -> [RdaHi,RdaLo]1 <= shift <= 32 | SQSHLL RdaLo, RdaHi, #shift | [RdaHi,RdaLo] -> result | MVE | 
| uint32_t [__arm_]uqshl( | value -> Rda1 <= shift <= 32 | UQSHL Rda, #shift | Rda -> result | MVE | 
| int32_t [__arm_]sqshl( | value -> Rda1 <= shift <= 32 | SQSHL Rda, #shift | Rda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vrshlq[_n_s8]( | a -> Qdab -> Rm | VRSHL.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vrshlq[_n_s16]( | a -> Qdab -> Rm | VRSHL.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vrshlq[_n_s32]( | a -> Qdab -> Rm | VRSHL.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vrshlq[_n_u8]( | a -> Qdab -> Rm | VRSHL.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vrshlq[_n_u16]( | a -> Qdab -> Rm | VRSHL.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vrshlq[_n_u32]( | a -> Qdab -> Rm | VRSHL.U32 Qda, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vrshlq_m_n[_s8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVRSHLT.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vrshlq_m_n[_s16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVRSHLT.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vrshlq_m_n[_s32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVRSHLT.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vrshlq_m_n[_u8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVRSHLT.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vrshlq_m_n[_u16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVRSHLT.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vrshlq_m_n[_u32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVRSHLT.U32 Qda, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vrshlq[_s8]( | a -> Qmb -> Qn | VRSHL.S8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vrshlq[_s16]( | a -> Qmb -> Qn | VRSHL.S16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vrshlq[_s32]( | a -> Qmb -> Qn | VRSHL.S32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vrshlq[_u8]( | a -> Qmb -> Qn | VRSHL.U8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vrshlq[_u16]( | a -> Qmb -> Qn | VRSHL.U16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vrshlq[_u32]( | a -> Qmb -> Qn | VRSHL.U32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vrshlq_m[_s8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.S8 Qd, Qm, Qn | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshlq_m[_s16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.S16 Qd, Qm, Qn | Qd -> result | MVE | 
| int32x4_t [__arm_]vrshlq_m[_s32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.S32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshlq_m[_u8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.U8 Qd, Qm, Qn | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshlq_m[_u16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.U16 Qd, Qm, Qn | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrshlq_m[_u32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.U32 Qd, Qm, Qn | Qd -> result | MVE | 
| int8x16_t [__arm_]vrshlq_x[_s8]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.S8 Qd, Qm, Qn | Qd -> result | MVE | 
| int16x8_t [__arm_]vrshlq_x[_s16]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.S16 Qd, Qm, Qn | Qd -> result | MVE | 
| int32x4_t [__arm_]vrshlq_x[_s32]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.S32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint8x16_t [__arm_]vrshlq_x[_u8]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.U8 Qd, Qm, Qn | Qd -> result | MVE | 
| uint16x8_t [__arm_]vrshlq_x[_u16]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.U16 Qd, Qm, Qn | Qd -> result | MVE | 
| uint32x4_t [__arm_]vrshlq_x[_u32]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVRSHLT.U32 Qd, Qm, Qn | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vshlcq[_s8]( | a -> Qda*b -> Rdm1 <= imm <= 32 | VSHLC Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| int16x8_t [__arm_]vshlcq[_s16]( | a -> Qda*b -> Rdm1 <= imm <= 32 | VSHLC Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| int32x4_t [__arm_]vshlcq[_s32]( | a -> Qda*b -> Rdm1 <= imm <= 32 | VSHLC Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| uint8x16_t [__arm_]vshlcq[_u8]( | a -> Qda*b -> Rdm1 <= imm <= 32 | VSHLC Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| uint16x8_t [__arm_]vshlcq[_u16]( | a -> Qda*b -> Rdm1 <= imm <= 32 | VSHLC Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| uint32x4_t [__arm_]vshlcq[_u32]( | a -> Qda*b -> Rdm1 <= imm <= 32 | VSHLC Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| int8x16_t [__arm_]vshlcq_m[_s8]( | a -> Qda*b -> Rdm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHLCT Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| int16x8_t [__arm_]vshlcq_m[_s16]( | a -> Qda*b -> Rdm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHLCT Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| int32x4_t [__arm_]vshlcq_m[_s32]( | a -> Qda*b -> Rdm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHLCT Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| uint8x16_t [__arm_]vshlcq_m[_u8]( | a -> Qda*b -> Rdm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHLCT Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| uint16x8_t [__arm_]vshlcq_m[_u16]( | a -> Qda*b -> Rdm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHLCT Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| uint32x4_t [__arm_]vshlcq_m[_u32]( | a -> Qda*b -> Rdm1 <= imm <= 32p -> Rp | VMSR P0, RpVPSTVSHLCT Qda, Rdm, #imm | Qda -> resultRdm -> *b | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int16x8_t [__arm_]vshllbq[_n_s8]( | a -> Qm1 <= imm <= 8 | VSHLLB.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshllbq[_n_s16]( | a -> Qm1 <= imm <= 16 | VSHLLB.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshllbq[_n_u8]( | a -> Qm1 <= imm <= 8 | VSHLLB.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshllbq[_n_u16]( | a -> Qm1 <= imm <= 16 | VSHLLB.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshllbq_m[_n_s8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLBT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshllbq_m[_n_s16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLBT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshllbq_m[_n_u8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLBT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshllbq_m[_n_u16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLBT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshllbq_x[_n_s8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLBT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshllbq_x[_n_s16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLBT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshllbq_x[_n_u8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLBT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshllbq_x[_n_u16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLBT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlltq[_n_s8]( | a -> Qm1 <= imm <= 8 | VSHLLT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlltq[_n_s16]( | a -> Qm1 <= imm <= 16 | VSHLLT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlltq[_n_u8]( | a -> Qm1 <= imm <= 8 | VSHLLT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlltq[_n_u16]( | a -> Qm1 <= imm <= 16 | VSHLLT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlltq_m[_n_s8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLTT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlltq_m[_n_s16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLTT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlltq_m[_n_u8]( | inactive -> Qda -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLTT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlltq_m[_n_u16]( | inactive -> Qda -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLTT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlltq_x[_n_s8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLTT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlltq_x[_n_s16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLTT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlltq_x[_n_u8]( | a -> Qm1 <= imm <= 8p -> Rp | VMSR P0, RpVPSTVSHLLTT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlltq_x[_n_u16]( | a -> Qm1 <= imm <= 16p -> Rp | VMSR P0, RpVPSTVSHLLTT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshlq[_s8]( | a -> Qmb -> Qn | VSHL.S8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vshlq[_s16]( | a -> Qmb -> Qn | VSHL.S16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vshlq[_s32]( | a -> Qmb -> Qn | VSHL.S32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vshlq[_u8]( | a -> Qmb -> Qn | VSHL.U8 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vshlq[_u16]( | a -> Qmb -> Qn | VSHL.U16 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vshlq[_u32]( | a -> Qmb -> Qn | VSHL.U32 Qd, Qm, Qn | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vshlq_m[_s8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.S8 Qd, Qm, Qn | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlq_m[_s16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.S16 Qd, Qm, Qn | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlq_m[_s32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.S32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshlq_m[_u8]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.U8 Qd, Qm, Qn | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlq_m[_u16]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.U16 Qd, Qm, Qn | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlq_m[_u32]( | inactive -> Qda -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.U32 Qd, Qm, Qn | Qd -> result | MVE | 
| int8x16_t [__arm_]vshlq_x[_s8]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.S8 Qd, Qm, Qn | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlq_x[_s16]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.S16 Qd, Qm, Qn | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlq_x[_s32]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.S32 Qd, Qm, Qn | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshlq_x[_u8]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.U8 Qd, Qm, Qn | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlq_x[_u16]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.U16 Qd, Qm, Qn | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlq_x[_u32]( | a -> Qmb -> Qnp -> Rp | VMSR P0, RpVPSTVSHLT.U32 Qd, Qm, Qn | Qd -> result | MVE | 
| int8x16_t [__arm_]vshlq_n[_s8]( | a -> Qm0 <= imm <= 7 | VSHL.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlq_n[_s16]( | a -> Qm0 <= imm <= 15 | VSHL.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlq_n[_s32]( | a -> Qm0 <= imm <= 31 | VSHL.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshlq_n[_u8]( | a -> Qm0 <= imm <= 7 | VSHL.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlq_n[_u16]( | a -> Qm0 <= imm <= 15 | VSHL.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlq_n[_u32]( | a -> Qm0 <= imm <= 31 | VSHL.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshlq_m_n[_s8]( | inactive -> Qda -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVSHLT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlq_m_n[_s16]( | inactive -> Qda -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVSHLT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlq_m_n[_s32]( | inactive -> Qda -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVSHLT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshlq_m_n[_u8]( | inactive -> Qda -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVSHLT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlq_m_n[_u16]( | inactive -> Qda -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVSHLT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlq_m_n[_u32]( | inactive -> Qda -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVSHLT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshlq_x_n[_s8]( | a -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVSHLT.S8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vshlq_x_n[_s16]( | a -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVSHLT.S16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vshlq_x_n[_s32]( | a -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVSHLT.S32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vshlq_x_n[_u8]( | a -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVSHLT.U8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vshlq_x_n[_u16]( | a -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVSHLT.U16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vshlq_x_n[_u32]( | a -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVSHLT.U32 Qd, Qm, #imm | Qd -> result | MVE | 
| int8x16_t [__arm_]vshlq_r[_s8]( | a -> Qdab -> Rm | VSHL.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vshlq_r[_s16]( | a -> Qdab -> Rm | VSHL.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vshlq_r[_s32]( | a -> Qdab -> Rm | VSHL.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vshlq_r[_u8]( | a -> Qdab -> Rm | VSHL.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vshlq_r[_u16]( | a -> Qdab -> Rm | VSHL.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vshlq_r[_u32]( | a -> Qdab -> Rm | VSHL.U32 Qda, Rm | Qda -> result | MVE | 
| int8x16_t [__arm_]vshlq_m_r[_s8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVSHLT.S8 Qda, Rm | Qda -> result | MVE | 
| int16x8_t [__arm_]vshlq_m_r[_s16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVSHLT.S16 Qda, Rm | Qda -> result | MVE | 
| int32x4_t [__arm_]vshlq_m_r[_s32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVSHLT.S32 Qda, Rm | Qda -> result | MVE | 
| uint8x16_t [__arm_]vshlq_m_r[_u8]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVSHLT.U8 Qda, Rm | Qda -> result | MVE | 
| uint16x8_t [__arm_]vshlq_m_r[_u16]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVSHLT.U16 Qda, Rm | Qda -> result | MVE | 
| uint32x4_t [__arm_]vshlq_m_r[_u32]( | a -> Qdab -> Rmp -> Rp | VMSR P0, RpVPSTVSHLT.U32 Qda, Rm | Qda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vsliq[_n_s8]( | a -> Qdb -> Qm0 <= imm <= 7 | VSLI.8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int16x8_t [__arm_]vsliq[_n_s16]( | a -> Qdb -> Qm0 <= imm <= 15 | VSLI.16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int32x4_t [__arm_]vsliq[_n_s32]( | a -> Qdb -> Qm0 <= imm <= 31 | VSLI.32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint8x16_t [__arm_]vsliq[_n_u8]( | a -> Qdb -> Qm0 <= imm <= 7 | VSLI.8 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint16x8_t [__arm_]vsliq[_n_u16]( | a -> Qdb -> Qm0 <= imm <= 15 | VSLI.16 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| uint32x4_t [__arm_]vsliq[_n_u32]( | a -> Qdb -> Qm0 <= imm <= 31 | VSLI.32 Qd, Qm, #imm | Qd -> result | MVE/NEON | 
| int8x16_t [__arm_]vsliq_m[_n_s8]( | a -> Qdb -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVSLIT.8 Qd, Qm, #imm | Qd -> result | MVE | 
| int16x8_t [__arm_]vsliq_m[_n_s16]( | a -> Qdb -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVSLIT.16 Qd, Qm, #imm | Qd -> result | MVE | 
| int32x4_t [__arm_]vsliq_m[_n_s32]( | a -> Qdb -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVSLIT.32 Qd, Qm, #imm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vsliq_m[_n_u8]( | a -> Qdb -> Qm0 <= imm <= 7p -> Rp | VMSR P0, RpVPSTVSLIT.8 Qd, Qm, #imm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vsliq_m[_n_u16]( | a -> Qdb -> Qm0 <= imm <= 15p -> Rp | VMSR P0, RpVPSTVSLIT.16 Qd, Qm, #imm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vsliq_m[_n_u32]( | a -> Qdb -> Qm0 <= imm <= 31p -> Rp | VMSR P0, RpVPSTVSLIT.32 Qd, Qm, #imm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int16x8_t [__arm_]vmovlbq[_s8](int8x16_t a) | a -> Qm | VMOVLB.S8 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmovlbq[_s16](int16x8_t a) | a -> Qm | VMOVLB.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovlbq[_u8](uint8x16_t a) | a -> Qm | VMOVLB.U8 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmovlbq[_u16](uint16x8_t a) | a -> Qm | VMOVLB.U16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovlbq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.S8 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmovlbq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovlbq_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.U8 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmovlbq_m[_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.U16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovlbq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.S8 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmovlbq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovlbq_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.U8 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmovlbq_x[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLBT.U16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovltq[_s8](int8x16_t a) | a -> Qm | VMOVLT.S8 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmovltq[_s16](int16x8_t a) | a -> Qm | VMOVLT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovltq[_u8](uint8x16_t a) | a -> Qm | VMOVLT.U8 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmovltq[_u16](uint16x8_t a) | a -> Qm | VMOVLT.U16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovltq_m[_s8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.S8 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmovltq_m[_s16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovltq_m[_u8]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.U8 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmovltq_m[_u16]( | inactive -> Qda -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.U16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovltq_x[_s8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.S8 Qd, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vmovltq_x[_s16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovltq_x[_u8]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.U8 Qd, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vmovltq_x[_u16]( | a -> Qmp -> Rp | VMSR P0, RpVPSTVMOVLTT.U16 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmovnbq[_s16]( | a -> Qdb -> Qm | VMOVNB.I16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovnbq[_s32]( | a -> Qdb -> Qm | VMOVNB.I32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmovnbq[_u16]( | a -> Qdb -> Qm | VMOVNB.I16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovnbq[_u32]( | a -> Qdb -> Qm | VMOVNB.I32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmovnbq_m[_s16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNBT.I16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovnbq_m[_s32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNBT.I32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmovnbq_m[_u16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNBT.I16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovnbq_m[_u32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNBT.I32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmovntq[_s16]( | a -> Qdb -> Qm | VMOVNT.I16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovntq[_s32]( | a -> Qdb -> Qm | VMOVNT.I32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmovntq[_u16]( | a -> Qdb -> Qm | VMOVNT.I16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovntq[_u32]( | a -> Qdb -> Qm | VMOVNT.I32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vmovntq_m[_s16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNTT.I16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vmovntq_m[_s32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNTT.I32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vmovntq_m[_u16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNTT.I16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vmovntq_m[_u32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVMOVNTT.I32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vqmovnbq[_s16]( | a -> Qdb -> Qm | VQMOVNB.S16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqmovnbq[_s32]( | a -> Qdb -> Qm | VQMOVNB.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovnbq[_u16]( | a -> Qdb -> Qm | VQMOVNB.U16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovnbq[_u32]( | a -> Qdb -> Qm | VQMOVNB.U32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqmovnbq_m[_s16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNBT.S16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqmovnbq_m[_s32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNBT.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovnbq_m[_u16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNBT.U16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovnbq_m[_u32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNBT.U32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqmovntq[_s16]( | a -> Qdb -> Qm | VQMOVNT.S16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqmovntq[_s32]( | a -> Qdb -> Qm | VQMOVNT.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovntq[_u16]( | a -> Qdb -> Qm | VQMOVNT.U16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovntq[_u32]( | a -> Qdb -> Qm | VQMOVNT.U32 Qd, Qm | Qd -> result | MVE | 
| int8x16_t [__arm_]vqmovntq_m[_s16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNTT.S16 Qd, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vqmovntq_m[_s32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNTT.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovntq_m[_u16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNTT.U16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovntq_m[_u32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVNTT.U32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovunbq[_s16]( | a -> Qdb -> Qm | VQMOVUNB.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovunbq[_s32]( | a -> Qdb -> Qm | VQMOVUNB.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovunbq_m[_s16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVUNBT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovunbq_m[_s32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVUNBT.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovuntq[_s16]( | a -> Qdb -> Qm | VQMOVUNT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovuntq[_s32]( | a -> Qdb -> Qm | VQMOVUNT.S32 Qd, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vqmovuntq_m[_s16]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVUNTT.S16 Qd, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vqmovuntq_m[_s32]( | a -> Qdb -> Qmp -> Rp | VMSR P0, RpVPSTVQMOVUNTT.S32 Qd, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vpnot(mve_pred16_t a) | a -> Rp | VMSR P0, RpVPNOTVMRS Rt, P0 | Rt -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int8x16_t [__arm_]vpselq[_s8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| int16x8_t [__arm_]vpselq[_s16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| int32x4_t [__arm_]vpselq[_s32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| int64x2_t [__arm_]vpselq[_s64]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| uint8x16_t [__arm_]vpselq[_u8]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| uint16x8_t [__arm_]vpselq[_u16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| uint32x4_t [__arm_]vpselq[_u32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| uint64x2_t [__arm_]vpselq[_u64]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| float16x8_t [__arm_]vpselq[_f16]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| float32x4_t [__arm_]vpselq[_f32]( | a -> Qnb -> Qmp -> Rp | VMSR P0, RpVPSEL Qd, Qn, Qm | Qd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| mve_pred16_t [__arm_]vctp8q(uint32_t a) | a -> Rn | VCTP.8 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp16q(uint32_t a) | a -> Rn | VCTP.16 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp32q(uint32_t a) | a -> Rn | VCTP.32 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp64q(uint32_t a) | a -> Rn | VCTP.64 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp8q_m( | a -> Rnp -> Rp | VMSR P0, RpVPSTVCTPT.8 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp16q_m( | a -> Rnp -> Rp | VMSR P0, RpVPSTVCTPT.16 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp32q_m( | a -> Rnp -> Rp | VMSR P0, RpVPSTVCTPT.32 RnVMRS Rd, P0 | Rd -> result | MVE | 
| mve_pred16_t [__arm_]vctp64q_m( | a -> Rnp -> Rp | VMSR P0, RpVPSTVCTPT.64 RnVMRS Rd, P0 | Rd -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| uint64_t [__arm_]lsll( | value -> [RdaHi,RdaLo]shift -> Rm | LSLL RdaLo, RdaHi, Rm | [RdaHi,RdaLo] -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int64_t [__arm_]asrl( | value -> [RdaHi,RdaLo]shift -> Rm | ASRL RdaLo, RdaHi, Rm | [RdaHi,RdaLo] -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| int64_t [__arm_]sqrshrl( | value -> [RdaHi,RdaLo]shift -> Rm | SQRSHRL RdaLo, RdaHi, #64, Rm | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]sqrshrl_sat48( | value -> [RdaHi,RdaLo]shift -> Rm | SQRSHRL RdaLo, RdaHi, #48, Rm | [RdaHi,RdaLo] -> result | MVE | 
| int32_t [__arm_]sqrshr( | value -> Rdashift -> Rm | SQRSHR Rda, Rm | Rda -> result | MVE | 
| Intrinsic | Argument preparation | Instruction | Result | Supported architectures | 
|---|---|---|---|---|
| uint64_t [__arm_]urshrl( | value -> [RdaHi,RdaLo]1 <= shift <= 32 | URSHRL RdaLo, RdaHi, #shift | [RdaHi,RdaLo] -> result | MVE | 
| int64_t [__arm_]srshrl( | value -> [RdaHi,RdaLo]1 <= shift <= 32 | SRSHRL RdaLo, RdaHi, #shift | [RdaHi,RdaLo] -> result | MVE | 
| uint32_t [__arm_]urshr( | value -> Rda1 <= shift <= 32 | URSHR Rda, #shift | Rda -> result | MVE | 
| int32_t [__arm_]srshr( | value -> Rda1 <= shift <= 32 | SRSHR Rda, #shift | Rda -> result | MVE |