| abs() | vsx::svec< 4, uint16_t > | |
| broadcast(int32_t index) | vsx::svec< 4, uint16_t > | |
| gather(svec< 4, void * > ptrs, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | static |
| gather_base_offsets(uint16_t *b, uint32_t scale, svec< 4, int32_t > offsets, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | static |
| gather_base_offsets(uint16_t *b, uint32_t scale, svec< 4, int64_t > offsets, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | static |
| gather_stride(uint16_t *b, int32_t off, int32_t stride) | vsx::svec< 4, uint16_t > | static |
| gather_stride(uint16_t *b, int64_t off, int64_t stride) | vsx::svec< 4, uint16_t > | static |
| load(svec< 4, uint16_t > *p) | vsx::svec< 4, uint16_t > | static |
| load_and_splat(uint16_t *p) | vsx::svec< 4, uint16_t > | static |
| load_const(const uint16_t *p) | vsx::svec< 4, uint16_t > | static |
| masked_load(svec< 4, uint16_t > *p, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | static |
| masked_store(svec< 4, uint16_t > *p, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | |
| operator!=(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator%(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator%(uint16_t s) | vsx::svec< 4, uint16_t > | |
| operator&(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator*(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator*(uint16_t s) | vsx::svec< 4, uint16_t > | |
| operator+(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator+(uint16_t s) | vsx::svec< 4, uint16_t > | |
| operator-() | vsx::svec< 4, uint16_t > | |
| operator-(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator-(uint16_t s) | vsx::svec< 4, uint16_t > | |
| operator/(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator/(uint16_t s) | vsx::svec< 4, uint16_t > | |
| operator<(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator<<(std::ostream &out, const svec< 4, uint16_t > &v) | vsx::svec< 4, uint16_t > | friend |
| operator<<(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator<<(int32_t s) | vsx::svec< 4, uint16_t > | |
| operator<=(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator==(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator>(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator>=(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator>>(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator>>(int32_t s) | vsx::svec< 4, uint16_t > | |
| operator[](int index) | vsx::svec< 4, uint16_t > | |
| operator[](int index) const | vsx::svec< 4, uint16_t > | |
| operator^(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| operator|(svec< 4, uint16_t > a) | vsx::svec< 4, uint16_t > | |
| reduce_add() | vsx::svec< 4, uint16_t > | |
| reduce_max() | vsx::svec< 4, uint16_t > | |
| reduce_min() | vsx::svec< 4, uint16_t > | |
| rotate(int32_t index) | vsx::svec< 4, uint16_t > | |
| scatter(svec< 4, void * > ptrs, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | |
| scatter_base_offsets(uint16_t *b, uint32_t scale, svec< 4, int32_t > offsets, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | |
| scatter_base_offsets(uint16_t *b, uint32_t scale, svec< 4, int64_t > offsets, svec< 4, bool > mask) | vsx::svec< 4, uint16_t > | |
| scatter_stride(uint16_t *b, int32_t off, int32_t stride) | vsx::svec< 4, uint16_t > | |
| scatter_stride(uint16_t *b, int64_t off, int64_t stride) | vsx::svec< 4, uint16_t > | |
| shuffle(svec< 4, int32_t > index) | vsx::svec< 4, uint16_t > | |
| store(svec< 4, uint16_t > *p) | vsx::svec< 4, uint16_t > | |
| svec() | vsx::svec< 4, uint16_t > | |
| svec(__vector unsigned short vv) | vsx::svec< 4, uint16_t > | |
| svec(uint16_t a, uint16_t b, uint16_t c, uint16_t d) | vsx::svec< 4, uint16_t > | |
| svec(uint16_t a) | vsx::svec< 4, uint16_t > | |
| v | vsx::svec< 4, uint16_t > | |