Home | History | Annotate | Download | only in rs6000

Lines Matching defs:__v4sf

93 typedef float __v4sf __attribute__ ((__vector_size__ (16)));
115 return ((__m128)vec_ld(0, (__v4sf*)__P));
129 __v4sf __tmp;
135 __tmp = vec_ld (0, (__v4sf *) __P);
144 return __extension__ (__m128)(__v4sf){ __F, __F, __F, __F };
157 return __extension__ (__m128)(__v4sf){ __W, __X, __Y, __Z };
164 return __extension__ (__m128)(__v4sf){ __Z, __Y, __X, __W };
172 vec_st((__v4sf)__A, 0, (__v4sf*)__P);
186 __v4sf __tmp;
200 __v4sf __va = vec_splat((__v4sf)__A, 0);
214 return __extension__ (__m128)(__v4sf){ __F, 0.0f, 0.0f, 0.0f };
223 return (vec_sel ((__v4sf)__A, (__v4sf)__B, __mask));
237 *__P = ((__v4sf)__A)[0];
352 return (__m128) ((__v4sf)__A + (__v4sf)__B);
358 return (__m128) ((__v4sf)__A - (__v4sf)__B);
364 return (__m128) ((__v4sf)__A * (__v4sf)__B);
370 return (__m128) ((__v4sf)__A / (__v4sf)__B);
376 return (vec_sqrt ((__v4sf)__A));
382 return (vec_re ((__v4sf)__A));
426 __v4sf __a, __b, __c;
432 __a = vec_splat ((__v4sf)__A, 0);
433 __b = vec_splat ((__v4sf)__B, 0);
437 return (vec_sel ((__v4sf)__A, __c, __mask));
443 __v4sf __a, __b, __c;
454 return (vec_sel ((__v4sf)__A, __c, __mask));
460 __vector __bool int __m = vec_cmpgt ((__v4sf) __B, (__v4sf) __A);
467 __vector __bool int __m = vec_cmpgt ((__v4sf) __A, (__v4sf) __B);
475 return ((__m128)vec_and ((__v4sf)__A, (__v4sf)__B));
482 return ((__m128)vec_andc ((__v4sf)__B, (__v4sf)__A));
488 return ((__m128)vec_or ((__v4sf)__A, (__v4sf)__B));
494 return ((__m128)vec_xor ((__v4sf)__A, (__v4sf)__B));
503 return ((__m128)vec_cmpeq ((__v4sf)__A,(__v4sf) __B));
509 return ((__m128)vec_cmplt ((__v4sf)__A, (__v4sf)__B));
515 return ((__m128)vec_cmple ((__v4sf)__A, (__v4sf)__B));
521 return ((__m128)vec_cmpgt ((__v4sf)__A, (__v4sf)__B));
527 return ((__m128)vec_cmpge ((__v4sf)__A, (__v4sf)__B));
533 __v4sf __temp = (__v4sf ) vec_cmpeq ((__v4sf) __A, (__v4sf)__B);
540 return ((__m128)vec_cmpge ((__v4sf)__A, (__v4sf)__B));
546 return ((__m128)vec_cmpgt ((__v4sf)__A, (__v4sf)__B));
552 return ((__m128)vec_cmple ((__v4sf)__A, (__v4sf)__B));
558 return ((__m128)vec_cmplt ((__v4sf)__A, (__v4sf)__B));
569 __a = (__vector unsigned int) vec_abs ((__v4sf)__A);
570 __b = (__vector unsigned int) vec_abs ((__v4sf)__B);
584 __a = (__vector unsigned int) vec_abs ((__v4sf)__A);
585 __b = (__vector unsigned int) vec_abs ((__v4sf)__B);
599 __v4sf __a, __b, __c;
604 __a = vec_splat ((__v4sf) __A, 0);
605 __b = vec_splat ((__v4sf) __B, 0);
606 __c = (__v4sf) vec_cmpeq (__a, __b);
609 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
617 __v4sf __a, __b, __c;
622 __a = vec_splat ((__v4sf) __A, 0);
623 __b = vec_splat ((__v4sf) __B, 0);
624 __c = (__v4sf) vec_cmplt(__a, __b);
627 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
635 __v4sf __a, __b, __c;
640 __a = vec_splat ((__v4sf) __A, 0);
641 __b = vec_splat ((__v4sf) __B, 0);
642 __c = (__v4sf) vec_cmple(__a, __b);
645 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
653 __v4sf __a, __b, __c;
658 __a = vec_splat ((__v4sf) __A, 0);
659 __b = vec_splat ((__v4sf) __B, 0);
660 __c = (__v4sf) vec_cmpgt(__a, __b);
663 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
671 __v4sf __a, __b, __c;
676 __a = vec_splat ((__v4sf) __A, 0);
677 __b = vec_splat ((__v4sf) __B, 0);
678 __c = (__v4sf) vec_cmpge(__a, __b);
681 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
689 __v4sf __a, __b, __c;
694 __a = vec_splat ((__v4sf) __A, 0);
695 __b = vec_splat ((__v4sf) __B, 0);
696 __c = (__v4sf) vec_cmpeq(__a, __b);
700 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
708 __v4sf __a, __b, __c;
713 __a = vec_splat ((__v4sf) __A, 0);
714 __b = vec_splat ((__v4sf) __B, 0);
715 __c = (__v4sf) vec_cmpge(__a, __b);
718 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
726 __v4sf __a, __b, __c;
731 __a = vec_splat ((__v4sf) __A, 0);
732 __b = vec_splat ((__v4sf) __B, 0);
733 __c = (__v4sf) vec_cmpgt(__a, __b);
736 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
744 __v4sf __a, __b, __c;
749 __a = vec_splat ((__v4sf) __A, 0);
750 __b = vec_splat ((__v4sf) __B, 0);
751 __c = (__v4sf) vec_cmple(__a, __b);
754 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
762 __v4sf __a, __b, __c;
767 __a = vec_splat ((__v4sf) __A, 0);
768 __b = vec_splat ((__v4sf) __B, 0);
769 __c = (__v4sf) vec_cmplt(__a, __b);
772 return ((__m128)vec_sel ((__v4sf)__A, __c, __mask));
785 __a = (__vector unsigned int) vec_abs ((__v4sf)__A);
786 __b = (__vector unsigned int) vec_abs ((__v4sf)__B);
792 return ((__m128)vec_sel ((__v4sf)__A, (__v4sf)__c, __mask));
805 __a = (__vector unsigned int) vec_abs ((__v4sf)__A);
806 __b = (__vector unsigned int) vec_abs ((__v4sf)__B);
812 return ((__m128)vec_sel ((__v4sf)__A, (__v4sf)__c, __mask));
900 return ((__v4sf)__A)[0];
965 return _mm_cvtss_si64 ((__v4sf) __A);
995 __v4sf __temp, __rounded;
999 __temp = (__v4sf) vec_splat ((__vector long long)__A, 0);
1053 __v4sf __temp;
1057 __temp = (__v4sf) vec_splat ((__vector long long)__A, 0);
1222 __v4sf __rounded;
1237 __v4sf __rounded;
1273 return vec_perm ((__v4sf) __A, (__v4sf)__B, (__vector unsigned char)__t);
1280 return (__m128) vec_vmrglw ((__v4sf) __A, (__v4sf)__B);
1287 return (__m128) vec_vmrghw ((__v4sf) __A, (__v4sf)__B);
1861 __v4sf __r0 = (row0), __r1 = (row1), __r2 = (row2), __r3 = (row3); \
1862 __v4sf __t0 = vec_vmrghw (__r0, __r1); \
1863 __v4sf __t1 = vec_vmrghw (__r2, __r3); \
1864 __v4sf __t2 = vec_vmrglw (__r0, __r1); \
1865 __v4sf __t3 = vec_vmrglw (__r2, __r3); \
1866 (row0) = (__v4sf)vec_mergeh ((__vector long long)__t0, \
1868 (row1) = (__v4sf)vec_mergel ((__vector long long)__t0, \
1870 (row2) = (__v4sf)vec_mergeh ((__vector long long)__t2, \
1872 (row3) = (__v4sf)vec_mergel ((__vector long long)__t2, \