Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 | // SPDX-License-Identifier: GPL-2.0-or-later /* * * Copyright (C) IBM Corporation, 2012 * * Author: Anton Blanchard <anton@au.ibm.com> */ /* * Sparse (as at v0.5.0) gets very, very confused by this file. * Make it a bit simpler for it. */ #if !defined(__CHECKER__) #include <altivec.h> #else #define vec_xor(a, b) a ^ b #define vector __attribute__((vector_size(16))) #endif #include "xor_vmx.h" typedef vector signed char unative_t; #define DEFINE(V) \ unative_t *V = (unative_t *)V##_in; \ unative_t V##_0, V##_1, V##_2, V##_3 #define LOAD(V) \ do { \ V##_0 = V[0]; \ V##_1 = V[1]; \ V##_2 = V[2]; \ V##_3 = V[3]; \ } while (0) #define STORE(V) \ do { \ V[0] = V##_0; \ V[1] = V##_1; \ V[2] = V##_2; \ V[3] = V##_3; \ } while (0) #define XOR(V1, V2) \ do { \ V1##_0 = vec_xor(V1##_0, V2##_0); \ V1##_1 = vec_xor(V1##_1, V2##_1); \ V1##_2 = vec_xor(V1##_2, V2##_2); \ V1##_3 = vec_xor(V1##_3, V2##_3); \ } while (0) void __xor_altivec_2(unsigned long bytes, unsigned long * __restrict v1_in, const unsigned long * __restrict v2_in) { DEFINE(v1); DEFINE(v2); unsigned long lines = bytes / (sizeof(unative_t)) / 4; do { LOAD(v1); LOAD(v2); XOR(v1, v2); STORE(v1); v1 += 4; v2 += 4; } while (--lines > 0); } void __xor_altivec_3(unsigned long bytes, unsigned long * __restrict v1_in, const unsigned long * __restrict v2_in, const unsigned long * __restrict v3_in) { DEFINE(v1); DEFINE(v2); DEFINE(v3); unsigned long lines = bytes / (sizeof(unative_t)) / 4; do { LOAD(v1); LOAD(v2); LOAD(v3); XOR(v1, v2); XOR(v1, v3); STORE(v1); v1 += 4; v2 += 4; v3 += 4; } while (--lines > 0); } void __xor_altivec_4(unsigned long bytes, unsigned long * __restrict v1_in, const unsigned long * __restrict v2_in, const unsigned long * __restrict v3_in, const unsigned long * __restrict v4_in) { DEFINE(v1); DEFINE(v2); DEFINE(v3); DEFINE(v4); unsigned long lines = bytes / (sizeof(unative_t)) / 4; do { LOAD(v1); LOAD(v2); LOAD(v3); LOAD(v4); XOR(v1, v2); XOR(v3, v4); XOR(v1, v3); STORE(v1); v1 += 4; v2 += 4; v3 += 4; v4 += 4; } while (--lines > 0); } void __xor_altivec_5(unsigned long bytes, unsigned long * __restrict v1_in, const unsigned long * __restrict v2_in, const unsigned long * __restrict v3_in, const unsigned long * __restrict v4_in, const unsigned long * __restrict v5_in) { DEFINE(v1); DEFINE(v2); DEFINE(v3); DEFINE(v4); DEFINE(v5); unsigned long lines = bytes / (sizeof(unative_t)) / 4; do { LOAD(v1); LOAD(v2); LOAD(v3); LOAD(v4); LOAD(v5); XOR(v1, v2); XOR(v3, v4); XOR(v1, v5); XOR(v1, v3); STORE(v1); v1 += 4; v2 += 4; v3 += 4; v4 += 4; v5 += 4; } while (--lines > 0); } |