tor-browser

The Tor Browser
git clone https://git.dasho.dev/tor-browser.git
Log | Files | Refs | README | LICENSE

quantize_x86.h (8490B)


      1 /*
      2 * Copyright (c) 2018, Alliance for Open Media. All rights reserved.
      3 *
      4 * This source code is subject to the terms of the BSD 2 Clause License and
      5 * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License
      6 * was not distributed with this source code in the LICENSE file, you can
      7 * obtain it at www.aomedia.org/license/software. If the Alliance for Open
      8 * Media Patent License 1.0 was not distributed with this source code in the
      9 * PATENTS file, you can obtain it at www.aomedia.org/license/patent.
     10 */
     11 
     12 #include <emmintrin.h>
     13 
     14 #include "aom/aom_integer.h"
     15 
     16 static inline void load_b_values(const int16_t *zbin_ptr, __m128i *zbin,
     17                                 const int16_t *round_ptr, __m128i *round,
     18                                 const int16_t *quant_ptr, __m128i *quant,
     19                                 const int16_t *dequant_ptr, __m128i *dequant,
     20                                 const int16_t *shift_ptr, __m128i *shift) {
     21  *zbin = _mm_load_si128((const __m128i *)zbin_ptr);
     22  *round = _mm_load_si128((const __m128i *)round_ptr);
     23  *quant = _mm_load_si128((const __m128i *)quant_ptr);
     24  *zbin = _mm_sub_epi16(*zbin, _mm_set1_epi16(1));
     25  *dequant = _mm_load_si128((const __m128i *)dequant_ptr);
     26  *shift = _mm_load_si128((const __m128i *)shift_ptr);
     27 }
     28 
     29 // With ssse3 and later abs() and sign() are preferred.
     30 static inline __m128i invert_sign_sse2(__m128i a, __m128i sign) {
     31  a = _mm_xor_si128(a, sign);
     32  return _mm_sub_epi16(a, sign);
     33 }
     34 
     35 static inline __m128i invert_sign_32_sse2(__m128i a, __m128i sign) {
     36  a = _mm_xor_si128(a, sign);
     37  return _mm_sub_epi32(a, sign);
     38 }
     39 
     40 static inline void calculate_qcoeff(__m128i *coeff, const __m128i round,
     41                                    const __m128i quant, const __m128i shift) {
     42  __m128i tmp, qcoeff;
     43  qcoeff = _mm_adds_epi16(*coeff, round);
     44  tmp = _mm_mulhi_epi16(qcoeff, quant);
     45  qcoeff = _mm_add_epi16(tmp, qcoeff);
     46  *coeff = _mm_mulhi_epi16(qcoeff, shift);
     47 }
     48 
     49 static inline void calculate_qcoeff_log_scale(__m128i *coeff,
     50                                              const __m128i round,
     51                                              const __m128i quant,
     52                                              const __m128i *shift,
     53                                              const int *log_scale) {
     54  __m128i tmp, tmp1, qcoeff;
     55  qcoeff = _mm_adds_epi16(*coeff, round);
     56  tmp = _mm_mulhi_epi16(qcoeff, quant);
     57  qcoeff = _mm_add_epi16(tmp, qcoeff);
     58  tmp = _mm_mullo_epi16(qcoeff, *shift);
     59  tmp = _mm_srli_epi16(tmp, (16 - *log_scale));
     60  tmp1 = _mm_mulhi_epi16(qcoeff, *shift);
     61  tmp1 = _mm_slli_epi16(tmp1, *log_scale);
     62  *coeff = _mm_or_si128(tmp, tmp1);
     63 }
     64 
     65 static inline __m128i calculate_dqcoeff(__m128i qcoeff, __m128i dequant) {
     66  return _mm_mullo_epi16(qcoeff, dequant);
     67 }
     68 
     69 static inline void calculate_dqcoeff_and_store_log_scale(__m128i qcoeff,
     70                                                         __m128i dequant,
     71                                                         const __m128i zero,
     72                                                         tran_low_t *dqcoeff,
     73                                                         const int *log_scale) {
     74  // calculate abs
     75  __m128i coeff_sign = _mm_srai_epi16(qcoeff, 15);
     76  __m128i coeff = invert_sign_sse2(qcoeff, coeff_sign);
     77 
     78  const __m128i sign_0 = _mm_unpacklo_epi16(coeff_sign, zero);
     79  const __m128i sign_1 = _mm_unpackhi_epi16(coeff_sign, zero);
     80 
     81  const __m128i low = _mm_mullo_epi16(coeff, dequant);
     82  const __m128i high = _mm_mulhi_epi16(coeff, dequant);
     83  __m128i dqcoeff32_0 = _mm_unpacklo_epi16(low, high);
     84  __m128i dqcoeff32_1 = _mm_unpackhi_epi16(low, high);
     85 
     86  dqcoeff32_0 = _mm_srli_epi32(dqcoeff32_0, *log_scale);
     87  dqcoeff32_1 = _mm_srli_epi32(dqcoeff32_1, *log_scale);
     88 
     89  dqcoeff32_0 = invert_sign_32_sse2(dqcoeff32_0, sign_0);
     90  dqcoeff32_1 = invert_sign_32_sse2(dqcoeff32_1, sign_1);
     91 
     92  _mm_store_si128((__m128i *)(dqcoeff), dqcoeff32_0);
     93  _mm_store_si128((__m128i *)(dqcoeff + 4), dqcoeff32_1);
     94 }
     95 
     96 // Scan 16 values for eob reference in scan_ptr. Use masks (-1) from comparing
     97 // to zbin to add 1 to the index in 'scan'.
     98 static inline __m128i scan_for_eob(__m128i *coeff0, __m128i *coeff1,
     99                                   const __m128i zbin_mask0,
    100                                   const __m128i zbin_mask1,
    101                                   const int16_t *scan_ptr, const int index,
    102                                   const __m128i zero) {
    103  const __m128i zero_coeff0 = _mm_cmpeq_epi16(*coeff0, zero);
    104  const __m128i zero_coeff1 = _mm_cmpeq_epi16(*coeff1, zero);
    105  __m128i scan0 = _mm_load_si128((const __m128i *)(scan_ptr + index));
    106  __m128i scan1 = _mm_load_si128((const __m128i *)(scan_ptr + index + 8));
    107  __m128i eob0, eob1;
    108  // Add one to convert from indices to counts
    109  scan0 = _mm_sub_epi16(scan0, zbin_mask0);
    110  scan1 = _mm_sub_epi16(scan1, zbin_mask1);
    111  eob0 = _mm_andnot_si128(zero_coeff0, scan0);
    112  eob1 = _mm_andnot_si128(zero_coeff1, scan1);
    113  return _mm_max_epi16(eob0, eob1);
    114 }
    115 
    116 static inline int16_t accumulate_eob(__m128i eob) {
    117  __m128i eob_shuffled;
    118  eob_shuffled = _mm_shuffle_epi32(eob, 0xe);
    119  eob = _mm_max_epi16(eob, eob_shuffled);
    120  eob_shuffled = _mm_shufflelo_epi16(eob, 0xe);
    121  eob = _mm_max_epi16(eob, eob_shuffled);
    122  eob_shuffled = _mm_shufflelo_epi16(eob, 0x1);
    123  eob = _mm_max_epi16(eob, eob_shuffled);
    124  return _mm_extract_epi16(eob, 1);
    125 }
    126 
    127 static inline __m128i load_coefficients(const tran_low_t *coeff_ptr) {
    128  assert(sizeof(tran_low_t) == 4);
    129  const __m128i coeff1 = _mm_load_si128((__m128i *)(coeff_ptr));
    130  const __m128i coeff2 = _mm_load_si128((__m128i *)(coeff_ptr + 4));
    131  return _mm_packs_epi32(coeff1, coeff2);
    132 }
    133 
    134 static inline void store_coefficients(__m128i coeff_vals,
    135                                      tran_low_t *coeff_ptr) {
    136  assert(sizeof(tran_low_t) == 4);
    137 
    138  __m128i one = _mm_set1_epi16(1);
    139  __m128i coeff_vals_hi = _mm_mulhi_epi16(coeff_vals, one);
    140  __m128i coeff_vals_lo = _mm_mullo_epi16(coeff_vals, one);
    141  __m128i coeff_vals_1 = _mm_unpacklo_epi16(coeff_vals_lo, coeff_vals_hi);
    142  __m128i coeff_vals_2 = _mm_unpackhi_epi16(coeff_vals_lo, coeff_vals_hi);
    143  _mm_store_si128((__m128i *)(coeff_ptr), coeff_vals_1);
    144  _mm_store_si128((__m128i *)(coeff_ptr + 4), coeff_vals_2);
    145 }
    146 
    147 static inline void update_mask1(__m128i *cmp_mask0, __m128i *cmp_mask1,
    148                                const int16_t *iscan_ptr, int *is_found,
    149                                __m128i *mask) {
    150  __m128i all_zero;
    151  __m128i temp_mask = _mm_setzero_si128();
    152  all_zero = _mm_or_si128(*cmp_mask0, *cmp_mask1);
    153  if (_mm_movemask_epi8(all_zero)) {
    154    __m128i iscan0 = _mm_load_si128((const __m128i *)(iscan_ptr));
    155    __m128i mask0 = _mm_and_si128(*cmp_mask0, iscan0);
    156    __m128i iscan1 = _mm_load_si128((const __m128i *)(iscan_ptr + 8));
    157    __m128i mask1 = _mm_and_si128(*cmp_mask1, iscan1);
    158    temp_mask = _mm_max_epi16(mask0, mask1);
    159    *is_found = 1;
    160  }
    161  *mask = _mm_max_epi16(temp_mask, *mask);
    162 }
    163 
    164 static inline void update_mask0(__m128i *qcoeff0, __m128i *qcoeff1,
    165                                __m128i *threshold, const int16_t *iscan_ptr,
    166                                int *is_found, __m128i *mask) {
    167  __m128i zero = _mm_setzero_si128();
    168  __m128i coeff[4], cmp_mask0, cmp_mask1, cmp_mask2, cmp_mask3;
    169 
    170  coeff[0] = _mm_unpacklo_epi16(*qcoeff0, zero);
    171  coeff[1] = _mm_unpackhi_epi16(*qcoeff0, zero);
    172  coeff[2] = _mm_unpacklo_epi16(*qcoeff1, zero);
    173  coeff[3] = _mm_unpackhi_epi16(*qcoeff1, zero);
    174 
    175  coeff[0] = _mm_slli_epi32(coeff[0], AOM_QM_BITS);
    176  cmp_mask0 = _mm_cmpgt_epi32(coeff[0], threshold[0]);
    177  coeff[1] = _mm_slli_epi32(coeff[1], AOM_QM_BITS);
    178  cmp_mask1 = _mm_cmpgt_epi32(coeff[1], threshold[1]);
    179  coeff[2] = _mm_slli_epi32(coeff[2], AOM_QM_BITS);
    180  cmp_mask2 = _mm_cmpgt_epi32(coeff[2], threshold[1]);
    181  coeff[3] = _mm_slli_epi32(coeff[3], AOM_QM_BITS);
    182  cmp_mask3 = _mm_cmpgt_epi32(coeff[3], threshold[1]);
    183 
    184  cmp_mask0 = _mm_packs_epi32(cmp_mask0, cmp_mask1);
    185  cmp_mask1 = _mm_packs_epi32(cmp_mask2, cmp_mask3);
    186 
    187  update_mask1(&cmp_mask0, &cmp_mask1, iscan_ptr, is_found, mask);
    188 }
    189 
    190 static inline int calculate_non_zero_count(__m128i mask) {
    191  __m128i mask0, mask1;
    192  int non_zero_count = 0;
    193  mask0 = _mm_unpackhi_epi64(mask, mask);
    194  mask1 = _mm_max_epi16(mask0, mask);
    195  mask0 = _mm_shuffle_epi32(mask1, 1);
    196  mask0 = _mm_max_epi16(mask0, mask1);
    197  mask1 = _mm_srli_epi32(mask0, 16);
    198  mask0 = _mm_max_epi16(mask0, mask1);
    199  non_zero_count = _mm_extract_epi16(mask0, 0) + 1;
    200 
    201  return non_zero_count;
    202 }