Home | History | Annotate | Download | only in compress

Lines Matching refs:ip1

168     const BYTE* ip1;
178 size_t hash1; /* hash for ip1 */
186 /* ip0 and ip1 are always adjacent. The targetLength skipping and
189 * between pairs of positions, from ip0 to ip2 or ip1 to ip3. */
210 ip1 = ip0 + 1;
219 hash1 = ZSTD_hashPtr(ip1, hlog, mls);
242 * This write is known to be safe because the ip1 is before the
244 hashTable[hash1] = (U32)(ip1 - base);
261 * This write is known to be safe because the ip1 == ip0 + 1, so
262 * we know we will resume searching after ip1 */
263 hashTable[hash1] = (U32)(ip1 - base);
276 ip0 = ip1;
277 ip1 = ip2;
303 * ip1 is ip0 + step - 1. If ip1 is >= ip0 + 4, we can't safely
306 hashTable[hash1] = (U32)(ip1 - base);
320 ip0 = ip1;
321 ip1 = ip2;
323 ip3 = ip1 + step;
328 PREFETCH_L1(ip1 + 64);
329 PREFETCH_L1(ip1 + 128);
477 const BYTE* ip1 = ip0 + stepSize; /* we assert below that stepSize >= 1 */
524 while (ip1 <= ilimit) { /* repcode check at (ip0 + 1) is safe because ip0 < ip1 */
545 const size_t hash1 = ZSTD_hashPtr(ip1, hlog, mls);
546 size_t const dictHashAndTag1 = ZSTD_hashPtr(ip1, dictHBits, mls);
604 if (ip1 >= nextStep) {
608 ip0 = ip1;
609 ip1 = ip1 + step;
610 if (ip1 > ilimit) goto _cleanup;
651 ip1 = ip0 + stepSize;
719 const BYTE* ip1;
726 size_t hash1; /* hash for ip1 */
760 ip1 = ip0 + 1;
769 hash1 = ZSTD_hashPtr(ip1, hlog, mls);
825 ip0 = ip1;
826 ip1 = ip2;
853 ip0 = ip1;
854 ip1 = ip2;
856 ip3 = ip1 + step;
861 PREFETCH_L1(ip1 + 64);
862 PREFETCH_L1(ip1 + 128);
914 if (ip1 < ip0) {
915 hashTable[hash1] = (U32)(ip1 - base);