/src/sys/external/bsd/sljit/dist/sljit_src/ |
sljitNativeARM_32.c | 267 static sljit_uw patch_pc_relative_loads(sljit_uw *last_pc_patch, sljit_uw *code_ptr, sljit_uw* const_pool, sljit_uw cpool_size) 275 SLJIT_ASSERT(const_pool - code_ptr <= CONST_POOL_ALIGNMENT); 282 while (last_pc_patch < code_ptr) { 393 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_uw *code_ptr, sljit_uw *code, sljit_sw executable_offset) 402 code_ptr--; 405 diff = ((sljit_sw)jump->u.target - (sljit_sw)(code_ptr + 2) - executable_offset); 408 diff = ((sljit_sw)(code + jump->u.label->size) - (sljit_sw)(code_ptr + 2)); 417 *code_ptr = (BL - CONDITIONAL) | (*(code_ptr + 1) & COND_MASK); 424 *code_ptr = (B - CONDITIONAL) | (*code_ptr & COND_MASK) 570 sljit_uw *code_ptr; local in function:sljit_generate_code [all...] |
sljitNativeARM_T2_32.c | 225 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_u16 *code_ptr, sljit_u16 *code, sljit_sw executable_offset) 236 diff = ((sljit_sw)jump->u.target - (sljit_sw)(code_ptr + 2) - executable_offset) >> 1; 240 diff = ((sljit_sw)(code + jump->u.label->size) - (sljit_sw)(code_ptr + 2)) >> 1; 349 sljit_u16 *code_ptr; local in function:sljit_generate_code 367 code_ptr = code; 379 *code_ptr = *buf_ptr++; 385 label->addr = ((sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset)) | 0x1; 386 label->size = code_ptr - code; 390 jump->addr = (sljit_uw)code_ptr - ((jump->flags & IS_COND) ? 10 : 8); 391 code_ptr -= detect_jump_type(jump, code_ptr, code, executable_offset) [all...] |
sljitNativePPC_common.c | 254 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset) 280 diff = ((sljit_sw)target_addr - (sljit_sw)(code_ptr) - executable_offset) & ~0x3l; 329 sljit_ins *code_ptr; local in function:sljit_generate_code 355 code_ptr = code; 367 *code_ptr = *buf_ptr++; 374 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 375 label->size = code_ptr - code; 380 jump->addr = (sljit_uw)(code_ptr - 3); 382 jump->addr = (sljit_uw)(code_ptr - 6); 384 if (detect_jump_type(jump, code_ptr, code, executable_offset)) [all...] |
sljitNativeSPARC_common.c | 204 static SLJIT_INLINE sljit_ins* detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset) 212 return code_ptr; 276 return code_ptr; 283 sljit_ins *code_ptr; local in function:sljit_generate_code 302 code_ptr = code; 314 *code_ptr = *buf_ptr++; 321 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset); 322 label->size = code_ptr - code; 327 jump->addr = (sljit_uw)(code_ptr - 3); 329 jump->addr = (sljit_uw)(code_ptr - 6) [all...] |
sljitNativeX86_common.c | 418 static sljit_u8* generate_far_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr, sljit_s32 type, sljit_sw executable_offset); 420 static sljit_u8* generate_far_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr, sljit_s32 type); 423 static sljit_u8* generate_near_jump_code(struct sljit_jump *jump, sljit_u8 *code_ptr, sljit_u8 *code, sljit_s32 type, sljit_sw executable_offset) 437 return generate_far_jump_code(jump, code_ptr, type); 442 *code_ptr++ = JMP_i8; 444 *code_ptr++ = JMP_i32; 449 *code_ptr++ = CALL_i32; 453 *code_ptr++ = get_jump_code(type) - 0x10; 457 *code_ptr++ = GROUP_0F; 458 *code_ptr++ = get_jump_code(type) 481 sljit_u8 *code_ptr; local in function:sljit_generate_code [all...] |
sljitNativeARM_64.c | 156 static SLJIT_INLINE sljit_s32 detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset) 173 diff = (sljit_sw)target_addr - (sljit_sw)(code_ptr + 4) - executable_offset; 178 code_ptr[-5] ^= (jump->flags & IS_CBZ) ? (0x1 << 24) : 0x1; 193 code_ptr[-5] -= (2 << 5); 194 code_ptr[-2] = code_ptr[0]; 199 code_ptr[-5] -= (1 << 5); 201 code_ptr[-1] = code_ptr[0]; 213 sljit_ins *code_ptr; local in function:sljit_generate_code [all...] |
sljitNativeMIPS_common.c | 218 static SLJIT_INLINE sljit_ins* detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code, sljit_sw executable_offset) 227 return code_ptr; 230 return code_ptr; 347 return code_ptr; 351 static __attribute__ ((noinline)) void sljit_cache_flush(void* code, void* code_ptr) 353 SLJIT_CACHE_FLUSH(code, code_ptr); 361 sljit_ins *code_ptr; local in function:sljit_generate_code 380 code_ptr = code; 392 *code_ptr = *buf_ptr++; 398 label->addr = (sljit_uw)SLJIT_ADD_EXEC_OFFSET(code_ptr, executable_offset) [all...] |
sljitNativeTILEGX_64.c | 905 static SLJIT_INLINE sljit_ins * detect_jump_type(struct sljit_jump *jump, sljit_ins *code_ptr, sljit_ins *code) 912 return code_ptr; 969 return code_ptr; 995 return code_ptr; 1002 sljit_ins *code_ptr; local in function:sljit_generate_code 1020 code_ptr = code; 1029 *code_ptr = *buf_ptr++; 1036 label->addr = (sljit_uw) code_ptr; 1037 label->size = code_ptr - code; 1043 jump->addr = (sljit_uw)(code_ptr - 4) [all...] |