sljitNativeX86_32.c revision 1.2 1 /*
2 * Stack-less Just-In-Time compiler
3 *
4 * Copyright 2009-2012 Zoltan Herczeg (hzmester (at) freemail.hu). All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without modification, are
7 * permitted provided that the following conditions are met:
8 *
9 * 1. Redistributions of source code must retain the above copyright notice, this list of
10 * conditions and the following disclaimer.
11 *
12 * 2. Redistributions in binary form must reproduce the above copyright notice, this list
13 * of conditions and the following disclaimer in the documentation and/or other materials
14 * provided with the distribution.
15 *
16 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDER(S) AND CONTRIBUTORS ``AS IS'' AND ANY
17 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT
19 * SHALL THE COPYRIGHT HOLDER(S) OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
20 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
21 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
22 * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
24 * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
25 */
26
27 /* x86 32-bit arch dependent functions. */
28
29 static int emit_do_imm(struct sljit_compiler *compiler, sljit_ub opcode, sljit_w imm)
30 {
31 sljit_ub *buf;
32
33 buf = (sljit_ub*)ensure_buf(compiler, 1 + 1 + sizeof(sljit_w));
34 FAIL_IF(!buf);
35 INC_SIZE(1 + sizeof(sljit_w));
36 *buf++ = opcode;
37 *(sljit_w*)buf = imm;
38 return SLJIT_SUCCESS;
39 }
40
41 static sljit_ub* generate_far_jump_code(struct sljit_jump *jump, sljit_ub *code_ptr, int type)
42 {
43 if (type == SLJIT_JUMP) {
44 *code_ptr++ = 0xe9;
45 jump->addr++;
46 }
47 else if (type >= SLJIT_FAST_CALL) {
48 *code_ptr++ = 0xe8;
49 jump->addr++;
50 }
51 else {
52 *code_ptr++ = 0x0f;
53 *code_ptr++ = get_jump_code(type);
54 jump->addr += 2;
55 }
56
57 if (jump->flags & JUMP_LABEL)
58 jump->flags |= PATCH_MW;
59 else
60 *(sljit_w*)code_ptr = jump->u.target - (jump->addr + 4);
61 code_ptr += 4;
62
63 return code_ptr;
64 }
65
66 SLJIT_API_FUNC_ATTRIBUTE int sljit_emit_enter(struct sljit_compiler *compiler, int args, int temporaries, int saveds, int local_size)
67 {
68 int size;
69 int locals_offset;
70 sljit_ub *buf;
71
72 CHECK_ERROR();
73 check_sljit_emit_enter(compiler, args, temporaries, saveds, local_size);
74
75 compiler->temporaries = temporaries;
76 compiler->saveds = saveds;
77 compiler->args = args;
78 compiler->flags_saved = 0;
79 #if (defined SLJIT_DEBUG && SLJIT_DEBUG)
80 compiler->logical_local_size = local_size;
81 #endif
82
83 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
84 size = 1 + (saveds <= 3 ? saveds : 3) + (args > 0 ? (args * 2) : 0) + (args > 2 ? 2 : 0);
85 #else
86 size = 1 + (saveds <= 3 ? saveds : 3) + (args > 0 ? (2 + args * 3) : 0);
87 #endif
88 buf = (sljit_ub*)ensure_buf(compiler, 1 + size);
89 FAIL_IF(!buf);
90
91 INC_SIZE(size);
92 PUSH_REG(reg_map[TMP_REGISTER]);
93 #if !(defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
94 if (args > 0) {
95 *buf++ = 0x8b;
96 *buf++ = 0xc4 | (reg_map[TMP_REGISTER] << 3);
97 }
98 #endif
99 if (saveds > 2)
100 PUSH_REG(reg_map[SLJIT_SAVED_REG3]);
101 if (saveds > 1)
102 PUSH_REG(reg_map[SLJIT_SAVED_REG2]);
103 if (saveds > 0)
104 PUSH_REG(reg_map[SLJIT_SAVED_REG1]);
105
106 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
107 if (args > 0) {
108 *buf++ = 0x8b;
109 *buf++ = 0xc0 | (reg_map[SLJIT_SAVED_REG1] << 3) | reg_map[SLJIT_TEMPORARY_REG3];
110 }
111 if (args > 1) {
112 *buf++ = 0x8b;
113 *buf++ = 0xc0 | (reg_map[SLJIT_SAVED_REG2] << 3) | reg_map[SLJIT_TEMPORARY_REG2];
114 }
115 if (args > 2) {
116 *buf++ = 0x8b;
117 *buf++ = 0x44 | (reg_map[SLJIT_SAVED_REG3] << 3);
118 *buf++ = 0x24;
119 *buf++ = sizeof(sljit_w) * (3 + 2); /* saveds >= 3 as well. */
120 }
121 #else
122 if (args > 0) {
123 *buf++ = 0x8b;
124 *buf++ = 0x40 | (reg_map[SLJIT_SAVED_REG1] << 3) | reg_map[TMP_REGISTER];
125 *buf++ = sizeof(sljit_w) * 2;
126 }
127 if (args > 1) {
128 *buf++ = 0x8b;
129 *buf++ = 0x40 | (reg_map[SLJIT_SAVED_REG2] << 3) | reg_map[TMP_REGISTER];
130 *buf++ = sizeof(sljit_w) * 3;
131 }
132 if (args > 2) {
133 *buf++ = 0x8b;
134 *buf++ = 0x40 | (reg_map[SLJIT_SAVED_REG3] << 3) | reg_map[TMP_REGISTER];
135 *buf++ = sizeof(sljit_w) * 4;
136 }
137 #endif
138
139 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
140 locals_offset = 2 * sizeof(sljit_uw);
141 #else
142 locals_offset = 2 * sizeof(sljit_uw) + FIXED_LOCALS_OFFSET;
143 #endif
144 compiler->temporaries_start = locals_offset;
145 if (temporaries > 3)
146 locals_offset += (temporaries - 3) * sizeof(sljit_uw);
147 compiler->saveds_start = locals_offset;
148 if (saveds > 3)
149 locals_offset += (saveds - 3) * sizeof(sljit_uw);
150 compiler->locals_offset = locals_offset;
151 local_size = locals_offset + ((local_size + sizeof(sljit_uw) - 1) & ~(sizeof(sljit_uw) - 1));
152
153 compiler->local_size = local_size;
154 #ifdef _WIN32
155 if (local_size > 1024) {
156 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
157 FAIL_IF(emit_do_imm(compiler, 0xb8 + reg_map[SLJIT_TEMPORARY_REG1], local_size));
158 #else
159 local_size -= FIXED_LOCALS_OFFSET;
160 FAIL_IF(emit_do_imm(compiler, 0xb8 + reg_map[SLJIT_TEMPORARY_REG1], local_size));
161 FAIL_IF(emit_non_cum_binary(compiler, SUB_r_rm, SUB_rm_r, SUB, SUB_EAX_i32,
162 SLJIT_LOCALS_REG, 0, SLJIT_LOCALS_REG, 0, SLJIT_IMM, FIXED_LOCALS_OFFSET));
163 #endif
164 FAIL_IF(sljit_emit_ijump(compiler, SLJIT_CALL1, SLJIT_IMM, SLJIT_FUNC_OFFSET(sljit_grow_stack)));
165 }
166 #endif
167
168 SLJIT_ASSERT(local_size > 0);
169 return emit_non_cum_binary(compiler, 0x2b, 0x29, 0x5 << 3, 0x2d,
170 SLJIT_LOCALS_REG, 0, SLJIT_LOCALS_REG, 0, SLJIT_IMM, local_size);
171 }
172
173 SLJIT_API_FUNC_ATTRIBUTE void sljit_set_context(struct sljit_compiler *compiler, int args, int temporaries, int saveds, int local_size)
174 {
175 int locals_offset;
176
177 CHECK_ERROR_VOID();
178 check_sljit_set_context(compiler, args, temporaries, saveds, local_size);
179
180 compiler->temporaries = temporaries;
181 compiler->saveds = saveds;
182 compiler->args = args;
183 #if (defined SLJIT_DEBUG && SLJIT_DEBUG)
184 compiler->logical_local_size = local_size;
185 #endif
186
187 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
188 locals_offset = 2 * sizeof(sljit_uw);
189 #else
190 locals_offset = 2 * sizeof(sljit_uw) + FIXED_LOCALS_OFFSET;
191 #endif
192 compiler->temporaries_start = locals_offset;
193 if (temporaries > 3)
194 locals_offset += (temporaries - 3) * sizeof(sljit_uw);
195 compiler->saveds_start = locals_offset;
196 if (saveds > 3)
197 locals_offset += (saveds - 3) * sizeof(sljit_uw);
198 compiler->locals_offset = locals_offset;
199 compiler->local_size = locals_offset + ((local_size + sizeof(sljit_uw) - 1) & ~(sizeof(sljit_uw) - 1));
200 }
201
202 SLJIT_API_FUNC_ATTRIBUTE int sljit_emit_return(struct sljit_compiler *compiler, int op, int src, sljit_w srcw)
203 {
204 int size;
205 sljit_ub *buf;
206
207 CHECK_ERROR();
208 check_sljit_emit_return(compiler, op, src, srcw);
209 SLJIT_ASSERT(compiler->args >= 0);
210
211 compiler->flags_saved = 0;
212 FAIL_IF(emit_mov_before_return(compiler, op, src, srcw));
213
214 SLJIT_ASSERT(compiler->local_size > 0);
215 FAIL_IF(emit_cum_binary(compiler, 0x03, 0x01, 0x0 << 3, 0x05,
216 SLJIT_LOCALS_REG, 0, SLJIT_LOCALS_REG, 0, SLJIT_IMM, compiler->local_size));
217
218 size = 2 + (compiler->saveds <= 3 ? compiler->saveds : 3);
219 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
220 if (compiler->args > 2)
221 size += 2;
222 #else
223 if (compiler->args > 0)
224 size += 2;
225 #endif
226 buf = (sljit_ub*)ensure_buf(compiler, 1 + size);
227 FAIL_IF(!buf);
228
229 INC_SIZE(size);
230
231 if (compiler->saveds > 0)
232 POP_REG(reg_map[SLJIT_SAVED_REG1]);
233 if (compiler->saveds > 1)
234 POP_REG(reg_map[SLJIT_SAVED_REG2]);
235 if (compiler->saveds > 2)
236 POP_REG(reg_map[SLJIT_SAVED_REG3]);
237 POP_REG(reg_map[TMP_REGISTER]);
238 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
239 if (compiler->args > 2)
240 RETN(sizeof(sljit_w));
241 else
242 RET();
243 #else
244 RET();
245 #endif
246
247 return SLJIT_SUCCESS;
248 }
249
250 /* --------------------------------------------------------------------- */
251 /* Operators */
252 /* --------------------------------------------------------------------- */
253
254 /* Size contains the flags as well. */
255 static sljit_ub* emit_x86_instruction(struct sljit_compiler *compiler, int size,
256 /* The register or immediate operand. */
257 int a, sljit_w imma,
258 /* The general operand (not immediate). */
259 int b, sljit_w immb)
260 {
261 sljit_ub *buf;
262 sljit_ub *buf_ptr;
263 int flags = size & ~0xf;
264 int inst_size;
265
266 /* Both cannot be switched on. */
267 SLJIT_ASSERT((flags & (EX86_BIN_INS | EX86_SHIFT_INS)) != (EX86_BIN_INS | EX86_SHIFT_INS));
268 /* Size flags not allowed for typed instructions. */
269 SLJIT_ASSERT(!(flags & (EX86_BIN_INS | EX86_SHIFT_INS)) || (flags & (EX86_BYTE_ARG | EX86_HALF_ARG)) == 0);
270 /* Both size flags cannot be switched on. */
271 SLJIT_ASSERT((flags & (EX86_BYTE_ARG | EX86_HALF_ARG)) != (EX86_BYTE_ARG | EX86_HALF_ARG));
272 #if (defined SLJIT_SSE2 && SLJIT_SSE2)
273 /* SSE2 and immediate is not possible. */
274 SLJIT_ASSERT(!(a & SLJIT_IMM) || !(flags & EX86_SSE2));
275 #endif
276
277 size &= 0xf;
278 inst_size = size;
279
280 #if (defined SLJIT_SSE2 && SLJIT_SSE2)
281 if (flags & EX86_PREF_F2)
282 inst_size++;
283 #endif
284 if (flags & EX86_PREF_66)
285 inst_size++;
286
287 /* Calculate size of b. */
288 inst_size += 1; /* mod r/m byte. */
289 if (b & SLJIT_MEM) {
290 if ((b & 0x0f) == SLJIT_UNUSED)
291 inst_size += sizeof(sljit_w);
292 else if (immb != 0 && !(b & 0xf0)) {
293 /* Immediate operand. */
294 if (immb <= 127 && immb >= -128)
295 inst_size += sizeof(sljit_b);
296 else
297 inst_size += sizeof(sljit_w);
298 }
299
300 if ((b & 0xf) == SLJIT_LOCALS_REG && !(b & 0xf0))
301 b |= SLJIT_LOCALS_REG << 4;
302
303 if ((b & 0xf0) != SLJIT_UNUSED)
304 inst_size += 1; /* SIB byte. */
305 }
306
307 /* Calculate size of a. */
308 if (a & SLJIT_IMM) {
309 if (flags & EX86_BIN_INS) {
310 if (imma <= 127 && imma >= -128) {
311 inst_size += 1;
312 flags |= EX86_BYTE_ARG;
313 } else
314 inst_size += 4;
315 }
316 else if (flags & EX86_SHIFT_INS) {
317 imma &= 0x1f;
318 if (imma != 1) {
319 inst_size ++;
320 flags |= EX86_BYTE_ARG;
321 }
322 } else if (flags & EX86_BYTE_ARG)
323 inst_size++;
324 else if (flags & EX86_HALF_ARG)
325 inst_size += sizeof(short);
326 else
327 inst_size += sizeof(sljit_w);
328 }
329 else
330 SLJIT_ASSERT(!(flags & EX86_SHIFT_INS) || a == SLJIT_PREF_SHIFT_REG);
331
332 buf = (sljit_ub*)ensure_buf(compiler, 1 + inst_size);
333 PTR_FAIL_IF(!buf);
334
335 /* Encoding the byte. */
336 INC_SIZE(inst_size);
337 #if (defined SLJIT_SSE2 && SLJIT_SSE2)
338 if (flags & EX86_PREF_F2)
339 *buf++ = 0xf2;
340 #endif
341 if (flags & EX86_PREF_66)
342 *buf++ = 0x66;
343
344 buf_ptr = buf + size;
345
346 /* Encode mod/rm byte. */
347 if (!(flags & EX86_SHIFT_INS)) {
348 if ((flags & EX86_BIN_INS) && (a & SLJIT_IMM))
349 *buf = (flags & EX86_BYTE_ARG) ? 0x83 : 0x81;
350
351 if ((a & SLJIT_IMM) || (a == 0))
352 *buf_ptr = 0;
353 #if (defined SLJIT_SSE2 && SLJIT_SSE2)
354 else if (!(flags & EX86_SSE2))
355 *buf_ptr = reg_map[a] << 3;
356 else
357 *buf_ptr = a << 3;
358 #else
359 else
360 *buf_ptr = reg_map[a] << 3;
361 #endif
362 }
363 else {
364 if (a & SLJIT_IMM) {
365 if (imma == 1)
366 *buf = 0xd1;
367 else
368 *buf = 0xc1;
369 } else
370 *buf = 0xd3;
371 *buf_ptr = 0;
372 }
373
374 if (!(b & SLJIT_MEM))
375 #if (defined SLJIT_SSE2 && SLJIT_SSE2)
376 *buf_ptr++ |= 0xc0 + ((!(flags & EX86_SSE2)) ? reg_map[b] : b);
377 #else
378 *buf_ptr++ |= 0xc0 + reg_map[b];
379 #endif
380 else if ((b & 0x0f) != SLJIT_UNUSED) {
381 if ((b & 0xf0) == SLJIT_UNUSED || (b & 0xf0) == (SLJIT_LOCALS_REG << 4)) {
382 if (immb != 0) {
383 if (immb <= 127 && immb >= -128)
384 *buf_ptr |= 0x40;
385 else
386 *buf_ptr |= 0x80;
387 }
388
389 if ((b & 0xf0) == SLJIT_UNUSED)
390 *buf_ptr++ |= reg_map[b & 0x0f];
391 else {
392 *buf_ptr++ |= 0x04;
393 *buf_ptr++ = reg_map[b & 0x0f] | (reg_map[(b >> 4) & 0x0f] << 3);
394 }
395
396 if (immb != 0) {
397 if (immb <= 127 && immb >= -128)
398 *buf_ptr++ = immb; /* 8 bit displacement. */
399 else {
400 *(sljit_w*)buf_ptr = immb; /* 32 bit displacement. */
401 buf_ptr += sizeof(sljit_w);
402 }
403 }
404 }
405 else {
406 *buf_ptr++ |= 0x04;
407 *buf_ptr++ = reg_map[b & 0x0f] | (reg_map[(b >> 4) & 0x0f] << 3) | (immb << 6);
408 }
409 }
410 else {
411 *buf_ptr++ |= 0x05;
412 *(sljit_w*)buf_ptr = immb; /* 32 bit displacement. */
413 buf_ptr += sizeof(sljit_w);
414 }
415
416 if (a & SLJIT_IMM) {
417 if (flags & EX86_BYTE_ARG)
418 *buf_ptr = imma;
419 else if (flags & EX86_HALF_ARG)
420 *(short*)buf_ptr = imma;
421 else if (!(flags & EX86_SHIFT_INS))
422 *(sljit_w*)buf_ptr = imma;
423 }
424
425 return !(flags & EX86_SHIFT_INS) ? buf : (buf + 1);
426 }
427
428 /* --------------------------------------------------------------------- */
429 /* Call / return instructions */
430 /* --------------------------------------------------------------------- */
431
432 static SLJIT_INLINE int call_with_args(struct sljit_compiler *compiler, int type)
433 {
434 sljit_ub *buf;
435
436 #if (defined SLJIT_X86_32_FASTCALL && SLJIT_X86_32_FASTCALL)
437 buf = (sljit_ub*)ensure_buf(compiler, type >= SLJIT_CALL3 ? 1 + 2 + 1 : 1 + 2);
438 FAIL_IF(!buf);
439 INC_SIZE(type >= SLJIT_CALL3 ? 2 + 1 : 2);
440
441 if (type >= SLJIT_CALL3)
442 PUSH_REG(reg_map[SLJIT_TEMPORARY_REG3]);
443 *buf++ = 0x8b;
444 *buf++ = 0xc0 | (reg_map[SLJIT_TEMPORARY_REG3] << 3) | reg_map[SLJIT_TEMPORARY_REG1];
445 #else
446 buf = (sljit_ub*)ensure_buf(compiler, 1 + 4 * (type - SLJIT_CALL0));
447 FAIL_IF(!buf);
448 INC_SIZE(4 * (type - SLJIT_CALL0));
449
450 *buf++ = 0x89;
451 *buf++ = 0x40 | (reg_map[SLJIT_TEMPORARY_REG1] << 3) | 0x4 /* SIB */;
452 *buf++ = (0x4 /* none*/ << 3) | reg_map[SLJIT_LOCALS_REG];
453 *buf++ = 0;
454 if (type >= SLJIT_CALL2) {
455 *buf++ = 0x89;
456 *buf++ = 0x40 | (reg_map[SLJIT_TEMPORARY_REG2] << 3) | 0x4 /* SIB */;
457 *buf++ = (0x4 /* none*/ << 3) | reg_map[SLJIT_LOCALS_REG];
458 *buf++ = sizeof(sljit_w);
459 }
460 if (type >= SLJIT_CALL3) {
461 *buf++ = 0x89;
462 *buf++ = 0x40 | (reg_map[SLJIT_TEMPORARY_REG3] << 3) | 0x4 /* SIB */;
463 *buf++ = (0x4 /* none*/ << 3) | reg_map[SLJIT_LOCALS_REG];
464 *buf++ = 2 * sizeof(sljit_w);
465 }
466 #endif
467 return SLJIT_SUCCESS;
468 }
469
470 SLJIT_API_FUNC_ATTRIBUTE int sljit_emit_fast_enter(struct sljit_compiler *compiler, int dst, sljit_w dstw)
471 {
472 sljit_ub *buf;
473
474 CHECK_ERROR();
475 check_sljit_emit_fast_enter(compiler, dst, dstw);
476 ADJUST_LOCAL_OFFSET(dst, dstw);
477
478 CHECK_EXTRA_REGS(dst, dstw, (void)0);
479
480 if (dst >= SLJIT_TEMPORARY_REG1 && dst <= SLJIT_NO_REGISTERS) {
481 buf = (sljit_ub*)ensure_buf(compiler, 1 + 1);
482 FAIL_IF(!buf);
483
484 INC_SIZE(1);
485 POP_REG(reg_map[dst]);
486 return SLJIT_SUCCESS;
487 }
488 else if (dst & SLJIT_MEM) {
489 buf = emit_x86_instruction(compiler, 1, 0, 0, dst, dstw);
490 FAIL_IF(!buf);
491 *buf++ = 0x8f;
492 return SLJIT_SUCCESS;
493 }
494
495 /* For UNUSED dst. Uncommon, but possible. */
496 buf = (sljit_ub*)ensure_buf(compiler, 1 + 1);
497 FAIL_IF(!buf);
498
499 INC_SIZE(1);
500 POP_REG(reg_map[TMP_REGISTER]);
501 return SLJIT_SUCCESS;
502 }
503
504 SLJIT_API_FUNC_ATTRIBUTE int sljit_emit_fast_return(struct sljit_compiler *compiler, int src, sljit_w srcw)
505 {
506 sljit_ub *buf;
507
508 CHECK_ERROR();
509 check_sljit_emit_fast_return(compiler, src, srcw);
510 ADJUST_LOCAL_OFFSET(src, srcw);
511
512 CHECK_EXTRA_REGS(src, srcw, (void)0);
513
514 if (src >= SLJIT_TEMPORARY_REG1 && src <= SLJIT_NO_REGISTERS) {
515 buf = (sljit_ub*)ensure_buf(compiler, 1 + 1 + 1);
516 FAIL_IF(!buf);
517
518 INC_SIZE(1 + 1);
519 PUSH_REG(reg_map[src]);
520 }
521 else if (src & SLJIT_MEM) {
522 buf = emit_x86_instruction(compiler, 1, 0, 0, src, srcw);
523 FAIL_IF(!buf);
524 *buf++ = 0xff;
525 *buf |= 6 << 3;
526
527 buf = (sljit_ub*)ensure_buf(compiler, 1 + 1);
528 FAIL_IF(!buf);
529 INC_SIZE(1);
530 }
531 else {
532 /* SLJIT_IMM. */
533 buf = (sljit_ub*)ensure_buf(compiler, 1 + 5 + 1);
534 FAIL_IF(!buf);
535
536 INC_SIZE(5 + 1);
537 *buf++ = 0x68;
538 *(sljit_w*)buf = srcw;
539 buf += sizeof(sljit_w);
540 }
541
542 RET();
543 return SLJIT_SUCCESS;
544 }
545