mirror of
https://github.com/php/php-src.git
synced 2025-08-15 13:38:49 +02:00
238 lines
8.2 KiB
C
238 lines
8.2 KiB
C
/*
|
|
+----------------------------------------------------------------------+
|
|
| Zend OPcache |
|
|
+----------------------------------------------------------------------+
|
|
| Copyright (c) The PHP Group |
|
|
+----------------------------------------------------------------------+
|
|
| This source file is subject to version 3.01 of the PHP license, |
|
|
| that is bundled with this package in the file LICENSE, and is |
|
|
| available through the world-wide-web at the following url: |
|
|
| https://www.php.net/license/3_01.txt |
|
|
| If you did not receive a copy of the PHP license and are unable to |
|
|
| obtain it through the world-wide-web, please send a note to |
|
|
| license@php.net so we can mail you a copy immediately. |
|
|
+----------------------------------------------------------------------+
|
|
| Authors: Andi Gutmans <andi@php.net> |
|
|
| Zeev Suraski <zeev@php.net> |
|
|
| Stanislav Malyshev <stas@zend.com> |
|
|
| Dmitry Stogov <dmitry@php.net> |
|
|
+----------------------------------------------------------------------+
|
|
*/
|
|
|
|
/* pass 3: (Jump optimization)
|
|
* - optimize series of JMPs
|
|
*/
|
|
|
|
#include "Optimizer/zend_optimizer.h"
|
|
#include "Optimizer/zend_optimizer_internal.h"
|
|
#include "zend_API.h"
|
|
#include "zend_constants.h"
|
|
#include "zend_execute.h"
|
|
#include "zend_vm.h"
|
|
|
|
/* we use "jmp_hitlist" to avoid infinity loops during jmp optimization */
|
|
static zend_always_inline bool in_hitlist(zend_op *target, zend_op **jmp_hitlist, int jmp_hitlist_count)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < jmp_hitlist_count; i++) {
|
|
if (jmp_hitlist[i] == target) {
|
|
return 1;
|
|
}
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#define CHECK_LOOP(target) \
|
|
if (EXPECTED(!in_hitlist(target, jmp_hitlist, jmp_hitlist_count))) { \
|
|
jmp_hitlist[jmp_hitlist_count++] = target; \
|
|
} else { \
|
|
break; \
|
|
}
|
|
|
|
void zend_optimizer_pass3(zend_op_array *op_array, zend_optimizer_ctx *ctx)
|
|
{
|
|
zend_op *opline;
|
|
zend_op *end;
|
|
zend_op *target;
|
|
zend_op **jmp_hitlist;
|
|
int jmp_hitlist_count;
|
|
ALLOCA_FLAG(use_heap);
|
|
|
|
jmp_hitlist = (zend_op**)do_alloca(sizeof(zend_op*)*op_array->last, use_heap);
|
|
opline = op_array->opcodes;
|
|
end = opline + op_array->last;
|
|
|
|
while (opline < end) {
|
|
|
|
switch (opline->opcode) {
|
|
case ZEND_JMP:
|
|
jmp_hitlist_count = 0;
|
|
|
|
target = ZEND_OP1_JMP_ADDR(opline);
|
|
while (1) {
|
|
if (target->opcode == ZEND_JMP) {
|
|
/* convert JMP L1 ... L1: JMP L2 to JMP L2 .. L1: JMP L2 */
|
|
target = ZEND_OP1_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == ZEND_NOP) {
|
|
target = target + 1;
|
|
} else {
|
|
break;
|
|
}
|
|
ZEND_SET_OP_JMP_ADDR(opline, opline->op1, target);
|
|
}
|
|
|
|
if (target == opline + 1) {
|
|
/* convert L: JMP L+1 to NOP */
|
|
MAKE_NOP(opline);
|
|
} else if ((target->opcode == ZEND_RETURN ||
|
|
target->opcode == ZEND_RETURN_BY_REF ||
|
|
target->opcode == ZEND_GENERATOR_RETURN) &&
|
|
!(op_array->fn_flags & ZEND_ACC_HAS_FINALLY_BLOCK)) {
|
|
/* JMP L, L: RETURN to immediate RETURN */
|
|
*opline = *target;
|
|
if (opline->op1_type == IS_CONST) {
|
|
zval zv;
|
|
ZVAL_COPY(&zv, &ZEND_OP1_LITERAL(opline));
|
|
opline->op1.constant = zend_optimizer_add_literal(op_array, &zv);
|
|
}
|
|
} else if (opline > op_array->opcodes &&
|
|
((opline-1)->opcode == ZEND_JMPZ ||
|
|
(opline-1)->opcode == ZEND_JMPNZ)) {
|
|
if (ZEND_OP2_JMP_ADDR(opline-1) == target) {
|
|
/* JMPZ(X,L1), JMP(L1) -> NOP, JMP(L1) */
|
|
zend_optimizer_convert_to_free_op1(op_array, opline - 1);
|
|
}
|
|
}
|
|
break;
|
|
|
|
case ZEND_JMP_SET:
|
|
case ZEND_COALESCE:
|
|
jmp_hitlist_count = 0;
|
|
|
|
target = ZEND_OP2_JMP_ADDR(opline);
|
|
while (1) {
|
|
if (target->opcode == ZEND_JMP) {
|
|
target = ZEND_OP1_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == ZEND_NOP) {
|
|
target = target + 1;
|
|
} else {
|
|
break;
|
|
}
|
|
ZEND_SET_OP_JMP_ADDR(opline, opline->op2, target);
|
|
}
|
|
break;
|
|
|
|
case ZEND_JMPZ:
|
|
case ZEND_JMPNZ:
|
|
jmp_hitlist_count = 0;
|
|
|
|
target = ZEND_OP2_JMP_ADDR(opline);
|
|
while (1) {
|
|
if (target->opcode == ZEND_JMP) {
|
|
/* plain JMP */
|
|
/* JMPZ(X,L1), L1: JMP(L2) => JMPZ(X,L2), L1: JMP(L2) */
|
|
target = ZEND_OP1_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == opline->opcode &&
|
|
SAME_VAR(opline->op1, target->op1)) {
|
|
/* same opcode and same var as this opcode */
|
|
/* JMPZ(X,L1), L1: JMPZ(X,L2) => JMPZ(X,L2), L1: JMPZ(X,L2) */
|
|
target = ZEND_OP2_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == INV_COND(opline->opcode) &&
|
|
SAME_VAR(opline->op1, target->op1)) {
|
|
/* convert JMPZ(X,L1), L1: JMPNZ(X,L2) to
|
|
JMPZ(X,L1+1) */
|
|
target = target + 1;
|
|
} else if (target->opcode == ZEND_NOP) {
|
|
target = target + 1;
|
|
} else {
|
|
break;
|
|
}
|
|
ZEND_SET_OP_JMP_ADDR(opline, opline->op2, target);
|
|
}
|
|
|
|
/* convert L: JMPZ L+1 to NOP */
|
|
if (target == opline + 1) {
|
|
zend_optimizer_convert_to_free_op1(op_array, opline);
|
|
}
|
|
break;
|
|
|
|
case ZEND_JMPZ_EX:
|
|
case ZEND_JMPNZ_EX:
|
|
jmp_hitlist_count = 0;
|
|
|
|
target = ZEND_OP2_JMP_ADDR(opline);
|
|
while (1) {
|
|
if (target->opcode == ZEND_JMP) {
|
|
/* plain JMP */
|
|
/* JMPZ_EX(X,L1), L1: JMP(L2) => JMPZ_EX(X,L2), L1: JMP(L2) */
|
|
target = ZEND_OP1_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == opline->opcode-3 &&
|
|
(SAME_VAR(target->op1, opline->result) ||
|
|
SAME_VAR(target->op1, opline->op1))) {
|
|
/* convert T=JMPZ_EX(X,L1), L1: JMPZ(T,L2) to
|
|
JMPZ_EX(X,L2) */
|
|
target = ZEND_OP2_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == opline->opcode &&
|
|
target->result.var == opline->result.var &&
|
|
(SAME_VAR(target->op1, opline->result) ||
|
|
SAME_VAR(target->op1, opline->op1))) {
|
|
/* convert T=JMPZ_EX(X,L1), L1: T=JMPZ_EX(T,L2) to
|
|
JMPZ_EX(X,L2) */
|
|
target = ZEND_OP2_JMP_ADDR(target);
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == INV_EX_COND(opline->opcode) &&
|
|
(SAME_VAR(target->op1, opline->result) ||
|
|
SAME_VAR(target->op1, opline->op1))) {
|
|
/* convert T=JMPZ_EX(X,L1), L1: JMPNZ(T,L2) to
|
|
JMPZ_EX(X,L1+1) */
|
|
target = target + 1;
|
|
} else if (target->opcode == INV_EX_COND_EX(opline->opcode) &&
|
|
target->result.var == opline->result.var &&
|
|
(SAME_VAR(target->op1, opline->result) ||
|
|
SAME_VAR(target->op1, opline->op1))) {
|
|
/* convert T=JMPZ_EX(X,L1), L1: T=JMPNZ_EX(T,L2) to
|
|
JMPZ_EX(X,L1+1) */
|
|
target = target + 1;
|
|
} else if (target->opcode == ZEND_BOOL &&
|
|
(SAME_VAR(target->op1, opline->result) ||
|
|
SAME_VAR(target->op1, opline->op1))) {
|
|
/* convert Y = JMPZ_EX(X,L1), L1: Z = BOOL(Y) to
|
|
Z = JMPZ_EX(X,L1+1) */
|
|
|
|
/* NOTE: This optimization pattern is not safe, but works, */
|
|
/* because result of JMPZ_EX instruction */
|
|
/* is not used on the following path and */
|
|
/* should be used once on the branch path. */
|
|
/* */
|
|
/* The pattern works well only if jumps processed in */
|
|
/* direct order, otherwise it breaks JMPZ_EX */
|
|
/* sequences too early. */
|
|
opline->result.var = target->result.var;
|
|
target = target + 1;
|
|
CHECK_LOOP(target);
|
|
} else if (target->opcode == ZEND_NOP) {
|
|
target = target + 1;
|
|
} else {
|
|
break;
|
|
}
|
|
ZEND_SET_OP_JMP_ADDR(opline, opline->op2, target);
|
|
}
|
|
|
|
/* convert L: T = JMPZ_EX X,L+1 to T = BOOL(X) */
|
|
if (target == opline + 1) {
|
|
opline->opcode = ZEND_BOOL;
|
|
opline->op2.num = 0;
|
|
}
|
|
break;
|
|
}
|
|
opline++;
|
|
}
|
|
free_alloca(jmp_hitlist, use_heap);
|
|
}
|