mirror of
https://github.com/ruby/ruby.git
synced 2025-09-15 16:44:01 +02:00

Our current implementation of rb_postponed_job_register suffers from some safety issues that can lead to interpreter crashes (see bug #1991). Essentially, the issue is that jobs can be called with the wrong arguments. We made two attempts to fix this whilst keeping the promised semantics, but: * The first one involved masking/unmasking when flushing jobs, which was believed to be too expensive * The second one involved a lock-free, multi-producer, single-consumer ringbuffer, which was too complex The critical insight behind this third solution is that essentially the only user of these APIs are a) internal, or b) profiling gems. For a), none of the usages actually require variable data; they will work just fine with the preregistration interface. For b), generally profiling gems only call a single callback with a single piece of data (which is actually usually just zero) for the life of the program. The ringbuffer is complex because it needs to support multi-word inserts of job & data (which can't be atomic); but nobody actually even needs that functionality, really. So, this comit: * Introduces a pre-registration API for jobs, with a GVL-requiring rb_postponed_job_prereigster, which returns a handle which can be used with an async-signal-safe rb_postponed_job_trigger. * Deprecates rb_postponed_job_register (and re-implements it on top of the preregister function for compatability) * Moves all the internal usages of postponed job register pre-registration
484 lines
14 KiB
C
484 lines
14 KiB
C
/**********************************************************************
|
|
|
|
rjit.c - Ruby JIT compiler functions
|
|
|
|
Copyright (C) 2023 Takashi Kokubun <k0kubun@ruby-lang.org>.
|
|
|
|
**********************************************************************/
|
|
|
|
#include "rjit.h" // defines USE_RJIT
|
|
|
|
#if USE_RJIT
|
|
|
|
#include "constant.h"
|
|
#include "id_table.h"
|
|
#include "internal.h"
|
|
#include "internal/class.h"
|
|
#include "internal/cmdlineopt.h"
|
|
#include "internal/cont.h"
|
|
#include "internal/file.h"
|
|
#include "internal/hash.h"
|
|
#include "internal/process.h"
|
|
#include "internal/warnings.h"
|
|
#include "vm_sync.h"
|
|
#include "ractor_core.h"
|
|
|
|
#ifdef __sun
|
|
#define __EXTENSIONS__ 1
|
|
#endif
|
|
|
|
#include "vm_core.h"
|
|
#include "vm_callinfo.h"
|
|
#include "rjit_c.h"
|
|
#include "ruby_assert.h"
|
|
#include "ruby/debug.h"
|
|
#include "ruby/thread.h"
|
|
#include "ruby/version.h"
|
|
#include "builtin.h"
|
|
#include "insns.inc"
|
|
#include "insns_info.inc"
|
|
#include "internal/compile.h"
|
|
#include "internal/gc.h"
|
|
|
|
#include <sys/wait.h>
|
|
#include <sys/time.h>
|
|
#include <dlfcn.h>
|
|
#include <errno.h>
|
|
#ifdef HAVE_FCNTL_H
|
|
#include <fcntl.h>
|
|
#endif
|
|
#ifdef HAVE_SYS_PARAM_H
|
|
# include <sys/param.h>
|
|
#endif
|
|
#include "dln.h"
|
|
|
|
// For mmapp(), sysconf()
|
|
#ifndef _WIN32
|
|
#include <unistd.h>
|
|
#include <sys/mman.h>
|
|
#endif
|
|
|
|
#include "ruby/util.h"
|
|
|
|
// A copy of RJIT portion of MRI options since RJIT initialization. We
|
|
// need them as RJIT threads still can work when the most MRI data were
|
|
// freed.
|
|
struct rb_rjit_options rb_rjit_opts;
|
|
|
|
// true if RJIT is enabled.
|
|
bool rb_rjit_enabled = false;
|
|
// true if --rjit-stats (used before rb_rjit_opts is set)
|
|
bool rb_rjit_stats_enabled = false;
|
|
// true if --rjit-trace-exits (used before rb_rjit_opts is set)
|
|
bool rb_rjit_trace_exits_enabled = false;
|
|
// true if JIT-ed code should be called. When `ruby_vm_event_enabled_global_flags & ISEQ_TRACE_EVENTS`
|
|
// and `rb_rjit_call_p == false`, any JIT-ed code execution is cancelled as soon as possible.
|
|
bool rb_rjit_call_p = false;
|
|
// A flag to communicate that rb_rjit_call_p should be disabled while it's temporarily false.
|
|
static bool rjit_cancel_p = false;
|
|
|
|
// `rb_ec_ractor_hooks(ec)->events` is moved to this variable during compilation.
|
|
rb_event_flag_t rb_rjit_global_events = 0;
|
|
|
|
// Basically rb_rjit_opts.stats, but this becomes false during RJIT compilation.
|
|
static bool rjit_stats_p = false;
|
|
|
|
// RubyVM::RJIT
|
|
static VALUE rb_mRJIT = 0;
|
|
// RubyVM::RJIT::C
|
|
static VALUE rb_mRJITC = 0;
|
|
// RubyVM::RJIT::Compiler
|
|
static VALUE rb_RJITCompiler = 0;
|
|
// RubyVM::RJIT::CPointer::Struct_rb_iseq_t
|
|
static VALUE rb_cRJITIseqPtr = 0;
|
|
// RubyVM::RJIT::CPointer::Struct_rb_control_frame_t
|
|
static VALUE rb_cRJITCfpPtr = 0;
|
|
// RubyVM::RJIT::Hooks
|
|
static VALUE rb_mRJITHooks = 0;
|
|
|
|
// Frames for --rjit-trace-exits
|
|
VALUE rb_rjit_raw_samples = 0;
|
|
// Line numbers for --rjit-trace-exits
|
|
VALUE rb_rjit_line_samples = 0;
|
|
|
|
// Postponed job handle for triggering rjit_iseq_update_references
|
|
static rb_postponed_job_handle_t rjit_iseq_update_references_pjob;
|
|
|
|
// A default threshold used to add iseq to JIT.
|
|
#define DEFAULT_CALL_THRESHOLD 10
|
|
// Size of executable memory block in MiB.
|
|
#define DEFAULT_EXEC_MEM_SIZE 64
|
|
|
|
#define opt_match_noarg(s, l, name) \
|
|
opt_match(s, l, name) && (*(s) ? (rb_warn("argument to --rjit-" name " is ignored"), 1) : 1)
|
|
#define opt_match_arg(s, l, name) \
|
|
opt_match(s, l, name) && (*(s) ? 1 : (rb_raise(rb_eRuntimeError, "--rjit-" name " needs an argument"), 0))
|
|
|
|
void
|
|
rb_rjit_setup_options(const char *s, struct rb_rjit_options *rjit_opt)
|
|
{
|
|
const size_t l = strlen(s);
|
|
if (l == 0) {
|
|
return;
|
|
}
|
|
else if (opt_match_arg(s, l, "call-threshold")) {
|
|
rjit_opt->call_threshold = atoi(s + 1);
|
|
}
|
|
else if (opt_match_arg(s, l, "exec-mem-size")) {
|
|
rjit_opt->exec_mem_size = atoi(s + 1);
|
|
}
|
|
else if (opt_match_noarg(s, l, "stats")) {
|
|
rjit_opt->stats = true;
|
|
}
|
|
else if (opt_match_noarg(s, l, "trace-exits")) {
|
|
rjit_opt->trace_exits = true;
|
|
}
|
|
else if (opt_match_noarg(s, l, "dump-disasm")) {
|
|
rjit_opt->dump_disasm = true;
|
|
}
|
|
else if (opt_match_noarg(s, l, "verify-ctx")) {
|
|
rjit_opt->verify_ctx = true;
|
|
}
|
|
// --rjit=pause is an undocumented feature for experiments
|
|
else if (opt_match_noarg(s, l, "pause")) {
|
|
rjit_opt->pause = true;
|
|
}
|
|
else {
|
|
rb_raise(rb_eRuntimeError,
|
|
"invalid RJIT option `%s' (--help will show valid RJIT options)", s);
|
|
}
|
|
}
|
|
|
|
#define M(shortopt, longopt, desc) RUBY_OPT_MESSAGE(shortopt, longopt, desc)
|
|
const struct ruby_opt_message rb_rjit_option_messages[] = {
|
|
M("--rjit-stats", "", "Enable collecting RJIT statistics"),
|
|
#if RJIT_STATS
|
|
M("--rjit-trace-exits", "", "Trace side exit locations"),
|
|
#endif
|
|
M("--rjit-exec-mem-size=num", "", "Size of executable memory block in MiB (default: " STRINGIZE(DEFAULT_EXEC_MEM_SIZE) ")"),
|
|
M("--rjit-call-threshold=num", "", "Number of calls to trigger JIT (default: " STRINGIZE(DEFAULT_CALL_THRESHOLD) ")"),
|
|
#ifdef HAVE_LIBCAPSTONE
|
|
M("--rjit-dump-disasm", "", "Dump all JIT code"),
|
|
#endif
|
|
{0}
|
|
};
|
|
#undef M
|
|
|
|
struct rb_rjit_runtime_counters rb_rjit_counters = { 0 };
|
|
|
|
#if RJIT_STATS
|
|
void
|
|
rb_rjit_collect_vm_usage_insn(int insn)
|
|
{
|
|
if (!rjit_stats_p) return;
|
|
rb_rjit_counters.vm_insns_count++;
|
|
}
|
|
#endif // YJIT_STATS
|
|
|
|
extern VALUE rb_gc_enable(void);
|
|
extern VALUE rb_gc_disable(void);
|
|
|
|
#define WITH_RJIT_ISOLATED(stmt) do { \
|
|
VALUE was_disabled = rb_gc_disable(); \
|
|
rb_hook_list_t *global_hooks = rb_ec_ractor_hooks(GET_EC()); \
|
|
rb_rjit_global_events = global_hooks->events; \
|
|
global_hooks->events = 0; \
|
|
bool original_call_p = rb_rjit_call_p; \
|
|
rjit_stats_p = false; \
|
|
rb_rjit_call_p = false; \
|
|
stmt; \
|
|
rb_rjit_call_p = (rjit_cancel_p ? false : original_call_p); \
|
|
rjit_stats_p = rb_rjit_opts.stats; \
|
|
global_hooks->events = rb_rjit_global_events; \
|
|
if (!was_disabled) rb_gc_enable(); \
|
|
} while (0);
|
|
|
|
void
|
|
rb_rjit_cancel_all(const char *reason)
|
|
{
|
|
if (!rb_rjit_enabled)
|
|
return;
|
|
|
|
rb_rjit_call_p = false;
|
|
rjit_cancel_p = true;
|
|
}
|
|
|
|
void
|
|
rb_rjit_bop_redefined(int redefined_flag, enum ruby_basic_operators bop)
|
|
{
|
|
if (!rb_rjit_call_p) return;
|
|
rb_rjit_call_p = false;
|
|
}
|
|
|
|
static void
|
|
rjit_cme_invalidate(void *data)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
WITH_RJIT_ISOLATED({
|
|
rb_funcall(rb_mRJITHooks, rb_intern("on_cme_invalidate"), 1, SIZET2NUM((size_t)data));
|
|
});
|
|
}
|
|
|
|
extern int rb_workqueue_register(unsigned flags, rb_postponed_job_func_t func, void *data);
|
|
|
|
void
|
|
rb_rjit_cme_invalidate(rb_callable_method_entry_t *cme)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
// Asynchronously hook the Ruby code since running Ruby in the middle of cme invalidation is dangerous.
|
|
rb_workqueue_register(0, rjit_cme_invalidate, (void *)cme);
|
|
}
|
|
|
|
void
|
|
rb_rjit_before_ractor_spawn(void)
|
|
{
|
|
if (!rb_rjit_call_p) return;
|
|
rb_rjit_call_p = false;
|
|
}
|
|
|
|
static void
|
|
rjit_constant_state_changed(void *data)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
RB_VM_LOCK_ENTER();
|
|
rb_vm_barrier();
|
|
|
|
WITH_RJIT_ISOLATED({
|
|
rb_funcall(rb_mRJITHooks, rb_intern("on_constant_state_changed"), 1, SIZET2NUM((size_t)data));
|
|
});
|
|
|
|
RB_VM_LOCK_LEAVE();
|
|
}
|
|
|
|
void
|
|
rb_rjit_constant_state_changed(ID id)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
// Asynchronously hook the Ruby code since this is hooked during a "Ruby critical section".
|
|
rb_workqueue_register(0, rjit_constant_state_changed, (void *)id);
|
|
}
|
|
|
|
void
|
|
rb_rjit_constant_ic_update(const rb_iseq_t *const iseq, IC ic, unsigned insn_idx)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
|
|
RB_VM_LOCK_ENTER();
|
|
rb_vm_barrier();
|
|
|
|
WITH_RJIT_ISOLATED({
|
|
rb_funcall(rb_mRJITHooks, rb_intern("on_constant_ic_update"), 3,
|
|
SIZET2NUM((size_t)iseq), SIZET2NUM((size_t)ic), UINT2NUM(insn_idx));
|
|
});
|
|
|
|
RB_VM_LOCK_LEAVE();
|
|
}
|
|
|
|
void
|
|
rb_rjit_tracing_invalidate_all(rb_event_flag_t new_iseq_events)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
WITH_RJIT_ISOLATED({
|
|
rb_funcall(rb_mRJITHooks, rb_intern("on_tracing_invalidate_all"), 1, UINT2NUM(new_iseq_events));
|
|
});
|
|
}
|
|
|
|
static void
|
|
rjit_iseq_update_references(void *data)
|
|
{
|
|
if (!rb_rjit_enabled || !rb_rjit_call_p || !rb_mRJITHooks) return;
|
|
WITH_RJIT_ISOLATED({
|
|
rb_funcall(rb_mRJITHooks, rb_intern("on_update_references"), 0);
|
|
});
|
|
}
|
|
|
|
void
|
|
rb_rjit_iseq_update_references(struct rb_iseq_constant_body *const body)
|
|
{
|
|
if (!rb_rjit_enabled) return;
|
|
|
|
if (body->rjit_blocks) {
|
|
body->rjit_blocks = rb_gc_location(body->rjit_blocks);
|
|
}
|
|
|
|
// Asynchronously hook the Ruby code to avoid allocation during GC.compact.
|
|
// Using _one because it's too slow to invalidate all for each ISEQ. Thus
|
|
// not giving an ISEQ pointer.
|
|
rb_postponed_job_trigger(rjit_iseq_update_references_pjob);
|
|
}
|
|
|
|
void
|
|
rb_rjit_iseq_mark(VALUE rjit_blocks)
|
|
{
|
|
if (!rb_rjit_enabled) return;
|
|
|
|
// Note: This wasn't enough for some reason.
|
|
// We actually rely on RubyVM::RJIT::GC_REFS to mark this.
|
|
if (rjit_blocks) {
|
|
rb_gc_mark_movable(rjit_blocks);
|
|
}
|
|
}
|
|
|
|
// Called by rb_vm_mark()
|
|
void
|
|
rb_rjit_mark(void)
|
|
{
|
|
if (!rb_rjit_enabled)
|
|
return;
|
|
RUBY_MARK_ENTER("rjit");
|
|
|
|
// Pin object pointers used in this file
|
|
rb_gc_mark(rb_RJITCompiler);
|
|
rb_gc_mark(rb_cRJITIseqPtr);
|
|
rb_gc_mark(rb_cRJITCfpPtr);
|
|
rb_gc_mark(rb_mRJITHooks);
|
|
rb_gc_mark(rb_rjit_raw_samples);
|
|
rb_gc_mark(rb_rjit_line_samples);
|
|
|
|
RUBY_MARK_LEAVE("rjit");
|
|
}
|
|
|
|
void
|
|
rb_rjit_free_iseq(const rb_iseq_t *iseq)
|
|
{
|
|
// TODO: implement this. GC_REFS should remove this iseq's mjit_blocks
|
|
}
|
|
|
|
// TODO: Use this in more places
|
|
VALUE
|
|
rb_rjit_iseq_new(rb_iseq_t *iseq)
|
|
{
|
|
return rb_funcall(rb_cRJITIseqPtr, rb_intern("new"), 1, SIZET2NUM((size_t)iseq));
|
|
}
|
|
|
|
void
|
|
rb_rjit_compile(const rb_iseq_t *iseq)
|
|
{
|
|
RB_VM_LOCK_ENTER();
|
|
rb_vm_barrier();
|
|
|
|
WITH_RJIT_ISOLATED({
|
|
VALUE iseq_ptr = rb_funcall(rb_cRJITIseqPtr, rb_intern("new"), 1, SIZET2NUM((size_t)iseq));
|
|
VALUE cfp_ptr = rb_funcall(rb_cRJITCfpPtr, rb_intern("new"), 1, SIZET2NUM((size_t)GET_EC()->cfp));
|
|
rb_funcall(rb_RJITCompiler, rb_intern("compile"), 2, iseq_ptr, cfp_ptr);
|
|
});
|
|
|
|
RB_VM_LOCK_LEAVE();
|
|
}
|
|
|
|
void *
|
|
rb_rjit_entry_stub_hit(VALUE branch_stub)
|
|
{
|
|
VALUE result;
|
|
|
|
RB_VM_LOCK_ENTER();
|
|
rb_vm_barrier();
|
|
|
|
rb_control_frame_t *cfp = GET_EC()->cfp;
|
|
|
|
WITH_RJIT_ISOLATED({
|
|
VALUE cfp_ptr = rb_funcall(rb_cRJITCfpPtr, rb_intern("new"), 1, SIZET2NUM((size_t)cfp));
|
|
result = rb_funcall(rb_RJITCompiler, rb_intern("entry_stub_hit"), 2, branch_stub, cfp_ptr);
|
|
});
|
|
|
|
RB_VM_LOCK_LEAVE();
|
|
|
|
return (void *)NUM2SIZET(result);
|
|
}
|
|
|
|
void *
|
|
rb_rjit_branch_stub_hit(VALUE branch_stub, int sp_offset, int target0_p)
|
|
{
|
|
VALUE result;
|
|
|
|
RB_VM_LOCK_ENTER();
|
|
rb_vm_barrier();
|
|
|
|
rb_control_frame_t *cfp = GET_EC()->cfp;
|
|
cfp->sp += sp_offset; // preserve stack values, also using the actual sp_offset to make jit.peek_at_stack work
|
|
|
|
WITH_RJIT_ISOLATED({
|
|
VALUE cfp_ptr = rb_funcall(rb_cRJITCfpPtr, rb_intern("new"), 1, SIZET2NUM((size_t)cfp));
|
|
result = rb_funcall(rb_RJITCompiler, rb_intern("branch_stub_hit"), 3, branch_stub, cfp_ptr, RBOOL(target0_p));
|
|
});
|
|
|
|
cfp->sp -= sp_offset; // reset for consistency with the code without the stub
|
|
|
|
RB_VM_LOCK_LEAVE();
|
|
|
|
return (void *)NUM2SIZET(result);
|
|
}
|
|
|
|
void
|
|
rb_rjit_init(const struct rb_rjit_options *opts)
|
|
{
|
|
VM_ASSERT(rb_rjit_enabled);
|
|
|
|
// Normalize options
|
|
rb_rjit_opts = *opts;
|
|
if (rb_rjit_opts.exec_mem_size == 0)
|
|
rb_rjit_opts.exec_mem_size = DEFAULT_EXEC_MEM_SIZE;
|
|
if (rb_rjit_opts.call_threshold == 0)
|
|
rb_rjit_opts.call_threshold = DEFAULT_CALL_THRESHOLD;
|
|
#ifndef HAVE_LIBCAPSTONE
|
|
if (rb_rjit_opts.dump_disasm)
|
|
rb_warn("libcapstone has not been linked. Ignoring --rjit-dump-disasm.");
|
|
#endif
|
|
|
|
// RJIT doesn't support miniruby, but it might reach here by RJIT_FORCE_ENABLE.
|
|
rb_mRJIT = rb_const_get(rb_cRubyVM, rb_intern("RJIT"));
|
|
if (!rb_const_defined(rb_mRJIT, rb_intern("Compiler"))) {
|
|
rb_warn("Disabling RJIT because RubyVM::RJIT::Compiler is not defined");
|
|
rb_rjit_enabled = false;
|
|
return;
|
|
}
|
|
rjit_iseq_update_references_pjob = rb_postponed_job_preregister(rjit_iseq_update_references, NULL);
|
|
if (rjit_iseq_update_references_pjob == POSTPONED_JOB_HANDLE_INVALID) {
|
|
rb_bug("Could not preregister postponed job for RJIT");
|
|
}
|
|
rb_mRJITC = rb_const_get(rb_mRJIT, rb_intern("C"));
|
|
VALUE rb_cRJITCompiler = rb_const_get(rb_mRJIT, rb_intern("Compiler"));
|
|
rb_RJITCompiler = rb_funcall(rb_cRJITCompiler, rb_intern("new"), 0);
|
|
rb_cRJITIseqPtr = rb_funcall(rb_mRJITC, rb_intern("rb_iseq_t"), 0);
|
|
rb_cRJITCfpPtr = rb_funcall(rb_mRJITC, rb_intern("rb_control_frame_t"), 0);
|
|
rb_mRJITHooks = rb_const_get(rb_mRJIT, rb_intern("Hooks"));
|
|
if (rb_rjit_opts.trace_exits) {
|
|
rb_rjit_raw_samples = rb_ary_new();
|
|
rb_rjit_line_samples = rb_ary_new();
|
|
}
|
|
|
|
// Enable RJIT and stats from here
|
|
rb_rjit_call_p = !rb_rjit_opts.pause;
|
|
rjit_stats_p = rb_rjit_opts.stats;
|
|
}
|
|
|
|
//
|
|
// Primitive for rjit.rb
|
|
//
|
|
|
|
// Same as `rb_rjit_opts.stats`, but this is used before rb_rjit_opts is set.
|
|
static VALUE
|
|
rjit_stats_enabled_p(rb_execution_context_t *ec, VALUE self)
|
|
{
|
|
return RBOOL(rb_rjit_stats_enabled);
|
|
}
|
|
|
|
// Same as `rb_rjit_opts.trace_exits`, but this is used before rb_rjit_opts is set.
|
|
static VALUE
|
|
rjit_trace_exits_enabled_p(rb_execution_context_t *ec, VALUE self)
|
|
{
|
|
return RBOOL(rb_rjit_trace_exits_enabled);
|
|
}
|
|
|
|
// Disable anything that could impact stats. It ends up disabling JIT calls as well.
|
|
static VALUE
|
|
rjit_stop_stats(rb_execution_context_t *ec, VALUE self)
|
|
{
|
|
rb_rjit_call_p = false;
|
|
rjit_stats_p = false;
|
|
return Qnil;
|
|
}
|
|
|
|
#include "rjit.rbinc"
|
|
|
|
#endif // USE_RJIT
|