Skip to content
/ bcgen Public

Bytecode generator in the vein of copy-and-patch

Notifications You must be signed in to change notification settings

Kimplul/bcgen

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Bytecode generator

A relatively simple script that generates pretty fast bytecode systems. Relies on some non-standard C, but should in theory be completely portable.

Usage

bcgen <rules.py>

The script is pretty bare bones at the moment. Essentially, it generates a number of files in a directory called gen in the working directory. These generated files are included by bcode.c, which you can compile into an object file that provides an interface to compiling and running your bytecode system.

rules.py is a file that describes which operations to generate. Here's an example rules.py:

g = BCGen(r = 4, f = 0)
g.rule('addi', 'rRR__I', 'R0 = R1 + IMM;')
g.write()

BCGen() is the generator class. It has a number of optional parameters:

  • r for number of general purpose registers, default to 3.

  • f for number of floating point registers, default to 3.

  • i for number of immediates an operation can take, default to 1.

  • ubcval_type for type to use for general purpose registers, default to unsigned long

  • sbcval_type for signed equivalent of ubcval_type, default to signed long.

  • head for user configurable code to be inserted at the head of the genrated bytecode, useful for defining helper functions etc.

  • extra for user configurable code to be inserted right before bytecode interpreting. Useful for defining variables to store temporary values to etc.

rule() is the workhorse here. It takes three parameters:

  • name for the instruction name.
  • fmt for the instruction format.
  • body for the instruction body.

The fmt parameter is a string of at least five characters, from left to right:

  1. Relocation slot, r or _

r means that the when compiling the instruction, it should return a relocation that can later be patched with a value. _ means that no relocation is necessary.

  1. Four register slots, R, F or _:

R refers to a general purpose register, and F refers to a floating point register. _ means the register slot is unused. When compiling an operation, the user can pass in a number for each register slot that is the index of the register to use. Each operation can take a maximum of four register arguments.

  1. Immediate slots, I, D or _, as many as the i parameter to BCGen():

I refers to an integer immediate value, D refers to a floating point immediate value. _ means the slot is unused. The user can pass in immediate values that are accessible to the operation at runtime via IMM(i)/SIMM(i)/FIMM(i)/DIMM(i), for unsigned integer, signed integer, float and double respectively. i specifies which immedate slot is used. Immediate slot 0 has special shorthands IMM0/SIMM0/FIMM0/DIMM0 to reduce typing.

Body

The body of the instruction is what gets executed. The register arguments are referenced via R + slot index for the general purpose registers and F + slot index for the floating point registers.

The index to use when specifying register arguments is the register slot index, so for exampl the format _R_FR_ makes registers R0, F2 and R3 available to the body. It's very important to note that R0 does not refer to the literal first register in the bytecode system, rather it's the first register argument the user gives when compiling the instruction.

Each instruction can be compiled via the generated procedure select_name(). For the example rules.py above, to compile the instruction addi you would call select_addi(0, 2, 500);. The first and second arguments are the register indexes to use, and map to R0 and R1. The immediate is always last, and in this case it's 500. One way to look at the situation is that we're requesting that a bytecode instruction be compiled where register 2 plus 500 be placed into register 0.

Note that it can be easy to mix up which parameters are for register indexes and which are for immediate values. A possible future improvement is wrapping register indexes in their own struct, similar to lightening.

Jumps

By default the bytecode instruction following the current one is executed automatically. You can however specify explicit jumps, such as in branches, by JUMP(target). The target can be patched in later or be specified as an argument. For example:

# rules.py
...
g.rule('beq', 'rRR__I', 'if (R0 == R1) JUMP(IMM);')
// file.c
bcval_t label = label();
...
select_beq(0, 1, label.g); // we already know where we want to jump

// OR
bcreloc_t reloc = select_beq(0, 1, 0); // zero is placeholder
...
bcval_t label = label();
patch(reloc, 0, label); // now we know where we want to jump

The patch() procedure takes a relocation to patch, which immediate slot we want to patch and the value to place in the relocation at the immediate slot. (Note that the patch system probably should/could be improved, but good enough for now.)

See example for a simple test program that generates a limited set of operations that are enough to sum the first billion integers.

Performance

As a very basic test, below is a comparison of example/exec found in this repo in relation to some other methods to accomplish summing the first billion integers.

'Method' Time (s)
C, -O3 -march=native 0.062
C, -O2 0.221
LuaJIT 0.656
C, -O0 0.709
example/exec 1.533
Lua 18.717
Perl 27.962

The full C program is

#include <stdio.h>
#include <assert.h>
#include <stdlib.h>

int main(int argc, char *argv[])
{
        assert(argc == 2);
        unsigned long sum = 0;
        unsigned long n = strtoull(argv[1], 0, 0);
        for (unsigned long i = 0; i < n; ++i)
                sum += i;

        printf("%llu\n", sum);
        return 0;
}

Note the user input so as to avoid completely optimizing away the loop.

Notes

This software is currently very crude and pretty much only suitable for a simple demonstration. However, I think the performance figures are pretty impressive for what it is. A JIT system will essentially always beat this system in speed, but this is trivially portable and the code generation is ligtning fast. See comparisons in https://github.com/Kimplul/jit-comparison.

A fairly close match in the JIT world is lightning (and its lighter fork, lightening). My intention is to try and use this tool to generate a bytecode backend for lightening as a fallback for platforms that lightening doesn't yet support to create a 'universal' low level virtual machine.

No restrictions are placed on register count or instruction count. The instruction count increases the size and compile time linearly, but register count affects the size and compile time exponentially so be careful. Also, to maintain the highest performance, the register count should be smaller than the register count of the underlying hardware. This allows the compiler to lower bytecode registers to hardware registers, speeding things up considerably.

Finally, note that the system doesn't assign any specific usage conventions to registers. You are responsible for maintaining an ABI of some kind, with stack/frame register(s), callee-save vs. caller-save, etc.

Slow compilation times

With massive systems, some optimizations passes have essentially no effect but take up an immense amount of time.

GCC users should compile bcode.c with -fno-tree-fre -fno-gcse -fno-expensive-optimizations. This seems to have minimal impact on runtime performance, but has massive cost savings on compilation time. -fno-gcse is recommended by the GCC manual for computed gotos, though I didn't see any improvement in the generated code on x86. Maybe other architectures benefit from it more? Dunno. -fno-tree-slp-vectorize can also be useful to decrease memory usage, but doesn't seem to have too much of an effect on the compilation speed.

LLVM users should use -fno-slp-vectorize which does help a bit, but -ftime-report doesn't show any obvious other flags that should be turned off. I'm sure there are, I just haven't looked hard enough. In any case, GCC I suppose is the currently recommended compiler.

Using tailcalls instead of a single massive function

I've seen some things online about using tailcalls to produce faster bytecode interpreters, so as an experiment, I modified the generator to output tailcalls instead, where each operation corresponds to a single function that uses tailcalls to the next operation. The modifications can be found in the tailcalls branch.

In practice, it seems that this method has a slightly higher runtime overhead due to the fact that calling conventions typically allow only a fairly small number of physical registers to map the virtual registers to, leading to more stack accesses than with a single huge function. Additionally, some architecture don't seem to support tailcalls, or at least not efficient ones, so tailcalls may be less portable than this method. Would be interesting to see if the situation changes with the musttail attribute possibly being standardized.

Seemingly main advantage of tailcalls seems to be that the library compiles quite a bit quicker, as compiler are typically more geared towards handling lots of small functions over single massive ones.

Although, please keep in mind that this bytecode generator is only interested in producing machine-like bytecode, and wouldn't necessarily be as well suited for a higher-level bytecode that languages like Python uses, where tailcalls may well be a better fit.

About

Bytecode generator in the vein of copy-and-patch

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published