For various reasons the Tier 2 optimizer and interpreter prefer to have an instruction format that only contains one argument per instruction -- the argument can be either the opcode or a cache value.
This is a substantial task, but luckily we can parallelize it.
I expect that we'll find we need some improvements to the generator, to ensure that the original bytecodes (which become macro-instructions) are as efficient as their original version.
For various reasons the Tier 2 optimizer and interpreter prefer to have an instruction format that only contains one argument per instruction -- the argument can be either the opcode or a cache value.
This is a substantial task, but luckily we can parallelize it.
I expect that we'll find we need some improvements to the generator, to ensure that the original bytecodes (which become macro-instructions) are as efficient as their original version.