summaryrefslogtreecommitdiffstats
path: root/share
Commit message (Collapse)AuthorAgeFilesLines
...
* list-builder: methods return object.Kaz Kylheku2021-02-261-7/+8
| | | | | | | | | | | | The list-builder methods, other than del, del* and get, now return the object instead of nil. * share/txr/stdlib/build.tl (list-builder (add, add*, pend, pend*, ncon, ncon*): Return the object, self. (list-builder-flets): Do not return the object out of the local functions which invoke the above methods. * txr.1: Documented.
* compiler: new optimization.Kaz Kylheku2021-02-261-25/+54
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Using liveness information, if we are very careful about the circumstances, we can can eliminate instructions of the form mov tN src and replace every subsequent occurrence of tN in the basic block by src. For instance, simple case: if a function ends with mov t13 d5 end t13 that can be rewriten as end d5 The most important condition is that t13 is not live on exit from that basic block. There are other conditions. For now, one of the conditions is that src cannot be a v register. * share/txr/stdlib/optimize.tl (struct live-info): New slot, def. This indicates which t register is being clobbered, if any, by the instruction to which this info is attached. (basic-blocks local-liveness): Adjust the propagation of the defined info. If an instruction both consumes a register and overwrites it, we track that as both a use and a definition. We set up the def fields of live-info. We do that by mutation, so we must be careful to copy the structure. The def field pertains to just one instruction, but the same info can be attached to multiple instructions. (subst-preserve): New function. (basic-blocks peephole-block): New optimization added. Now takes a basic-block argument, bl. (basic-blocks peephole): Pass bl to peephole-block.
* compiler: data flow analysis for t registers.Kaz Kylheku2021-02-242-22/+197
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | The optimizer now calculates t liveness information for the t registers. In every basic block, it now knows which t regs are live on exit, and which are used in the block, at every instruction. One small optimization is based on this so far: the removal of a move instruction targeting a dead register. This appears stable. * share/txr/stdlib/compiler.tl (compiler comp-unwind-protect): The protected code of a uwprot must terminate with a regular end instruction, rather than the jend pseudo-instruction. This is because the clean-up block is executed after the protected block and references values generated in it: t registers are live between he pfrag and the cfrag. Without this, the compile-file-conditionally function was wrongly optimized, causing it to return false due to the setting of the success flag (that having been moved into a t register) having being optimized away. (compiler optimize): Add the call the basic-blocks method to calculate liveness. * share/txr/stdlib/optimize.tl (struct live-info, struct basic-block): New structure types. The basic-block structure type now representes basic blocks instead of raw lists. (struct basic-blocks): New slots, root, li-hash. (basic-blocks jump-ops): We add few instructions that reference labels, just to be safe. (basic-blocks :postinit): Refactor division into basic blocks so that it generates basic-block objects instead of just lists of instructions. Also, the new method link-graph is called which analyzes the tail instructions of all the blocks to determine connectivity and sets the next and links fields of the objects to build a graph. (basic-blocks (get-insns, cut-blocks)): Refactor for struct represenation of basic blocks. (basic-blocks (link-graph, local-liveness, calc-liveness): New methods. (basic-blocks thread-jumps-block): Refactor for struct representation of basic blocks. (basic-blocks peephole-blocks): Likewise, and new pattern for removing moves into dead t-registers, assisted by liveness information. (basic-blocks (peephole, thread-jumps)): Refactor for basic-blocks representation.
* compiler: bugfix: internal error in sys:switch.Kaz Kylheku2021-02-241-1/+1
| | | | | | | | | | | | | | Compiling a form like (caseq op ((a b c d e f g h i j k) 42))) Results in a run-time error in the compiler, similar to: list-vec: (#:l0048) is not of type vec * share/txr/stdlib/compiler.tl (compiler comp-switch): Make sure cases is also still a vector in the complex case when it's not just a copy of cases-vec.
* compiler: wrong close pattern in jump threading.Kaz Kylheku2021-02-231-2/+2
| | | | | | | * share/txr/stdlib/optimize.tl (thread-jumps-block): Add missing argument to close instruction pattern. This causes us to miss a threading opportunity due to the new ntregs parameter being mistaken for a label, which is not found.
* compiler: constant folding: avoid shadowed funs.Kaz Kylheku2021-02-201-7/+7
| | | | | | | | | | | | * share/txr/stdlib/compiler.tl (compiler comp-arith-form): Pass env to reduce-constant. (compiler comp-fun-form): Likewise, and don't bother checking %const-foldable% because reduce-constant does that again. (compiler comp-apply-call): Pass env to reduce-constant. (reduce-constant): Take env argument. If the function is constant foldable, check that there is no lexical function call binding shadowing it. If so, it's not the function we think it is, and we must not constant-fold it.
* compiler: constant-fold [...] forms.Kaz Kylheku2021-02-201-23/+28
| | | | | | | | | | * share/txr/stdlib/compiler.tl (compiler comp-apply-call): Constant-fold the arguments. Check for special cases involving call and route to regular function call. (compiler comp-dwim): Don't wrap all arguments with sys:lisp1-value, only those that are bindable symbols. This way constant expressions, including keywords, t and nil, are not wrapped, and detectable by constantp.
* compiler: constant-fold more basic functions.Kaz Kylheku2021-02-201-1/+17
| | | | | | | | * share/txr/stdlib/compiler.tl (%const-foldable-funs%): Add numerous eligible functions that are registered in eval.c. We avoid anything with functional arguments, environmental dependencies or anything that may be relied upon to produce a fresh object.
* compiler: constant-fold all cadr cons accessors.Kaz Kylheku2021-02-191-1/+7
| | | | | | | * share/txr/stdlib/compiler.tl (%const-foldable-funs%): Add all of the cadr, caddr, and other functions. Take out first and second; these will be later added together with other things that are being registered in eval.c.
* compiler: constant-fold most arithmetic functionsKaz Kylheku2021-02-191-6/+15
| | | | | | | | * share/txr/stdlib/compiler.tl (%const-foldable-funs%): Add most functions from arith module. (%const-foldable%): New variable, hash built from list. (compiler comp-fun-form, reduce-constant): Refer to %const-foldable% hash instead of %const-foldable-funs% list.
* compiler: constant fold pred/succ functions.Kaz Kylheku2021-02-193-5/+8
| | | | | | | | | | | | * share/txr/stdlib/compiler.tl (%const-foldable-funs%): Add pred, succ and their sisters. * share/txr/stdlib/vm-param.tl (%max-lev-idx%, %max-v-lev%, %max-sm-lev-idx%): Get rid of macro-time wrapping in calculation, which are there for manual constant folding. * share/txr/stdlib/asm.tl (with-lev-idx): Remove macro-time providing manual constant folding.
* compiler: start of constant-folding implementation.Kaz Kylheku2021-02-181-0/+46
| | | | | | | | | | | | | | | | | Introducing folding of certain expressions that can be evaluated at compile time, with some special handling for common arithmetic functions, in which we can collapse consecutive arguments that are constant integer expressions. * share/txr/stdlib/compiler.tl (%const-foldable-funs%): New global variable. (compiler compile): Send multiplication and division through new methods that that treat integer arguments. (compiler comp-arith-form, compiler comp-neg-arith-form): New methods. (comp-fun-form): Apply constant folding to a proper function call whose operator is listed in %const-foldable-funs%. (reduce-constant): New function.
* compiler: reduce (if (not (eq ...) ...)).Kaz Kylheku2021-02-181-0/+3
| | | | | | | * share/txr/stdlib/compiler.tl (compiler comp-if): Recognize the pattern (if (not (eq ...) ..), and convert to (if (neq ...) ...) and likewise for eql and equal. This is fed back to comp-if, whereby it may be further reduced.
* compiler: use fixed-point macro for reduce-lisp.Kaz Kylheku2021-02-181-27/+32
| | | | | | * share/txr/stdlib/compiler.tl (fixed-point): New macro. (reduce-lisp): Hide irrelevant iteration details by using fixed-point macro.
* compiler: reduce (not (eq ...)) and related exprs.Kaz Kylheku2021-02-181-0/+3
| | | | | * share/txr/stdlib/compiler.tl (compiler comp-fun-form): Reduce negated eq, eql, equal to neq, neql, nequal.
* compiler: strength reduction of nequal.Kaz Kylheku2021-02-181-4/+6
| | | | | * share/txr/stdlib/compiler.tl (compiler comp-if): Support reduction of nequal in the same way as equal.
* compiler: reduce (list* x) to x.Kaz Kylheku2021-02-181-1/+1
| | | | | | | * share/txr/stdlib/compiler.tl (reduce-lisp): Add one more reduction case. There is a "hit" for this somewhere, because even though this adds code, overall 200 bytes are saved over the entire library.
* compiler: add list construction optimizations.Kaz Kylheku2021-02-171-0/+35
| | | | | | | | | | | | The raw size of the library compiled files shrinks by over 2% from this optimization, not to mention that some list construction code is faster. * share/txr/stdlib/compiler.tl (compiler comp-fun-form): Reduce common list construction primitives via reduce-lisp function which algebraically transforms to a form with fewer function calls. (reduce-lisp): New function.
* compiler: condense if with pattern matching.Kaz Kylheku2021-02-171-59/+47
| | | | | | | | | | * share/txr/stdlib/compiler.tl (compiler comp-if): Remove the pointless cases which check for test being nil, since that is subsumed under constantp. Move all the constantp cases up, making them match-case clauses. The handling of %test-funs% in several places becomes a single pattern case. The remaining cases don't have any more sub-cases to test, so the cond forms are gone.
* compiler: strength reduction of equal.Kaz Kylheku2021-02-171-6/+29
| | | | | | | | | | | | | | | Here, we look for (equal x y) expressions that can be reduced to (eql x y) or (eq x y) and compiled that way. Also, we look for (if (equal x y) ...) expressions that can be turned into (if (eql x y) ...) or (if (eq x y) ...) which then compile into ifq or ifql instructions. * share/txr/stdlib/compiler.tl (compiler comp-if): Convert tree-case into match case, and then handle the (if (equal ...)) pattern. (comp-fun-form): Add recognition for (equal x y) expressions, and reduce their strength, if possible. (eq-comparable, eql-comparable): New functions.
* compiler: fix jump-threading regression.Kaz Kylheku2021-02-171-48/+48
| | | | | | | | | | * share/txr/stdlib/optimize.tl (basic-blocks thread-jumps-block): We want a set here, not a pset, otherwise we are processing the old-instruction again rather than iterating. This breaks jump threading where multiple iterations are required to get to the ultimate target. It showed up as a difference in the compiled image of the sys:compile-match function.
* compiler: use pattern matching for function formKaz Kylheku2021-02-171-11/+7
| | | | | * share/txr/stdlib/compiler.tl (compiler comp-fun-form): Rewritten more compactly and extensibly using match-case.
* compiler: separate jump threading from peepholeKaz Kylheku2021-02-172-51/+65
| | | | | | | | | | | | | | | Jump threading just needs to looks at the last instruction in a basic blocks now; it's a waste of cycles to be pattern matching on jump intruction patterns while peephole scanning. * share/txr/stdlib/compiler.tl (compiler optimize): Invoke new thread-jumps after peephole. * share/txr/stdlib/optimize.tl (basic-blocks thread-jumps-block): New method. (basic-blocks peephole-block): Remove jump-threading cases; they are in thread-jumps block. (basic-blocks thread-jumps): New method.
* compiler: remove unreachable optimization.Kaz Kylheku2021-02-161-7/+0
| | | | | | | | * share/txr/stdlib/optimize.tl (basic-blocks peephole-block): Remove the special optimization involving an unconditional jump followed by an if, to a block which tests the same register with another if. This optimization can't match because a jmp and if cannot be in a basic block together.
* compiler: re-scan block altered by frame-move.Kaz Kylheku2021-02-151-0/+1
| | | | | | | | | | | * share/txr/stdlib/optimize.tl (basic-blocks peephole-block): If we move a frame instruction past a jump into the next block, we must add that block's label to the rescan list. There may be an opportunity to propagate the frame instruction deeper into that block. I'm not seeing a difference from this change in the compilation of the standard library, which indicates that this is happening by fluke; the alteration of that block is happening before it has been visited.
* compiler: close instructions terminate basic block.Kaz Kylheku2021-02-151-1/+1
| | | | | | | | * share/txr/stdlib/optimize.tl (struct basic-blocks): Include the close instruction in the set which terminate a basic block. A close is an unconditional jump; execution never continues after a close instruction, but goes unconditionally to a branch target.
* compiler: peephole newly added blocks.Kaz Kylheku2021-02-151-90/+100
| | | | | | | | | | | | | | | | If cut-block is called during peephole optimization, it can introduce blocks that can be missed, in which there might be some opportunity for peephole reduction. Let's keep track of newly added blocks in a re-scan list. * share/txr/stdlib/optimize.tl (struct basic-blocks): New slot, rescan. (basic-blocks cut-block): Add new block's label to rescan list. (basic-blocks peephole-block): New method, formed out of the bulk of basic-blocks peephole. (basic-blocks peephole): After processing the blocks from the hash table, iterate on the rescan list.
* compiler: basic blocks replace extended basic blocks.Kaz Kylheku2021-02-151-12/+27
| | | | | | | | | | | | | | | | * share/txr/stdlib/optimize.tl (struct basic-blocks): jump-ops, new static member. (basic-blocks :postinit): Cut the code into basic blocks rather than extended basic blocks. This means that the insruction which follows every jumping instructions is now a block leader. Every block needs a label, so we add them. (basic-blocks peephole): The optimization which slides a frame instruction past a jump must be refactored to move the frame instruction into the next block. Firstly, moving anything past a jump instruction is no longer allowed, because the result is no longer a basic block. Secondly, doing so prevents further frame movements, because the block no longer has any instructions after the jump over which the frame can be moved.
* asm/vm/compiler: introduce jend pseudo-instruction.Kaz Kylheku2021-02-142-5/+10
| | | | | | | | | | | | | | | | | | | | | | The jend pseudo-instruction is a simple alias for end. It indicates a jumping end: an end that does not fall through to the next instruction but could go somewhere else. This is for "future correctness" as well as clarity. The difference is important in analysis of code into basic blocks. Currently this won't make a difference because all the jend instructions except for the one at the end of compiled top-level form are followed by a label which kicks off a basic block anyway. * share/txr/stdlib/asm.tl (defopcode-alias): New macro. (jend): New opcode, defined as alias for end. * share/txr/stdlib/compiler.tl (comp-unwind-protect, comp-lambda-impl, compile-toplevel): Use jend instruction for a jumping end: the one after the protected code block of a uwprot, the one at the end of a function, and the one at the end of a top-level form.
* vm/asm: housecleaning: remove deprecated opcodes.Kaz Kylheku2021-02-141-46/+1
| | | | | | | | | | | | | | | | | | Since we have are breaking binary compatibility in the upcoming TXR 252, we might as well take the opportunity to remove deprecated opcodes that the compiler doesn't use. * share/txr/stdlib/asm.tl (op-fin): Opcode removed. (op-pprof): Derive directly from op-end rather than op-fin. (op-movrsi, op-movsmi, op-movrbi, op-movi-pseudo): Opcodes removed. * vm.c (vm_fin, vm_movrsi, vm_movsmi, vm_movrbi): Functions removed. (vm_execute): FIN, MOVRSI, MOVSMI, MOVRBI cases removed. * vmop.h: Regenerated. (vm_op_t): Enum members FIN, MOVRSI, MOVSMI, MOVRBI removed.
* compiler: eliminate block from recursive functions.Kaz Kylheku2021-02-121-7/+9
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | The block elimination logic doesn't work for self-recursive functions, even if they invoke no block returning, and use only system functions that don't have anything to do with block returns. This is because the recursive call is not recognized, and treated as a call to an unknown function. Let's put in a simple hack. The defun and defmacro operators will use a new secret special operator called sys:blk instead of block to generate the block. The compilation of sys:blk will assume that (sys:blk name ...) is only used in a defun or defmacro by that same name, and include name in the list of OK functions. So that functions created using the interpreter and then dynamically compiled will also benefit, we add this operator to the interpreter. * eval.c (sys_blk_s): New symbol variable. (op_defun): For defun and defmacro, use sys:blk for the block for the block (eval_init): Initialize sys_blk_s with the interned symbol sys:blk. Register the sys:blk operator. * share/txr/stdlib/compiler.tl (compiler compile): Recognize the sys:blk special form and handle via comp-block. (comp-block): If sys:blk is being compiled, then include the block name in the list of functions that do not perform block returns. (If this is false, other checks will fail before use that.) (expand-defun): Use sys:blk for defun and defmacro.
* compiler/vm: more compact frame size for closures.Kaz Kylheku2021-02-112-121/+143
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Closures do not share t-registers with surrounding code; they do not store a value into such a register that code outside the closure would read and vice versa. When compiling closures, we can can temporarily reset the compiler's t-register allocator machinery to get low t-register values. Then, when executing the closure, we reserve space just for the registers it needs, not based off the containing vm description. Here we make a backwards-incompatible change. The VM close instruction needs an extra parameter indicating the number of t-regisers. This is stored into the closure and used for allocating the frame when it is dispatched. * parser.c (read_file_common): We read nothing but version 6 tlo files now. * share/txr/stdlib/asm.tl (op-close asm): Parse new ntreg argument from close syntax, and put it out as an extra word. Here is where we pay for this improvement in extra code size. (op-close dis): Extract the new argument from the machine code and add it to the disassembled format. * share/txr/stdlib/compiler.tl (compile-in-toplevel): Save and restore the t-reg discards list also. Don't bother with a gensym for the compiler; the argument is always a symbol, which we can use unhygienically like in with-var-spy. (compile-with-fresh-tregs): New macro based on compile-in-toplevel: almost the same but doesn't reset the level. (comp-lambda-impl): Use compile-with-fresh-tregs to compile the entire closure with a minimized register set. Place the treg-cntr into the closure instruction to indicate the number of registers the closure requires. * vm.c (struct vm): New member, nreg. (vm_make_closure): New parameter, nreg, stored into the closure. (vm_close): Extract a third opcode word, and pull the nreg value from the bottom half. Pass this to vm_make_closure. (vm_execute_closure, vm_funcall_common): Calculate frame size based on the closur's nreg rather than the VM description's. * txr.1: Document that the upcoming version 252 produces version 6.0 object files and only loads version 6.
* compiler: frame-eliminating optimization.Kaz Kylheku2021-02-111-149/+260
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This optimization identifies let blocks whose variables are not captured by closures. The variables are relocated to registers and the frame M N ... end reg wrapping is removed. * parser.c (read_file_common): Load version 6 files. We remain backwards-compatible. * share/txr/stdlib/compiler.tl (var-spy, capture-var-spy): New structure types. (struct compiler): New slot, var-spies. (with-var-spy): New macro. (compiler (alloc-new-treg, unalloc-reg-count, push-var-spy, pop-var-spy)): New methods. (compiler (comp-atom, compt-setq, comp-list-setq, comp-lisp1-value)): Inform the spies in the spy notification stack about assignments and accesses. (compiler eliminate-frame): New method. (compiler comp-let): Use spies to determine which variables from this frame are captured, and if none are, then use eliminate-frame to rename all the variables to t-registers and drop the frame setup/teardown. (compiler comp-lambda): Set up a capture-var-spy which intercepts accesses and assignments within a lambda, and informs other spies about the captures. (%tlo-ver%): Bump compiled file version to to (6 0), because of some behavioral changes necessary in the VM. We might revert this if the issues are solved differently. * vm.c (vm_getz): Do not null out T registers. (vm_execute_toplevel, vm_execute_closure): Use zalloca to allocate the register part of the frame, so T registers are initialized to nil.
* compiler: use rewrite-case in dedup-labels.Kaz Kylheku2021-02-101-8/+6
| | | | | | | * share/txr/stdlib/optimize.tl (dedup-labels): Use rewrite-case macro defined in the same file instead of rewrite/lambda/match-case. Also change two-argument list* to cons.
* compiler: bug: trivial unwind-protect return valueKaz Kylheku2021-02-101-1/+1
| | | | | | | | | | | | | Bad test case: (unwind-protect 42 1 2 3) -> 3 ;; should be 42 * share/txr/stdlib/compiler.tl (compile comp-unwind-protect): In the case when the protected code compiles to zero code, because it is a simple variable or constant, the code that we return must still nominate the that fragment's output register as its output, and not the output register of the cleanup forms.
* matcher: diagnose syntax problems in hash pattern.Kaz Kylheku2021-02-091-1/+1
| | | | | | * share/txr/stdlib/match.tl (compile-hash-match): Use mac-param-bind instead of tree-bind, like in the other functions.
* compiler: rename variable in optimizer code.Kaz Kylheku2021-02-091-2/+2
| | | | | | * share/txr/stdlib/optimize.tl (basic-blocks peephole): Rename jlabel3 variable to jlabel2, so it is in sequence after jlabel0 and jlabel1.
* struct: changing meaning of obj.[fun ...] syntax.Kaz Kylheku2021-02-092-6/+16
| | | | | | | | | | | | | | | | | | | | | | | | | Until now, the obj.[fun ...] syntax has uselessly denoted exactly the same thing as [obj.fun ...]. This latter syntax is what should be used for that meaning. The new meaning of obj.[fun ...] will be that it performs method dispatch, where obj is passed to obj.fun as the leftmost argument: obj.[fun ...] is [obj.fun obj ...], with obj evaluated once. * share/txr/stdlib/struct.tl (qref): Expansion change done here, with backward compat switch. * share/txr/stdlib/termios.tl (termios (go-raw, go-cbreak)): Some a.[b c] turned to [a.b c] here. * tests/012/oop.tl (animal print): Likewise. * tests/012/struct.tl: Likewise, and some expansion tests updated to reflect the new expansion. * txr.1: Documentation revised in multiple places and compat note added.
* matcher: replace wrg function with reduce-right.Kaz Kylheku2021-02-091-23/+14
| | | | | | | * share/txr/stdlib/match.tl (wrap-expr): Remove wrg local function. Replace call with simple reduce-right, which doesn't require a reversal of the original list. (compiled-match): Likewise.
* matcher: wrap-guards type case to method dispatch.Kaz Kylheku2021-02-091-44/+48
| | | | | | | | * share/txr/stdlib/match.tl (match-guards wrap-expr): New method. (guard-distinction wrap-expr): New method. (compiled-match wrap-guards): Reduce type-case to wrap-expr method call.
* matcher: match-guard: use op list, and zip.Kaz Kylheku2021-02-091-2/+2
| | | | | | | | * share/txr/stdlib/match.tl (match-guard assignments): Use simpler op expression to generate a function that produces set assignments. (match-guard lets): Use zip function instead of mapcar with ret and quasiquote.
* matcher: get rid of undocumented @(range) op.Kaz Kylheku2021-02-091-5/+4
| | | | | | | | * share/txr/stdlib/match.tl (compile-match): Remove the rcons entry which was supposed to be already gone in version 250, and is no longer documented. (compile-range-match): Edit parameter name to remove misleading reference to rcons.
* Version 251txr-251Kaz Kylheku2021-02-081-1/+1
| | | | | | | | | | * RELNOTES: Updated. * configure, txr.1: Bumped version and date. * share/txr/stdlib/ver.tl: Re-synced to 251. * txr.vim, tl.vim: Regenerated.
* matcher: fix backreferencing in predicate.Kaz Kylheku2021-02-081-1/+1
| | | | | | | | | * share/txr/stdlib/match.tl (compile-predicate-match): Always allocate res-var as a gensym; do not use resvar. Otherwise we will freshly bind resvar as a local, failing to back-reference. * tests/011/patmatch.tl: Add test cases, the second of which fails before this change.
* compiler: frame optimizations apply to dframe too.Kaz Kylheku2021-02-081-2/+2
| | | | | * share/txr/stdlib/optimize.tl (basic-blocks peephole): Use or pattern to look for dframe as well as frame.
* compiler: jump skid optimization.Kaz Kylheku2021-02-071-0/+8
| | | | | | | | | | | | | | | | | | If an (if reg label0) target branches due to reg being nil, and the target of the branch is another branch instrution of the form (ifq reg nil label1), we know that that branch is not taken: the code following that instruction is executed. THus can jump right to that code. (if reg label0) (if reg xlabel) ... ... label0 label0 (ifq reg nil label1) --> (ifq reg nil label1) ... xlabel ... * share/txr/stdlib/optimize.tl (basic-blocks peephole): New sub-case under (jmp @reg @jlabel).
* compiler: frame reduction optimizations.Kaz Kylheku2021-02-071-1/+36
| | | | | | | | | | | | | | These optimizations have to do with moving a (frame x y) instruction past the next instruction. The goal is to move the frame past a conditional branch, under the right circumstances, so that the frame is eliminated when the branch is taken. * share/txr/stdlib/optimize.tl (basic-blocks (cut-block, next-block)): New methods. (basic-block peephole): Add two patterns: one to move a frame past a mov, call or gcall. Another more complicated one to move it past an if which jumps to an end.
* matcher: exprs-syntax: process trivial matches first.Kaz Kylheku2021-02-071-12/+15
| | | | | | | * share/txr/stdlib/match.tl (compile-exprs-match): Sort the expressions and patterns so trivial matches are processed first. The original order is used for evaluating the expressions.
* matcher: eliminate use of flags.Kaz Kylheku2021-02-071-27/+18
| | | | | | | | | * share/txr/stdlib/match.tl (if-match, match-case, lambda-match): Instead of returning the result from the case(s), which gets stored in a result variable, and setting a flag to t, set the result variable inside the case, and return t. This eliminates the flag. In match-case and lambda-match, the cases can then be combined into an or form.
* matcher: remove @(op ...) pattern.Kaz Kylheku2021-02-061-7/+0
| | | | | | | | | | | | | | | All he typical uses of this are better served by the new predicate match. If op is really needed, it can be used with the DWIM form of the predicate, as in @[(op ...) ...]. * share/txr/stdlib/match.tl (compile-op-match): Function removed. (compile-match): Remove op case. * tests/011/patmatch.tl: Keep op test cases by converting them to predicate test cases. * txr.1: Documentation removed.