Hi Hans, all,

this was just announced on lua-l. Might be relevant to Context
since the advertised functionality intersects with mtx-package /
util-mrg.lua:

    Potential uses include:
    - Compressing Lua programs by removing comments and whitespace (strip.lua)
    - Removing assertions (assert.lua)
    - Adding new syntax sugar (self.lua)
    - Experimenting with new syntax without hacking the Lua source 
(reserved.lua)

I especially like the idea of having some kind of a preprocessor.

The original announcement can be found at Gmane in case the
attachment is eaten by the list:

    http://thread.gmane.org/gmane.comp.lang.lua.general/121329

Have fun,
Philipp

----- Forwarded message from Luiz Henrique de Figueiredo 
<l...@tecgraf.puc-rio.br> -----

Date: Wed, 4 May 2016 02:17:22 -0300
From: Luiz Henrique de Figueiredo <l...@tecgraf.puc-rio.br>
To: lu...@lists.lua.org
Subject: [ANN] ltokenp, a token processor for Lua
User-Agent: Mutt/1.5.20 (2009-12-10)

----- End forwarded message -----
ltokenp is now available at
        http://www.tecgraf.puc-rio.br/~lhf/ftp/lua/5.3/ltokenp.tar.gz

It should work for Lua 5.1, 5.2, and 5.3.

The README is attached. Enjoy. All feedback welcome.
--lhf

ltokenp is a token processor for Lua: it allows you to process the
stream of tokens coming from the Lua lexer.

Potential uses include:
- Compressing Lua programs by removing comments and whitespace (strip.lua)
- Removing assertions (assert.lua)
- Adding new syntax sugar (self.lua)
- Experimenting with new syntax without hacking the Lua source (reserved.lua)

ltokenp accepts Lua scripts to run and files to process.
The scripts and files are executed and processed as seen.
Each script appears as a separate argument after '-s', one '-s' per script.

Scripts typically output the contents of the files with some modifications.
Unfortunately, all comments and whitespace are eaten by the lexer and never
reach the token stream.

A typical usage is
        ltokenp -s script.lua [file.lua ...]
but you can also do
        ltokenp -s s1.lua f1.lua -s s2.lua f2.lua

A global function named FILTER will be called once for each token seen
in the files. So, scripts typically define FILTER. If no scripts are
given, ltokenp just dumps the token stream with this:

        function FILTER(line,token,text,value)
                print(line,token,text,value)
        end

As can been above, the FILTER function receives 4 arguments from the lexer:
- the line number where the token appears
- the token as a number
- the token as text
- the value of names, numbers, and strings;
  for other tokens, the value is the same as the text.

To try ltokenp, just edit Makefile to reflect your installation of Lua
and then run make. This will build ltokenp and run a simple test. Note
that ltokenp needs inside information from the Lua private headers and
so it needs to be built using a Lua build directory, not just the usual
installation directories.

This code is hereby placed in the public domain.
Please send comments, suggestions, and bug reports to l...@tecgraf.puc-rio.br .

Attachment: pgpxKNS940IqD.pgp
Description: PGP signature

___________________________________________________________________________________
If your question is of interest to others as well, please add an entry to the 
Wiki!

maillist : ntg-context@ntg.nl / http://www.ntg.nl/mailman/listinfo/ntg-context
webpage  : http://www.pragma-ade.nl / http://tex.aanhet.net
archive  : http://foundry.supelec.fr/projects/contextrev/
wiki     : http://contextgarden.net
___________________________________________________________________________________

Reply via email to