I've verified that Rubber *correctly* interprets the filename when the
parent directory has spaces in it. If I run
rubber /home/a folder/foo.tex
then rubber invokes the command
latex "\nonstopmode" "\input{/home/a folder/foo.tex}" .
This command fails. You can verify, by running it from the command line,
that the following command also fails
latex "\nonstopmode" "\input{"/home/a folder/foo.tex"}" .
But, if you are already in the directory "/home/a folder/" and run
latex "\nonstopmode" "\input{foo}"
then everything compiles correctly. I have attached a copy of the
file /rubber-1.1/src/rules/latex/__init__.py (re-compile and re-install
after replacing this file) that has been changed to invoke latex this
way. This is not a fix, as it will surely not work for most of rubber's
features. However, as long as you run rubber from within the same
directory as your .tex file it will properly compile.
The correct way to address this problem is to figure out how invoke
LaTeX with "\input{%s}" where %s is a file (with path) that may have
spaces in the name. My solution is messy and temporary.
Thanks,
Bob
# This file is part of Rubber and thus covered by the GPL
# (c) Emmanuel Beffara, 2002--2006
"""
LaTeX document building system for Rubber.
This module contains all the code in Rubber that actually does the job of
building a LaTeX document from start to finish.
"""
# Stop python 2.2 from calling "yield" statements syntax errors.
from __future__ import generators
import os, sys, posix
from os.path import *
import re
import string
from rubber import _
from rubber import *
from rubber.version import moddir
#---- Module handler ----{{{1
class Modules (Plugins):
"""
This class gathers all operations related to the management of modules.
The modules are searched for first in the current directory, then as
scripts in the 'modules' directory in the program's data directort, then
as a Python module in the package `rubber.latex'.
"""
def __init__ (self, env):
Plugins.__init__(self, rubber.rules.latex.__path__)
self.env = env
self.objects = {}
self.commands = {}
def __getitem__ (self, name):
"""
Return the module object of the given name.
"""
return self.objects[name]
def has_key (self, name):
"""
Check if a given module is loaded.
"""
return self.objects.has_key(name)
def register (self, name, dict={}):
"""
Attempt to register a package with the specified name. If a module is
found, create an object from the module's class called `Module',
passing it the environment and `dict' as arguments, and execute all
delayed commands for this module. The dictionary describes the
command that caused the registration.
"""
if self.has_key(name):
msg.debug(_("module %s already registered") % name)
return 2
# First look for a script
mod = None
for path in "", join(moddir, "modules"):
file = join(path, name + ".rub")
if exists(file):
mod = ScriptModule(self.env, file)
msg.log(_("script module %s registered") % name)
break
# Then look for a Python module
if not mod:
if Plugins.register(self, name) == 0:
msg.debug(_("no support found for %s") % name)
return 0
mod = self.modules[name].Module(self.env, dict)
msg.log(_("built-in module %s registered") % name)
# Run any delayed commands.
if self.commands.has_key(name):
for (cmd, args, vars) in self.commands[name]:
msg.push_pos(vars)
try:
mod.command(cmd, args)
except AttributeError:
msg.warn(_("unknown directive '%s.%s'") % (name, cmd))
except TypeError:
msg.warn(_("wrong syntax for '%s.%s'") % (name, cmd))
msg.pop_pos()
del self.commands[name]
self.objects[name] = mod
return 1
def clear (self):
"""
Unregister all modules.
"""
Plugins.clear(self)
self.objects = {}
self.commands = {}
def command (self, mod, cmd, args):
"""
Send a command to a particular module. If this module is not loaded,
store the command so that it will be sent when the module is register.
"""
if self.objects.has_key(mod):
self.objects[mod].command(cmd, args)
else:
if not self.commands.has_key(mod):
self.commands[mod] = []
self.commands[mod].append((cmd, args, self.env.vars.copy()))
#---- Log parser ----{{{1
re_loghead = re.compile("This is [0-9a-zA-Z-]*(TeX|Omega)")
re_rerun = re.compile("LaTeX Warning:.*Rerun")
re_file = re.compile("(\\((?P<file>[^ \n\t(){}]*)|\\))")
re_badbox = re.compile(r"(Ov|Und)erfull \\[hv]box ")
re_line = re.compile(r"(l\.(?P<line>[0-9]+)( (?P<code>.*))?$|<\*>)")
re_cseq = re.compile(r".*(?P<seq>\\[^ ]*) ?$")
re_page = re.compile("\[(?P<num>[0-9]+)\]")
re_atline = re.compile(
"( detected| in paragraph)? at lines? (?P<line>[0-9]*)(--(?P<last>[0-9]*))?")
re_reference = re.compile("LaTeX Warning: Reference `(?P<ref>.*)' \
on page (?P<page>[0-9]*) undefined on input line (?P<line>[0-9]*)\\.$")
re_label = re.compile("LaTeX Warning: (?P<text>Label .*)$")
re_warning = re.compile(
"(LaTeX|Package)( (?P<pkg>.*))? Warning: (?P<text>.*)$")
re_online = re.compile("(; reported)? on input line (?P<line>[0-9]*)")
re_ignored = re.compile("; all text was ignored after line (?P<line>[0-9]*).$")
class LogCheck (object):
"""
This class performs all the extraction of information from the log file.
For efficiency, the instances contain the whole file as a list of strings
so that it can be read several times with no disk access.
"""
#-- Initialization {{{2
def __init__ (self):
self.lines = None
def read (self, name):
"""
Read the specified log file, checking that it was produced by the
right compiler. Returns true if the log file is invalid or does not
exist.
"""
self.lines = None
try:
file = open(name)
except IOError:
return 2
line = file.readline()
if not line:
file.close()
return 1
if not re_loghead.match(line):
file.close()
return 1
self.lines = file.readlines()
file.close()
return 0
#-- Process information {{{2
def errors (self):
"""
Returns true if there was an error during the compilation.
"""
skipping = 0
for line in self.lines:
if line.strip() == "":
skipping = 0
continue
if skipping:
continue
m = re_badbox.match(line)
if m:
skipping = 1
continue
if line[0] == "!":
# We check for the substring "pdfTeX warning" because pdfTeX
# sometimes issues warnings (like undefined references) in the
# form of errors...
if string.find(line, "pdfTeX warning") == -1:
return 1
return 0
def run_needed (self):
"""
Returns true if LaTeX indicated that another compilation is needed.
"""
for line in self.lines:
if re_rerun.match(line):
return 1
return 0
#-- Information extraction {{{2
def continued (self, line):
"""
Check if a line in the log is continued on the next line. This is
needed because TeX breaks messages at 79 characters per line. We make
this into a method because the test is slightly different in Metapost.
"""
return len(line) == 79
def parse (self, errors=0, boxes=0, refs=0, warnings=0):
"""
Parse the log file for relevant information. The named arguments are
booleans that indicate which information should be extracted:
- errors: all errors
- boxes: bad boxes
- refs: warnings about references
- warnings: all other warnings
The function returns a generator. Each generated item is a dictionary
that contains (some of) the following entries:
- kind: the kind of information ("error", "box", "ref", "warning")
- text: the text of the error or warning
- code: the piece of code that caused an error
- file, line, last, pkg: as used by Message.format_pos.
"""
if not self.lines:
return
last_file = None
pos = [last_file]
page = 1
parsing = 0 # 1 if we are parsing an error's text
skipping = 0 # 1 if we are skipping text until an empty line
something = 0 # 1 if some error was found
prefix = None # the prefix for warning messages from packages
accu = "" # accumulated text from the previous line
for line in self.lines:
line = line[:-1] # remove the line feed
# TeX breaks messages at 79 characters, just to make parsing
# trickier...
if self.continued(line):
accu += line
continue
line = accu + line
accu = ""
# Text that should be skipped (from bad box messages)
if prefix is None and line == "":
skipping = 0
continue
if skipping:
continue
# Errors (including aborted compilation)
if parsing:
if error == "Undefined control sequence.":
# This is a special case in order to report which control
# sequence is undefined.
m = re_cseq.match(line)
if m:
error = "Undefined control sequence %s." % m.group("seq")
m = re_line.match(line)
if m:
parsing = 0
skipping = 1
pdfTeX = string.find(line, "pdfTeX warning") != -1
if (pdfTeX and warnings) or (errors and not pdfTeX):
if pdfTeX:
d = {
"kind": "warning",
"pkg": "pdfTeX",
"text": error[error.find(":")+2:]
}
else:
d = {
"kind": "error",
"text": error
}
d.update( m.groupdict() )
m = re_ignored.search(error)
if m:
d["file"] = last_file
if d.has_key("code"):
del d["code"]
d.update( m.groupdict() )
elif pos[-1] is None:
d["file"] = last_file
else:
d["file"] = pos[-1]
yield d
elif line[0] == "!":
error = line[2:]
elif line[0:3] == "***":
parsing = 0
skipping = 1
if errors:
yield {
"kind": "abort",
"text": error,
"why" : line[4:],
"file": last_file
}
elif line[0:15] == "Type X to quit ":
parsing = 0
skipping = 0
if errors:
yield {
"kind": "error",
"text": error,
"file": pos[-1]
}
continue
if len(line) > 0 and line[0] == "!":
error = line[2:]
parsing = 1
continue
if line == "Runaway argument?":
error = line
parsing = 1
continue
# Long warnings
if prefix is not None:
if line[:len(prefix)] == prefix:
text.append(string.strip(line[len(prefix):]))
else:
text = " ".join(text)
m = re_online.search(text)
if m:
info["line"] = m.group("line")
text = text[:m.start()] + text[m.end():]
if warnings:
info["text"] = text
d = { "kind": "warning" }
d.update( info )
yield d
prefix = None
continue
# Undefined references
m = re_reference.match(line)
if m:
if refs:
d = {
"kind": "warning",
"text": _("Reference `%s' undefined.") % m.group("ref"),
"file": pos[-1]
}
d.update( m.groupdict() )
yield d
continue
m = re_label.match(line)
if m:
if refs:
d = {
"kind": "warning",
"file": pos[-1]
}
d.update( m.groupdict() )
yield d
continue
# Other warnings
if line.find("Warning") != -1:
m = re_warning.match(line)
if m:
info = m.groupdict()
info["file"] = pos[-1]
info["page"] = page
if info["pkg"] is None:
del info["pkg"]
prefix = ""
else:
prefix = ("(%s)" % info["pkg"])
prefix = prefix.ljust(m.start("text"))
text = [info["text"]]
continue
# Bad box messages
m = re_badbox.match(line)
if m:
if boxes:
mpos = { "file": pos[-1], "page": page }
m = re_atline.search(line)
if m:
md = m.groupdict()
for key in "line", "last":
if md[key]: mpos[key] = md[key]
line = line[:m.start()]
d = {
"kind": "warning",
"text": line
}
d.update( mpos )
yield d
skipping = 1
continue
# If there is no message, track source names and page numbers.
last_file = self.update_file(line, pos, last_file)
page = self.update_page(line, page)
def get_errors (self):
return self.parse(errors=1)
def get_boxes (self):
return self.parse(boxes=1)
def get_references (self):
return self.parse(refs=1)
def get_warnings (self):
return self.parse(warnings=1)
def update_file (self, line, stack, last):
"""
Parse the given line of log file for file openings and closings and
update the list `stack'. Newly opened files are at the end, therefore
stack[1] is the main source while stack[-1] is the current one. The
first element, stack[0], contains the value None for errors that may
happen outside the source. Return the last file from which text was
read (the new stack top, or the one before the last closing
parenthesis).
"""
m = re_file.search(line)
while m:
if line[m.start()] == '(':
last = m.group("file")
stack.append(last)
else:
last = stack[-1]
del stack[-1]
line = line[m.end():]
m = re_file.search(line)
return last
def update_page (self, line, before):
"""
Parse the given line and return the number of the page that is being
built after that line, assuming the current page before the line was
`before'.
"""
ms = re_page.findall(line)
if ms == []:
return before
return int(ms[-1]) + 1
#---- Parsing and compiling ----{{{1
re_comment = re.compile(r"(?P<line>([^\\%]|\\%|\\)*)(%.*)?")
re_command = re.compile("[% ]*(rubber: *(?P<cmd>[^ ]*) *(?P<arg>.*))?.*")
re_input = re.compile("\\\\input +(?P<arg>[^{} \n\\\\]+)")
class EndDocument:
""" This is the exception raised when \\end{document} is found. """
pass
class EndInput:
""" This is the exception raised when \\endinput is found. """
pass
class LaTeXDep (Depend):
"""
This class represents dependency nodes for LaTeX compilation. It handles
the cyclic LaTeX compilation until a stable output, including actual
compilation (with a parametrable executable) and possible processing of
compilation results (e.g. running BibTeX).
Before building (or cleaning) the document, the method `parse' must be
called to load and configure all required modules. Text lines are read
from the files and parsed to extract LaTeX macro calls. When such a macro
is found, a handler is searched for in the `hooks' dictionary. Handlers
are called with one argument: the dictionary for the regular expression
that matches the macro call.
"""
#-- Initialization {{{2
def __init__ (self, env):
"""
Initialize the environment. This prepares the processing steps for the
given file (all steps are initialized empty) and sets the regular
expressions and the hook dictionary.
"""
Depend.__init__(self, env)
self.log = LogCheck()
self.modules = Modules(self)
if env.caching:
if not env.cache.has_key("latex"):
env.cache["latex"] = {}
self.vars = env.vars.copy()
self.vars.update({
"program": "latex",
"engine": "TeX",
"paper": "" })
self.vars_stack = []
self.cache_list = []
self.cmdline = ["\\nonstopmode", "\\input{%s}"]
# the initial hooks:
self.comment_mark = "%"
self.hooks = {
"input" : self.h_input,
"include" : self.h_include,
"includeonly": self.h_includeonly,
"usepackage" : self.h_usepackage,
"RequirePackage" : self.h_usepackage,
"documentclass" : self.h_documentclass,
"tableofcontents" : self.h_tableofcontents,
"listoffigures" : self.h_listoffigures,
"listoftables" : self.h_listoftables,
"bibliography" : self.h_bibliography,
"bibliographystyle" : self.h_bibliographystyle,
"begin{verbatim}" : self.h_begin_verbatim,
"begin{verbatim*}" :
lambda d: self.h_begin_verbatim(d, end="end{verbatim*}"),
"endinput" : self.h_endinput,
"end{document}" : self.h_end_document
}
self.update_seq()
self.include_only = {}
# description of the building process:
self.aux_md5 = {}
self.aux_old = {}
self.watched_files = {}
self.onchange_md5 = {}
self.onchange_cmd = {}
self.removed_files = []
self.not_included = [] # dependencies that don't trigger latex
# state of the builder:
self.processed_sources = {}
self.must_compile = 0
self.something_done = 0
self.failed_module = None
def set_source (self, path):
"""
Specify the main source for the document. The exact path and file name
are determined, and the source building process is updated if needed,
according the the source file's extension.
"""
name = self.env.find_file(path, ".tex")
if not name:
return 1
self.sources = {}
(self.src_path, name) = split(name)
(self.src_base, self.src_ext) = splitext(name)
if self.src_path == "":
self.src_path = "."
self.src_pbase = self.src_base
else:
self.env.path.append(self.src_path)
self.src_pbase = join(self.src_path, self.src_base)
self.prods = [self.src_base + ".dvi"]
self.vars['job'] = self.src_base
self.vars['base'] = self.src_pbase
return 0
def includeonly (self, files):
"""
Use partial compilation, by appending a call to \\inlcudeonly on the
command line on compilation.
"""
if self.vars["engine"] == "VTeX":
msg.error(_("I don't know how to do partial compilation on VTeX."))
return
if self.cmdline[-2][:13] == "\\includeonly{":
self.cmdline[-2] = "\\includeonly{" + ",".join(files) + "}"
else:
self.cmdline.insert(-1, "\\includeonly{" + ",".join(files) + "}")
for f in files:
self.include_only[f] = None
def source (self):
"""
Return the main source's complete filename.
"""
return self.src_pbase + self.src_ext
#-- Variable handling {{{2
def push_vars (self, **dict):
"""
For each named argument "key=val", save the value of variable "key"
and assign it the value "val".
"""
saved = {}
for (key, val) in dict.items():
saved[key] = self.vars[key]
self.vars[key] = val
self.vars_stack.append(saved)
def pop_vars (self):
"""
Restore the last set of variables saved using "push_vars".
"""
self.vars.update(self.vars_stack[-1])
del self.vars_stack[-1]
def abspath (self, name, ref=None):
"""
Return the absolute path of a given filename. Relative paths are
considered relative to the file currently processed, the optional
argument "ref" can be used to override the reference file name.
"""
path = self.vars["cwd"]
if ref is None and self.vars.has_key("file"):
ref = self.vars["file"]
if ref is not None:
path = join(path, dirname(ref))
return abspath(join(path, expanduser(name)))
#-- LaTeX source parsing {{{2
def parse (self):
"""
Parse the source for packages and supported macros.
"""
self.vars["file"] = None
self.vars["line"] = None
try:
self.process(self.source())
except EndDocument:
pass
del self.vars["file"]
del self.vars["line"]
self.set_date()
msg.log(_("dependencies: %r") % self.sources.keys())
def parse_file (self, file, dump=None):
"""
Process a LaTeX source. The file must be open, it is read to the end
calling the handlers for the macro calls. This recursively processes
the included sources.
If the optional argument 'dump' is not None, then it is considered as
a stream on which all text not matched as a macro is written.
"""
lines = file.readlines()
lineno = 0
vars = self.vars
# If a line ends with braces open, we read on until we get a correctly
# braced text. We also stop accumulating on paragraph breaks, the way
# non-\long macros do in TeX.
brace_level = 0
accu = ""
for line in lines:
lineno = lineno + 1
# Lines that start with a comment are the ones where directives
# may be found.
if line[0] == self.comment_mark:
m = re_command.match(string.rstrip(line))
if m.group("cmd"):
vars['line'] = lineno
args = parse_line(m.group("arg"), vars)
if self.env.caching:
self.cache_list.append(("cmd", m.group("cmd"), args, vars))
self.command(m.group("cmd"), args, vars)
continue
# Otherwise we accumulate lines (with comments stripped) until
# bracing is correct.
line = re_comment.match(line).group("line")
if accu != "" and accu[-1] != '\n':
line = string.lstrip(line)
brace_level = brace_level + count_braces(line)
if brace_level <= 0 or string.strip(line) == "":
brace_level = 0
line = accu + line
accu = ""
else:
accu = accu + line
continue
# Then we check for supported macros in the text.
match = self.seq.search(line)
while match:
dict = match.groupdict()
name = dict["name"]
# The case of \input is handled specifically, because of the
# TeX syntax with no braces
if name == "input" and not dict["arg"]:
match2 = re_input.search(line)
if match2:
match = match2
dict = match.groupdict()
if dump: dump.write(line[:match.start()])
dict["match"] = line[match.start():match.end()]
dict["line"] = line[match.end():]
dict["pos"] = { 'file': self.vars["file"], 'line': lineno }
dict["dump"] = dump
if self.env.caching:
self.cache_list.append(("hook", name, dict))
self.hooks[name](dict)
line = dict["line"]
match = self.seq.search(line)
if dump: dump.write(line)
def process (self, path, loc={}):
"""
This method is called when an included file is processed. The argument
must be a valid file name.
"""
if self.processed_sources.has_key(path):
msg.debug(_("%s already parsed") % path)
return
self.processed_sources[path] = None
if not self.sources.has_key(path):
self.sources[path] = DependLeaf(self.env, path, loc=loc)
if self.env.caching:
if self.env.cache["latex"].has_key(path):
(date, list) = self.env.cache["latex"][path]
fdate = getmtime(path)
if fdate <= date:
msg.log(_("using cache for %s") % path)
for elem in list:
if elem[0] == "hook":
try:
self.hooks[elem[1]](elem[2])
except EndInput:
pass
elif elem[0] == "cmd":
self.command(*elem[1:])
return
else:
msg.log(_("cache for %s is obsolete") % path)
saved_cache = self.cache_list
self.cache_list = []
try:
try:
msg.log(_("parsing %s") % path)
self.push_vars(file=path, line=None)
file = open(path)
try:
self.parse_file(file)
finally:
file.close()
finally:
self.pop_vars()
msg.debug(_("end of %s") % path)
if self.env.caching:
self.env.cache["latex"][path] = (
getmtime(path), self.cache_list)
self.cache_list = saved_cache
except EndInput:
pass
def input_file (self, name, loc={}):
"""
Treat the given name as a source file to be read. If this source can
be the result of some conversion, then the conversion is performed,
otherwise the source is parsed. The returned value is a couple
(name,dep) where `name' is the actual LaTeX source and `dep' is
its dependency node. The return value is (None,None) is the source
could neither be read nor built.
"""
if name.find("\\") >= 0 or name.find("#") >= 0:
return None, None
for path in self.env.path:
pname = join(path, name)
dep = self.env.convert(pname, suffixes=[".tex",""], doc=self)
if dep:
dep.loc = loc
file = dep.prods[0]
self.sources[file] = dep
else:
file = self.env.find_file(name, ".tex")
if not file:
continue
dep = None
if dep is None or isinstance(dep, DependLeaf):
self.process(file, loc)
if dep is None:
return file, self.sources[file]
else:
return file, dep
return None, None
#-- Directives {{{2
def command (self, cmd, args, pos=None):
"""
Execute the rubber command 'cmd' with arguments 'args'. This is called
when a command is found in the source file or in a configuration file.
A command name of the form 'foo.bar' is considered to be a command
'bar' for module 'foo'. The argument 'pos' describes the position
(file and line) where the command occurs.
"""
if pos is None:
pos = self.vars
# Calls to this method are actually translated into calls to "do_*"
# methods, except for calls to module directives.
lst = string.split(cmd, ".", 1)
try:
if len(lst) > 1:
self.modules.command(lst[0], lst[1], args)
else:
getattr(self, "do_" + cmd)(*args)
except AttributeError:
msg.warn(_("unknown directive '%s'") % cmd, **pos)
except TypeError:
msg.warn(_("wrong syntax for '%s'") % cmd, **pos)
def do_alias (self, name, val):
if self.hooks.has_key(val):
self.hooks[name] = self.hooks[val]
self.update_seq()
def do_clean (self, *args):
for file in args:
self.removed_files.append(self.abspath(file))
def do_depend (self, *args):
for arg in args:
file = self.env.find_file(arg)
if file:
self.sources[file] = DependLeaf(self.env, file)
else:
msg.warn(_("dependency '%s' not found") % arg, **self.vars)
def do_make (self, file, *args):
file = self.abspath(file)
vars = { "target": file }
while len(args) > 1:
if args[0] == "from":
vars["source"] = self.abspath(args[1])
elif args[0] == "with":
vars["name"] = args[1]
else:
break
args = args[2:]
if len(args) != 0:
msg.error(_("invalid syntax for 'make'"), **self.vars)
return
self.env.conv_set(file, vars)
def do_module (self, mod, opt=None):
dict = { 'arg': mod, 'opt': opt }
self.modules.register(mod, dict)
def do_onchange (self, file, cmd):
file = self.abspath(file)
self.onchange_cmd[file] = cmd
if exists(file):
self.onchange_md5[file] = md5_file(file)
else:
self.onchange_md5[file] = None
def do_paper (self, arg):
self.vars["paper"] = arg
def do_path (self, name):
self.env.path.append(self.abspath(name))
def do_read (self, name):
path = self.abspath(name)
self.push_vars(file=path, line=None)
try:
file = open(path)
lineno = 0
for line in file.readlines():
lineno += 1
line = line.strip()
if line == "" or line[0] == "%":
continue
self.vars["line"] = lineno
lst = parse_line(line, self.vars)
self.command(lst[0], lst[1:])
file.close()
except IOError:
msg.warn(_("cannot read option file %s") % name, **self.vars)
self.pop_vars()
def do_rules (self, file):
name = self.env.find_file(file)
if name is None:
msg.warn(_("cannot read rule file %s") % file, **self.vars)
else:
self.env.user_rules.read_ini(name)
def do_set (self, name, val):
self.vars[name] = val
def do_watch (self, *args):
for arg in args:
self.watch_file(self.abspath(arg))
#-- Macro handling {{{2
def update_seq (self):
"""
Update the regular expression used to match macro calls using the keys
in the `hook' dictionary. We don't match all control sequences for
obvious efficiency reasons.
"""
clean = map(lambda x: x.replace("*", "\\*"), self.hooks.keys())
self.seq = re.compile("\
\\\\(?P<name>%s)\*?\
*(\\[(?P<opt>[^\\]]*)\\])?\
*({(?P<arg>[^{}]*)}|(?=[^A-Za-z]))"
% string.join(clean, "|"))
def add_hook (self, name, fun):
"""
Register a given function to be called (with no arguments) when a
given macro is found.
"""
self.hooks[name] = fun
self.update_seq()
# Now the macro handlers:
def h_input (self, dict):
"""
Called when an \\input macro is found. This calls the `process' method
if the included file is found.
"""
if dict["arg"]:
self.input_file(dict["arg"], dict)
def h_include (self, dict):
"""
Called when an \\include macro is found. This includes files into the
source in a way very similar to \\input, except that LaTeX also
creates .aux files for them, so we have to notice this.
"""
if not dict["arg"]:
return
if self.include_only and not self.include_only.has_key(dict["arg"]):
return
file, _ = self.input_file(dict["arg"], dict)
if file:
aux = dict["arg"] + ".aux"
self.removed_files.append(aux)
self.aux_old[aux] = None
if exists(aux):
self.aux_md5[aux] = md5_file(aux)
else:
self.aux_md5[aux] = None
def h_includeonly (self, dict):
"""
Called when the macro \\includeonly is found, indicates the
comma-separated list of files that should be included, so that the
othe \\include are ignored.
"""
if not dict["arg"]:
return
self.include_only = {}
for name in dict["arg"].split(","):
name = name.strip()
if name != "":
self.include_only[name] = None
def h_documentclass (self, dict):
"""
Called when the macro \\documentclass is found. It almost has the same
effect as `usepackage': if the source's directory contains the class
file, in which case this file is treated as an input, otherwise a
module is searched for to support the class.
"""
if not dict["arg"]: return
file = self.env.find_file(dict["arg"] + ".cls")
if file:
self.process(file)
else:
self.modules.register(dict["arg"], dict)
def h_usepackage (self, dict):
"""
Called when a \\usepackage macro is found. If there is a package in the
directory of the source file, then it is treated as an include file
unless there is a supporting module in the current directory,
otherwise it is treated as a package.
"""
if not dict["arg"]: return
for name in string.split(dict["arg"], ","):
name = name.strip()
file = self.env.find_file(name + ".sty")
if file and not exists(name + ".py"):
self.process(file)
else:
self.modules.register(name, dict)
def h_tableofcontents (self, dict):
self.watch_file(self.src_base + ".toc")
def h_listoffigures (self, dict):
self.watch_file(self.src_base + ".lof")
def h_listoftables (self, dict):
self.watch_file(self.src_base + ".lot")
def h_bibliography (self, dict):
"""
Called when the macro \\bibliography is found. This method actually
registers the module bibtex (if not already done) and registers the
databases.
"""
if dict["arg"]:
self.modules.register("bibtex", dict)
for db in dict["arg"].split(","):
self.modules["bibtex"].add_db(db.strip())
def h_bibliographystyle (self, dict):
"""
Called when \\bibliographystyle is found. This registers the module
bibtex (if not already done) and calls the method set_style() of the
module.
"""
if dict["arg"]:
self.modules.register("bibtex", dict)
self.modules["bibtex"].set_style(dict["arg"])
def h_begin_verbatim (self, dict, end="end{verbatim}"):
"""
Called when \\begin{verbatim} is found. This disables all macro
handling and comment parsing until the end of the environment. The
optional argument 'end' specifies the end marker, by default it is
"\\end{verbatim}".
"""
def end_verbatim (dict, self=self, hooks=self.hooks):
self.hooks = hooks
self.comment_mark = "%"
self.update_seq()
self.hooks = { end : end_verbatim }
self.update_seq()
self.comment_mark = None
def h_endinput (self, dict):
"""
Called when \\endinput is found. This stops the processing of the
current input file, thus ignoring any code that appears afterwards.
"""
raise EndInput
def h_end_document (self, dict):
"""
Called when \\end{document} is found. This stops the processing of any
input file, thus ignoring any code that appears afterwards.
"""
raise EndDocument
#-- Compilation steps {{{2
def compile (self):
"""
Run one LaTeX compilation on the source. Return true if errors
occured, and false if compilaiton succeeded.
"""
msg.progress(_("compiling %s") % msg.simplify(self.source()))
file = self.source()
fakefile = self.src_base
cmd = [self.vars["program"]]
cmd += map(lambda x: x.replace("%s",fakefile), self.cmdline)
inputs = string.join(self.env.path, ":")
if inputs == "":
env = {}
else:
inputs = inputs + ":" + os.getenv("TEXINPUTS", "")
env = {"TEXINPUTS": inputs}
self.env.execute(cmd, env, kpse=1)
self.something_done = 1
if self.log.read(self.src_base + ".log"):
msg.error(_("Could not run %s.") % cmd[0])
return 1
if self.log.errors():
return 1
for aux, md5 in self.aux_md5.items():
self.aux_old[aux] = md5
self.aux_md5[aux] = md5_file(aux)
return 0
def pre_compile (self, force):
"""
Prepare the source for compilation using package-specific functions.
This function must return true on failure. This function sets
`must_compile' to 1 if we already know that a compilation is needed,
because it may avoid some unnecessary preprocessing (e.g. BibTeXing).
"""
aux = self.src_base + ".aux"
if os.path.exists(aux):
self.aux_md5[aux] = md5_file(aux)
else:
self.aux_md5[aux] = None
self.aux_old[aux] = None
self.log.read(self.src_base + ".log")
self.must_compile = force
self.must_compile = self.compile_needed()
msg.log(_("building additional files..."))
for mod in self.modules.objects.values():
if mod.pre_compile():
self.failed_module = mod
return 1
return 0
def post_compile (self):
"""
Run the package-specific operations that are to be performed after
each compilation of the main source. Returns true on failure.
"""
msg.log(_("running post-compilation scripts..."))
for file, md5 in self.onchange_md5.items():
if not exists(file):
continue
new = md5_file(file)
if md5 != new:
msg.progress(_("running %s") % self.onchange_cmd[file])
self.env.execute(["sh", "-c", self.onchange_cmd[file]])
self.onchange_md5[file] = new
for mod in self.modules.objects.values():
if mod.post_compile():
self.failed_module = mod
return 1
return 0
def clean (self, all=0):
"""
Remove all files that are produced by compilation.
"""
self.remove_suffixes([".log", ".aux", ".toc", ".lof", ".lot"])
for file in self.prods + self.removed_files:
if exists(file):
msg.log(_("removing %s") % file)
os.unlink(file)
msg.log(_("cleaning additional files..."))
for dep in self.sources.values():
dep.clean()
for mod in self.modules.objects.values():
mod.clean()
#-- Building routine {{{2
def force_run (self):
return self.run(1)
def run (self, force=0):
"""
Run the building process until the last compilation, or stop on error.
This method supposes that the inputs were parsed to register packages
and that the LaTeX source is ready. If the second (optional) argument
is true, then at least one compilation is done. As specified by the
class Depend, the method returns 0 on success and 1 on failure.
"""
if self.pre_compile(force):
return 1
# If an error occurs after this point, it will be while LaTeXing.
self.failed_dep = self
self.failed_module = None
if force or self.compile_needed():
self.must_compile = 0
if self.compile(): return 1
if self.post_compile(): return 1
while self.recompile_needed():
self.must_compile = 0
if self.compile(): return 1
if self.post_compile(): return 1
# Finally there was no error.
self.failed_dep = None
if self.something_done:
self.date = int(time.time())
return 0
def compile_needed (self):
"""
Returns true if a first compilation is needed. This method supposes
that no compilation was done (by the script) yet.
"""
if self.must_compile:
return 1
msg.log(_("checking if compiling is necessary..."))
if not exists(self.prods[0]):
msg.debug(_("the output file doesn't exist"))
return 1
if not exists(self.src_base + ".log"):
msg.debug(_("the log file does not exist"))
return 1
if getmtime(self.prods[0]) < getmtime(self.source()):
msg.debug(_("the source is younger than the output file"))
return 1
if self.log.read(self.src_base + ".log"):
msg.debug(_("the log file is not produced by TeX"))
return 1
return self.recompile_needed()
def recompile_needed (self):
"""
Returns true if another compilation is needed. This method is used
when a compilation has already been done.
"""
if self.must_compile:
self.update_watches()
return 1
if self.log.errors():
msg.debug(_("last compilation failed"))
self.update_watches()
return 1
if self.deps_modified(getmtime(self.prods[0])):
msg.debug(_("dependencies were modified"))
self.update_watches()
return 1
suffix = self.update_watches()
if suffix:
msg.debug(_("the %s file has changed") % suffix)
return 1
if self.log.run_needed():
msg.debug(_("LaTeX asks to run again"))
aux_changed = 0
for aux, md5 in self.aux_md5.items():
if md5 is not None and md5 != self.aux_old[aux]:
aux_changed = 1
break
if not aux_changed:
msg.debug(_("but the aux files are unchanged"))
return 0
return 1
msg.debug(_("no new compilation is needed"))
return 0
def deps_modified (self, date):
"""
Returns true if any of the dependencies is younger than the specified
date.
"""
for name, dep in self.sources.items():
if name not in self.not_included and dep.date > date:
return 1
return 0
#-- Utility methods {{{2
def get_errors (self):
if self.failed_module is None:
return self.log.get_errors()
else:
return self.failed_module.get_errors()
def watch_file (self, file):
"""
Register the given file (typically "jobname.toc" or such) to be
watched. When the file changes during a compilation, it means that
another compilation has to be done.
"""
if exists(file):
self.watched_files[file] = md5_file(file)
else:
self.watched_files[file] = None
def update_watches (self):
"""
Update the MD5 sums of all files watched, and return the name of one
of the files that changed, or None of they didn't change.
"""
changed = None
for file in self.watched_files.keys():
if exists(file):
new = md5_file(file)
if self.watched_files[file] != new:
changed = file
self.watched_files[file] = new
return changed
def remove_suffixes (self, list):
"""
Remove all files derived from the main source with one of the
specified suffixes.
"""
for suffix in list:
file = self.src_base + suffix
if exists(file):
msg.log(_("removing %s") % file)
os.unlink(file)
#---- Base classes for modules ----{{{1
class Module (object):
"""
This is the base class for modules. Each module should define a class
named 'Module' that derives from this one. The default implementation
provides all required methods with no effects.
"""
def __init__ (self, env, dict):
"""
The constructor receives two arguments: 'env' is the compiling
environment, 'dict' is a dictionary that describes the command that
caused the module to load.
"""
def pre_compile (self):
"""
This method is called before the first LaTeX compilation. It is
supposed to build any file that LaTeX would require to compile the
document correctly. The method must return true on failure.
"""
return 0
def post_compile (self):
"""
This method is called after each LaTeX compilation. It is supposed to
process the compilation results and possibly request a new
compilation. The method must return true on failure.
"""
return 0
def clean (self):
"""
This method is called when cleaning the compiled files. It is supposed
to remove all the files that this modules generates.
"""
def command (self, cmd, args):
"""
This is called when a directive for the module is found in the source.
The method can raise 'AttributeError' when the directive does not
exist and 'TypeError' if the syntax is wrong. By default, when called
with argument "foo" it calls the method "do_foo" if it exists, and
fails otherwise.
"""
getattr(self, "do_" + cmd)(*args)
def get_errors (self):
"""
This is called if something has failed during an operation performed
by this module. The method returns a generator with items of the same
form as in LaTeXDep.get_errors.
"""
if None:
yield None
class ScriptModule (Module):
"""
This class represents modules that are defined as Rubber scripts.
"""
def __init__ (self, env, filename):
vars = env.vars.copy()
vars['file'] = filename
lineno = 0
file = open(filename)
for line in file.readlines():
line = line.strip()
lineno = lineno + 1
if line == "" or line[0] == "%":
continue
vars['line'] = lineno
lst = parse_line(line, vars)
env.command(lst[0], lst[1:], vars)
file.close()