I think there are many good reasons why XSLT is absolutely the wrong tool for
the job of indexing MARC records for Solr.
1) Performance/Speed: In my experience even just transforming from MARCXML to
MODS takes a second or two (using the LoC stylesheet), due to the stylesheet's
complexity and inefficiency of doing heavy-duty string manipulation in XSL.
That means you're looking at an indexing speed of around 1 record/second. If
you've got 1,000,000 bib records, it'll take a couple of weeks just to index
your data. For comparison, the indexer of our commercial OPAC does about 50
records per second (~6 hours for a million records) and the one I've written in
Jython (by no means the fastest language out there) that doesn't use XSL can do
about 150 records a second (about 2 hours for 1 million records).
2) Reusability: What if you want to change how a field is indexed? You would
have to edit the XSLT directly (or have the XSL stylesheet automatically
generated based on settings stored elsewhere).
a) Users of the indexer shouldn't have to actually mess with programming logic
to change how it indexes. You shouldn't have to know a thing about programming
to change the setup of an index.
b) It should be easy for an external application to know how your indexes have
been built. This would be very difficult with an XSL stylesheet. Burying
configuration inside of programming logic is a bad idea.
c) The Solr schema should be automatically generated from your index setup so
all your index configuration is in one place. I guess you could write
*another* XSL stylesheet that would transform your indexing stylesheet into the
Solr schema file, but that seems ridiculous.
d) Automatic code generation is evil. Blanchard's law: "Systems that require
code generation lack sufficient power to tackle the problem at hand." If you
find yourself considering automatic code generation, you should instead be
considering a more dynamic programming language.
3) Ease of programming.
a) Heavy-duty string manipulation is a pain in pure XSLT. To index MARC
records have to do normalization on dates and names and you probably want to do
some translation between MARC codes and their meanings (for the audience &
language codes, for instance). Is it doable? Yes, especially if you use XSL
extension functions. But if you're going to have huge chunks of your logic
buried in extension functions, why not go whole hog and do it all outside of
XSLT, instead of having half your programming logic in an extension function
and half in the XSLT itself?
b) Using XSLT makes object-oriented programming with your data harder. Your
indexer should be able to give you a nice object representation of a record (so
you can use that object representation within other code). If you go the XSLT
route, you'd have to parse the MARC record, transform it to your Solr record
XML format, then parse that XML and map the XML to an object. If you avoid
XSLT, you just parse the MARC record and transform it to an object
programmatically (with the object having a method to print itself out as a Solr
XML record).
Honestly, all this talk of using XSLT for indexing MARC records reminds me of
that guy who rode across the United States on a riding lawnmower. I am looking
forward to there being a standard, well-tested MARC record indexer for Solr
(and would be excited to contribute to such a project), but I don't think that
XSL is the right tool to use.
--Casey