Shared Job /tmp location should be configurable
-----------------------------------------------
Key: PIG-103
URL: https://issues.apache.org/jira/browse/PIG-103
Project: Pig
Issue Type: Improvement
Components: impl
Environment: Partially shared file:// filesystem (eg NFS)
Reporter: Craig Macdonald
Hello,
I'm investigating running pig in an environment where various parts of the
file:// filesystem are available on all nodes. I can tell hadoop to use a
file:// file system location for it's default, by seting
fs.default.name=file://path/to/shared/folder
However, this creates issues for Pig, as Pig writes it's job information in a
folder that it assumes is a shared FS (eg DFS). However, in this scenario /tmp
is not shared on each machine.
So /tmp should either be configurable, or Hadoop should tell you the actual
full location set in fs.default.name?
Straightforward solution is to make "/tmp/" a property in
src/org/apache/pig/impl/io/FileLocalizer.java init(PigContext)
Any suggestions of property names?
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.