Yes understood. I do not have a problem in defining the parameters in the code. But the problem is, I am using PowerPivot as the visualization engine.
Now, when I give the query as a set like: -------------------- add jar /usr/local/hadoop/src/retweetlink1.jar; create temporary function link as 'retweetlink'; select link(tweet),count(*) as countlink from tweetsdata where tweet like '%RT%' group by link(tweet) sort by countlink desc limit 20; -------------------- The error then is that /usr/local/hadoop/src/retweetlink1.jar; does not exist So, I need to give them as seperate queries like add jar first, execute that followed by others. Now, the problem is it runs fine for the first instance of 3 mapreduce jobs, but when the second instance runs, for importing data, then throws up error saying that retweetlink class not found.