Hi Peter,

Of course, though you might want to think a bit on the approach.

1) If you want to do it in XWiki using the scheduler plugin, you have to
write some groovy script [1] in the scheduled task that you create in order
to use the filter module [2] (or the old packager module [3], whichever you
choose) in order to export the pages you are interested in and save the
result (on the filesystem or as an attachment in a page).

As inspiration, you could check out the source code (in java) of the export
action [4] which is exactly what is going on when you are exporting a page
from the UI.

2) A simpler choice, IMO, would be to schedule a cron [5] task on a linux
machine (possible the same one running the XWiki instance) and have that
task simply download (using something like cURL [6]) the result of a
manually crafter export URL (as detailed in our documentation [7])
containing the list of pages to export.

Hope this helps,

[1] http://platform.xwiki.org/xwiki/bin/view/DevGuide/Scripting
[2] http://extensions.xwiki.org/xwiki/bin/view/Extension/Filter+Module
[5] https://en.wikipedia.org/wiki/Cron
[6] https://en.wikipedia.org/wiki/CURL
[7] http://platform.xwiki.org/xwiki/bin/view/Features/Exports#HXARExport

On Thu, Sep 17, 2015 at 3:49 PM, Peter Huisman <p.huis...@ximm.nl> wrote:

> Hi,
> Is there a way to generate XAR’s using XWiki’s scheduling function? I’m
> not looking for a DB based backup but merely for a backup of spaces / pages.
> With kind regards,
> Peter
> _______________________________________________
> users mailing list
> users@xwiki.org
> http://lists.xwiki.org/mailman/listinfo/users
users mailing list

Reply via email to