Hi,
On 02.08.2011, at 13:19, Boian Mihailov wrote:
> Hello everyone. I am fairly new to puppet, and i just love it.
>
> I am trying to distribute websites across my web servers.
> I've tried file copy with recursion but its fairly slow process and i
> believe its not quite right to do it this way.
>
> Now i am thinking of making tar.gz of each site and distribute it to
> server than extract it there to the webroot.
>
> My exact question is how to make puppet unarchive this tar.gz only
> when its changed, so its not done every time the puppet resyncs
>
> something like:
>
> class pwiki {
> file { "/srv/www/pwiki.tar.gz":
> source => "puppet://puppet/files/srv/www/pwiki.tar.gz"
> }
>
> exec { "pwiki.tar.gz":
> command => "tar zxf pwiki.tar.gz",
> subscribe => File["/srv/www/pwiki.tar.gz"]
> }
> }
add the following line to your exec resource:
refreshonly => true,
Regards,
Martin
>
> --
> You received this message because you are subscribed to the Google Groups
> "Puppet Users" group.
> To post to this group, send email to [email protected].
> To unsubscribe from this group, send email to
> [email protected].
> For more options, visit this group at
> http://groups.google.com/group/puppet-users?hl=en.
>
--
You received this message because you are subscribed to the Google Groups
"Puppet Users" group.
To post to this group, send email to [email protected].
To unsubscribe from this group, send email to
[email protected].
For more options, visit this group at
http://groups.google.com/group/puppet-users?hl=en.