Revision: 6110
http://ipcop.svn.sourceforge.net/ipcop/?rev=6110&view=rev
Author: gespinasse
Date: 2011-11-18 10:29:09 +0000 (Fri, 18 Nov 2011)
Log Message:
-----------
Remove patch now unused
Removed Paths:
-------------
ipcop/trunk/src/patches/oinkmaster-2.0_less-space-usage.patch
Deleted: ipcop/trunk/src/patches/oinkmaster-2.0_less-space-usage.patch
===================================================================
--- ipcop/trunk/src/patches/oinkmaster-2.0_less-space-usage.patch
2011-11-18 10:21:16 UTC (rev 6109)
+++ ipcop/trunk/src/patches/oinkmaster-2.0_less-space-usage.patch
2011-11-18 10:29:09 UTC (rev 6110)
@@ -1,106 +0,0 @@
---- oinkmaster-2.0/oinkmaster.pl.old 2007-04-16 13:15:51.000000000 +0000
-+++ oinkmaster-2.0/oinkmaster.pl 2007-06-08 17:10:39.000000000 +0000
-@@ -1062,101 +1062,10 @@
- my $dir = dirname($archive);
- chdir("$dir") or clean_exit("$url: could not change directory to
\"$dir\": $!");
-
-- if ($config{use_external_bins}) {
--
-- # Run integrity check on the gzip file.
-- clean_exit("$url: integrity check on gzip file failed (file transfer
failed or ".
-- "file in URL not in gzip format?).")
-- if (system("gzip", "-t", "$archive"));
--
-- # Decompress it.
-- system("gzip", "-d", "$archive")
-- and clean_exit("$url: unable to uncompress $archive.");
--
-- # Suffix has now changed from .tar.gz|.tgz to .tar.
-- $archive =~ s/\.gz$//;
--
-- # Make sure the .tar file now exists.
-- # (Gzip may not return an error if it was not a gzipped file...)
-- clean_exit("$url: failed to unpack gzip file (file transfer failed or
".
-- "file in URL not in tar'ed gzip format?).")
-- unless (-e "$archive");
--
-- my $stdout_file = "$tmpdir/tar_content.out";
--
-- open(OLDOUT, ">&STDOUT") or clean_exit("could not dup STDOUT:
$!");
-- open(STDOUT, ">$stdout_file") or clean_exit("could not redirect
STDOUT: $!");
--
-- my $ret = system("tar", "tf", "$archive");
--
-- close(STDOUT);
-- open(STDOUT, ">&OLDOUT") or clean_exit("could not dup STDOUT: $!");
-- close(OLDOUT);
--
-- clean_exit("$url: could not list files in tar archive (is it
broken?)")
-- if ($ret);
--
-- open(TAR, "$stdout_file") or clean_exit("failed to open $stdout_file:
$!");
-- @tar_content = <TAR>;
-- close(TAR);
--
-- # use_external_bins=0
-- } else {
-- $tar = Archive::Tar->new($archive, 1);
-- clean_exit("$url: failed to read $archive (file transfer failed or ".
-- "file in URL not in tar'ed gzip format?).")
-- unless (defined($tar));
-- @tar_content = $tar->list_files();
-- }
--
-- # Make sure we could grab some content from the tarball.
-- clean_exit("$url: could not list files in tar archive (is it broken?)")
-- if ($#tar_content < 0);
--
-- # For each filename in the archive, do some basic sanity checks.
-- foreach my $filename (@tar_content) {
-- chomp($filename);
--
-- # We don't want absolute filename.
-- clean_exit("$url: rules archive contains absolute filename. ".
-- "Offending file/line:\n$filename")
-- if ($filename =~ /^\//);
--
-- # We don't want to have any weird characters anywhere in the filename.
-- clean_exit("$url: illegal character in filename in tar archive.
Allowed are ".
-- "$OK_PATH_CHARS\nOffending file/line:\n$filename")
-- if ($config{use_path_checks} && $filename =~ /[^$OK_PATH_CHARS]/);
--
-- # We don't want to unpack any "../../" junk (check is useless now
though).
-- clean_exit("$url: filename in tar archive contains \"..\".\n".
-- "Offending file/line:\n$filename")
-- if ($filename =~ /\.\./);
-- }
--
-- # Looks good. Now we can untar it.
-- print STDERR "Archive successfully downloaded, unpacking... "
-- unless ($config{quiet});
--
-+# start of IPCop change to use less disk space and unneed as we use md5
- if ($config{use_external_bins}) {
- clean_exit("failed to untar $archive.")
-- if system("tar", "xf", "$archive");
-- } else {
-- mkdir("$rules_dir") or clean_exit("could not create \"$rules_dir\"
directory: $!\n");
-- foreach my $file ($tar->list_files) {
-- next unless ($file =~ /^$rules_dir\/[^\/]+$/); # only
^rules/<file>$
--
-- my $content = $tar->get_content($file);
--
-- # Symlinks in the archive will make get_content return undef.
-- clean_exit("could not get content from file \"$file\" in
downloaded archive, ".
-- "make sure it is a regular file\n")
-- unless (defined($content));
--
-- open(RULEFILE, ">", "$file")
-- or clean_exit("could not open \"$file\" for writing: $!\n");
-- print RULEFILE $content;
-- close(RULEFILE);
-- }
-+ if system("/bin/tar xzf $archive --exclude='doc'
--exclude='so_rules'");
- }
-
- # Make sure that non-empty rules directory existed in archive.
This was sent by the SourceForge.net collaborative development platform, the
world's largest Open Source development site.
------------------------------------------------------------------------------
All the data continuously generated in your IT infrastructure
contains a definitive record of customers, application performance,
security threats, fraudulent activity, and more. Splunk takes this
data and makes sense of it. IT sense. And common sense.
http://p.sf.net/sfu/splunk-novd2d
_______________________________________________
Ipcop-svn mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/ipcop-svn