From c2c1c41add0f4c3f95fbb6716fc9f8da9aa0e720 Mon Sep 17 00:00:00 2001
From: jsm28 <jsm28@138bc75d-0d04-0410-961f-82ee72b054a4>
Date: Sat, 3 Jan 2004 21:16:13 +0000
Subject: [PATCH] 	* update_web_docs: Create tarballs of documentation
 sources and 	HTML docs.

git-svn-id: svn+ssh://gcc.gnu.org/svn/gcc/trunk@75372 138bc75d-0d04-0410-961f-82ee72b054a4
---
 maintainer-scripts/ChangeLog       |  5 +++++
 maintainer-scripts/update_web_docs | 14 +++++++++-----
 2 files changed, 14 insertions(+), 5 deletions(-)

diff --git a/maintainer-scripts/ChangeLog b/maintainer-scripts/ChangeLog
index 4372b66e49bb..1537ad9da530 100644
--- a/maintainer-scripts/ChangeLog
+++ b/maintainer-scripts/ChangeLog
@@ -1,3 +1,8 @@
+2004-01-03  Joseph S. Myers  <jsm@polyomino.org.uk>
+
+	* update_web_docs: Create tarballs of documentation sources and
+	HTML docs.
+
 2004-01-02  Gerald Pfeifer  <gp@suse.de>
 
 	* update_version: Add some documentation.
diff --git a/maintainer-scripts/update_web_docs b/maintainer-scripts/update_web_docs
index 1cd10c228c64..998885c6935a 100755
--- a/maintainer-scripts/update_web_docs
+++ b/maintainer-scripts/update_web_docs
@@ -88,7 +88,10 @@ find $CVSROOT/gcc -name \*.texi,v -print | fgrep -v -f/home/gccadmin/scripts/doc
 
 
 # Checkout all the texi files.
-cvs -Q co -r$RELEASE `cat FILES` gcc/gcc/doc/install.texi2html gcc/gcc/doc/include/texinfo.tex
+cvs -Q export -r$RELEASE `cat FILES` gcc/gcc/doc/install.texi2html gcc/gcc/doc/include/texinfo.tex
+
+# Build a tarball of the sources.
+tar cf docs-sources.tar gcc
 
 # The directory to pass to -I; this is the one with texinfo.tex
 # and fdl.texi.
@@ -101,13 +104,14 @@ for file in $MANUALS; do
   filename=`find . -name ${file}.texi`
   if [ "${filename}" ]; then
     makeinfo --html -I ${includedir} -I `dirname ${filename}` ${filename}
+    tar cf ${file}-html.tar ${file}/*.html
     texi2dvi -I ${includedir} ${filename} </dev/null && dvips -o ${file}.ps ${file}.dvi
     mkdir -p $DOCSDIR/$file
   fi
 done
 
-# Then build a gzipped copy of each of the resulting .html and .ps files
-for file in */*.html *.ps; do
+# Then build a gzipped copy of each of the resulting .html, .ps and .tar files
+for file in */*.html *.ps *.tar; do
   cat $file | gzip --best > $file.gz
 done
 
@@ -121,8 +125,8 @@ if test $today = 15; then
   done
 fi
 
-# And copy the resulting html files to the web server
-for file in */*.html *.ps; do
+# And copy the resulting files to the web server
+for file in */*.html *.ps *.tar; do
   cat $DOCSDIR/$file | 
     sed -e '/^<meta name=generator/d' \
         -e '/^%DVIPSSource:/d' > file1
-- 
GitLab