Subversion Repositories svn LFS-FR

Rev

Rev 5432 | Blame | Compare with Previous | Last modification | View Log | RSS feed

# Makefile for BLFS Book generation.
# By Tushar Teredesai <tushar@linuxfromscratch.org>
# 2004-01-31
# $LastChangedBy: randy $
# $Date: 2012-08-22 18:33:50 +0200 (mer. 22 août 2012) $

# Adjust these to suit your installation
BASEDIR ?= $(HOME)/public_html/blfs-book-xsl
DUMPDIR ?= $(HOME)/blfs-commands
RENDERTMP ?= tmp
CHUNK_QUIET = 1
ROOT_ID =
PDF_OUTPUT = BLFS-BOOK.pdf
NOCHUNKS_OUTPUT = BLFS-BOOK.html
SHELL = /bin/bash

ALLXML := $(filter-out $(RENDERTMP)/%, \
        $(wildcard *.xml */*.xml */*/*.xml */*/*/*.xml */*/*/*/*.xml))
ALLXSL := $(filter-out $(RENDERTMP)/%, \
        $(wildcard *.xsl */*.xsl */*/*.xsl */*/*/*.xsl */*/*/*/*.xsl))

ifdef V
  Q =
else
  Q = @
endif

blfs: html wget-list
all: blfs nochunks pdf
#all: blfs nochunks
world: all blfs-patch-list dump-commands test-links

html: $(BASEDIR)/index.html
$(BASEDIR)/index.html: $(RENDERTMP)/blfs-html.xml
        @echo "Generating chunked XHTML files..."
        $(Q)xsltproc --nonet -stringparam chunk.quietly $(CHUNK_QUIET) \
          -stringparam rootid "$(ROOT_ID)" -stringparam base.dir $(BASEDIR)/ \
          stylesheets/blfs-chunked.xsl $(RENDERTMP)/blfs-html.xml

        @echo "Copying CSS code and images..."
        $(Q)if [ ! -e $(BASEDIR)/stylesheets ]; then \
          mkdir -p $(BASEDIR)/stylesheets; \
        fi;
        $(Q)cp stylesheets/lfs-xsl/*.css $(BASEDIR)/stylesheets
        $(Q)if [ ! -e $(BASEDIR)/images ]; then \
          mkdir -p $(BASEDIR)/images; \
        fi;
        $(Q)cp images/*.png $(BASEDIR)/images
        $(Q)cd $(BASEDIR)/; sed -i -e "s@../stylesheets@stylesheets@g" *.html
        $(Q)cd $(BASEDIR)/; sed -i -e "s@../images@images@g" *.html

        @echo "Running Tidy and obfuscate.sh on chunked XHTML..."
        $(Q)for filename in `find $(BASEDIR) -name "*.html"`; do \
          tidy -config tidy.conf $$filename; \
          true; \
          bash obfuscate.sh $$filename; \
          sed -i -e "s@text/html@application/xhtml+xml@g" $$filename; \
        done;

pdf: $(BASEDIR)/$(PDF_OUTPUT)
$(RENDERTMP)/blfs-pdf.xml: $(RENDERTMP)/blfs-full.xml
        @echo "Generating profiled XML for PDF..."
        $(Q)xsltproc --nonet --stringparam profile.condition pdf \
        --output $(RENDERTMP)/blfs-pdf.xml stylesheets/lfs-xsl/profile.xsl \
        $(RENDERTMP)/blfs-full.xml

$(RENDERTMP)/blfs-pdf.fo: $(RENDERTMP)/blfs-pdf.xml
        @echo "Generating FO file..."
        $(Q)xsltproc --nonet -stringparam rootid "$(ROOT_ID)" \
          --output $(RENDERTMP)/blfs-pdf.fo stylesheets/blfs-pdf.xsl \
          $(RENDERTMP)/blfs-pdf.xml
        $(Q)sed -i -e 's/span="inherit"/span="all"/' $(RENDERTMP)/blfs-pdf.fo

$(BASEDIR)/$(PDF_OUTPUT): $(RENDERTMP)/blfs-pdf.fo
        @echo "Generating PDF file..."
        $(Q)if [ ! -e $(BASEDIR) ]; then \
          mkdir -p $(BASEDIR); \
        fi;
        $(Q)fop $(RENDERTMP)/blfs-pdf.fo $(BASEDIR)/$(PDF_OUTPUT)

nochunks: $(BASEDIR)/$(NOCHUNKS_OUTPUT)
$(BASEDIR)/$(NOCHUNKS_OUTPUT): $(RENDERTMP)/blfs-html.xml
        @echo "Generating non-chunked XHTML file..."
        $(Q)xsltproc --nonet -stringparam rootid "$(ROOT_ID)" \
          --output $(BASEDIR)/$(NOCHUNKS_OUTPUT) \
          stylesheets/blfs-nochunks.xsl $(RENDERTMP)/blfs-html.xml

        @echo "Running Tidy and obfuscate.sh on non-chunked XHTML..."
        $(Q)tidy -config tidy.conf $(BASEDIR)/$(NOCHUNKS_OUTPUT) || true
        $(Q)bash obfuscate.sh $(BASEDIR)/$(NOCHUNKS_OUTPUT)
        $(Q)sed -i -e "s@text/html@application/xhtml+xml@g" \
          $(BASEDIR)/$(NOCHUNKS_OUTPUT)

tmpdir: $(RENDERTMP)
$(RENDERTMP):
        @echo "Creating $(RENDERTMP)"
        $(Q)[ -d $(RENDERTMP) ] || mkdir -p $(RENDERTMP)

clean:
        @echo "Cleaning $(RENDERTMP)"
        $(Q)rm -f $(RENDERTMP)/blfs-{full,html}.xml
#       $(Q)rm -f $(RENDERTMP)/blfs-{full,html,pdf}.xml
#       $(Q)rm -f $(RENDERTMP)/blfs-pdf.fo
        $(Q)rm -f $(RENDERTMP)/blfs-{patch-list,patches}
        $(Q)rmdir $(RENDERTMP) 2>/dev/null || :

validxml: $(RENDERTMP)/blfs-full.xml
$(RENDERTMP)/blfs-full.xml: general.ent $(ALLXML) $(ALLXSL)
        @echo "Validating the book..."
        $(Q)[ -d $(RENDERTMP) ] || mkdir -p $(RENDERTMP)
        $(Q)xmllint --nonet --noent --xinclude --postvalid \
          -o $(RENDERTMP)/blfs-full.xml index.xml

profile-html: $(RENDERTMP)/blfs-html.xml
$(RENDERTMP)/blfs-html.xml: $(RENDERTMP)/blfs-full.xml
        @echo "Generating profiled XML for XHTML..."
        $(Q)xsltproc --nonet --stringparam profile.condition html \
          --output $(RENDERTMP)/blfs-html.xml stylesheets/lfs-xsl/profile.xsl \
          $(RENDERTMP)/blfs-full.xml

blfs-patch-list: blfs-patches.sh
        @echo "Generating blfs patch list..."
        $(Q)awk '{if ($$1 == "copy") {sub(/.*\//, "", $$2); print $$2}}' \
          blfs-patches.sh > blfs-patch-list

blfs-patches.sh: $(RENDERTMP)/blfs-full.xml
        @echo "Generating blfs patch script..."
        $(Q)xsltproc --nonet --output blfs-patches.sh \
          stylesheets/patcheslist.xsl $(RENDERTMP)/blfs-full.xml

wget-list: $(BASEDIR)/wget-list
$(BASEDIR)/wget-list: $(RENDERTMP)/blfs-full.xml
        @echo "Generating wget list..."
        $(Q)mkdir -p $(BASEDIR)
        $(Q)xsltproc --nonet --output $(BASEDIR)/wget-list \
          stylesheets/wget-list.xsl $(RENDERTMP)/blfs-full.xml

test-links: $(BASEDIR)/test-links
$(BASEDIR)/test-links: $(RENDERTMP)/blfs-full.xml
        @echo "Generating test-links file..."
        $(Q)mkdir -p $(BASEDIR)
        $(Q)xsltproc --nonet --stringparam list_mode full \
          --output $(BASEDIR)/test-links stylesheets/wget-list.xsl \
          $(RENDERTMP)/blfs-full.xml

        @echo "Checking URLs, first pass..."
        $(Q)rm -f $(BASEDIR)/{good,bad,true_bad}_urls
        $(Q)for URL in `cat $(BASEDIR)/test-links`; do \
            wget --spider --tries=2 --timeout=60 $$URL >>/dev/null 2>&1; \
            if test $$? -ne 0 ; then echo $$URL >> $(BASEDIR)/bad_urls ; \
            else echo $$URL >> $(BASEDIR)/good_urls 2>&1; \
            fi; \
        done

        @echo "Checking URLs, second pass..."
        $(Q)for URL2 in `cat $(BASEDIR)/bad_urls`; do \
            wget --spider --tries=2 --timeout=60 $$URL2 >>/dev/null 2>&1; \
            if test $$? -ne 0 ; then echo $$URL2 >> $(BASEDIR)/true_bad_urls ; \
            else echo $$URL2 >> $(BASEDIR)/good_urls 2>&1; \
            fi; \
        done
        
test-all-links: $(BASEDIR)/test-all-links
$(BASEDIR)/test-all-links: $(RENDERTMP)/blfs-full.xml
        @echo "Generating test-all-links file..."
        $(Q)mkdir -p $(BASEDIR)
        $(Q)xsltproc --nonet \
          --output $(BASEDIR)/test-all-links stylesheets/list_url.xsl \
          $(RENDERTMP)/blfs-full.xml

        @echo "Checking URLs, first pass..."
        $(Q)rm -f $(BASEDIR)/{good,bad,true_bad,broken}_urls
        $(Q)for URL in `uniq $(BASEDIR)/test-all-links`; do \
            wget --spider --tries=2 --timeout=60 $$URL >>/dev/null 2>&1; \
            if test $$? -ne 0 ; then echo $$URL >> $(BASEDIR)/bad_urls ; \
            else echo $$URL >> $(BASEDIR)/good_urls 2>&1; \
            fi; \
        done

        @echo "Checking URLs, second pass..."
        $(Q)for URL2 in `cat $(BASEDIR)/bad_urls`; do \
            wget --spider --tries=2 --timeout=60 $$URL2 >>/dev/null 2>&1; \
            if test $$? -ne 0 ; then echo $$URL2 >> $(BASEDIR)/true_bad_urls ; \
            else echo $$URL2 >> $(BASEDIR)/good_urls 2>&1; \
            fi; \
        done
        
        @echo "Generating broken_urls file ..."
        $(Q)for URL in `cat $(BASEDIR)/true_bad_urls`; do \
                grep -r "$$URL" * >> $(BASEDIR)/broken_urls ; \
        done
        
        $(Q)sed -i $(BASEDIR)/broken_urls -e "/^tmp/d" -e "/xml~/d"


dump-commands: $(DUMPDIR)
$(DUMPDIR): $(RENDERTMP)/blfs-full.xml
        @echo "Dumping book commands..."
        $(Q)xsltproc --output $(DUMPDIR)/ \
           stylesheets/dump-commands.xsl $(RENDERTMP)/blfs-full.xml
        $(Q)touch $(DUMPDIR)

validate:
        @echo "Validating the book..."
        $(Q)xmllint --noout --nonet --xinclude --postvalid index.xml

.PHONY: blfs all world html nochunks tmpdir clean validxml \
        profile-html wget-list test-links dump-commands validate

#.PHONY: blfs all world html pdf nochunks tmpdir clean validxml \
        profile-html wget-list test-links dump-commands validate