From c8bb0015d0bee72d707686058e0585afdf3c5ffb Mon Sep 17 00:00:00 2001 From: Luke Shumaker Date: Sat, 7 Jan 2017 01:13:12 -0500 Subject: Do more efficient crawling. --- make | 8 +++++--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/make b/make index 844d28b..6e1fd47 100755 --- a/make +++ b/make @@ -6,9 +6,11 @@ require 'page_index' # Initialize the site generator Sitegen::init -# This should trigger a full recursive crawl, loadng everything into -# the sitegen. -IndexPage::new('src').local_input +# Load all pages +def crawl(page) + page.index_pages.select{|p|p.is_a?(IndexPage)}.each{|p|crawl(p)} +end +crawl(IndexPage::new('src')) # Tell the sitegen which files we want Sitegen.pages.each do |page| -- cgit v1.2.3-2-g168b