X-Git-Url: http://git.tremily.us/?p=ikiwiki.git;a=blobdiff_plain;f=ikiwiki;h=ce89112de9fb02e263abd5804323bb645756bbb0;hp=cb8295cf9586ba76148541da7e69d5de2bf4da77;hb=e7900a6f1735bc922f2a25571bda7f9855e4825b;hpb=d5566303d6b416fb4b0f49a4a7eae2c81bddf17e diff --git a/ikiwiki b/ikiwiki index cb8295cf9..ce89112de 100755 --- a/ikiwiki +++ b/ikiwiki @@ -5,53 +5,75 @@ use strict; use File::Find; use Memoize; use File::Spec; - -$ENV{PATH}="/usr/local/bin:/usr/bin:/bin"; +use HTML::Template; BEGIN { $blosxom::version="is a proper perl module too much to ask?"; do "/usr/bin/markdown"; } -my ($srcdir, $destdir, %links, %oldlinks, %oldpagemtime, %renderedfiles, - %pagesources); -my $link=qr/\[\[([^\s]+)\]\]/; +$ENV{PATH}="/usr/local/bin:/usr/bin:/bin"; +my ($srcdir, $templatedir, $destdir, %links, %oldlinks, %oldpagemtime, + %renderedfiles, %pagesources); +my $wiki_link_regexp=qr/\[\[([^\s]+)\]\]/; +my $wiki_file_regexp=qr/(^[-A-Za-z0-9_.:\/+]+$)/; +my $wiki_file_prune_regexp=qr!((^|/).svn/|\.\.|^\.|\/\.|\.html?$)!; my $verbose=0; my $wikiname="wiki"; +my $default_pagetype=".mdwn"; +my $cgi=0; +my $url=""; +my $cgiurl=""; +my $historyurl=""; +my $svn=1; +my $anonok=0; + +sub usage { #{{{ + die "usage: ikiwiki [options] source templates dest\n"; +} #}}} + +sub error ($) { #{{{ + if ($cgi) { + print "Content-type: text/html\n\n"; + print "Error: @_\n"; + exit 1; + } + else { + die @_; + } +} #}}} -sub usage { - die "usage: ikiwiki [options] source dest\n"; -} - -sub error ($) { - die @_; -} - -sub debug ($) { +sub debug ($) { #{{{ print "@_\n" if $verbose; -} +} #}}} -sub mtime ($) { +sub mtime ($) { #{{{ my $page=shift; return (stat($page))[9]; -} +} #}}} + +sub possibly_foolish_untaint ($) { #{{{ + my $tainted=shift; + my ($untainted)=$tainted=~/(.*)/; + return $untainted; +} #}}} -sub basename { +sub basename ($) { #{{{ my $file=shift; $file=~s!.*/!!; return $file; -} +} #}}} -sub dirname { +sub dirname ($) { #{{{ my $file=shift; $file=~s!/?[^/]+$!!; return $file; -} +} #}}} -sub pagetype ($) { +sub pagetype ($) { #{{{ my $page=shift; if ($page =~ /\.mdwn$/) { @@ -60,38 +82,38 @@ sub pagetype ($) { else { return "unknown"; } -} +} #}}} -sub pagename ($) { +sub pagename ($) { #{{{ my $file=shift; my $type=pagetype($file); my $page=$file; $page=~s/\Q$type\E*$// unless $type eq 'unknown'; return $page; -} +} #}}} -sub htmlpage ($) { +sub htmlpage ($) { #{{{ my $page=shift; return $page.".html"; -} +} #}}} -sub readpage ($) { - my $page=shift; +sub readfile ($) { #{{{ + my $file=shift; local $/=undef; - open (PAGE, "$srcdir/$page") || error("failed to read $page: $!"); - my $ret=; - close PAGE; + open (IN, "$file") || error("failed to read $file: $!"); + my $ret=; + close IN; return $ret; -} +} #}}} -sub writepage ($$) { - my $page=shift; +sub writefile ($$) { #{{{ + my $file=shift; my $content=shift; - my $dir=dirname("$destdir/$page"); + my $dir=dirname($file); if (! -d $dir) { my $d=""; foreach my $s (split(m!/+!, $dir)) { @@ -102,26 +124,26 @@ sub writepage ($$) { } } - open (PAGE, ">$destdir/$page") || error("failed to write $page: $!"); - print PAGE $content; - close PAGE; -} + open (OUT, ">$file") || error("failed to write $file: $!"); + print OUT $content; + close OUT; +} #}}} -sub findlinks { +sub findlinks ($) { #{{{ my $content=shift; my @links; - while ($content =~ /$link/g) { + while ($content =~ /$wiki_link_regexp/g) { push @links, lc($1); } return @links; -} +} #}}} # Given a page and the text of a link on the page, determine which existing # page that link best points to. Prefers pages under a subdirectory with # the same name as the source page, failing that goes down the directory tree # to the base looking for matching pages. -sub bestlink ($$) { +sub bestlink ($$) { #{{{ my $page=shift; my $link=lc(shift); @@ -139,17 +161,18 @@ sub bestlink ($$) { #print STDERR "warning: page $page, broken link: $link\n"; return ""; -} +} #}}} -sub isinlinableimage ($) { +sub isinlinableimage ($) { #{{{ my $file=shift; $file=~/\.(png|gif|jpg|jpeg)$/; -} +} #}}} -sub htmllink ($$) { +sub htmllink { #{{{ my $page=shift; my $link=shift; + my $noimagelink=shift; my $bestlink=bestlink($page, $link); @@ -163,27 +186,27 @@ sub htmllink ($$) { $bestlink=htmlpage($bestlink); } if (! grep { $_ eq $bestlink } values %renderedfiles) { - return "?$link" + return "?$link" } $bestlink=File::Spec->abs2rel($bestlink, dirname($page)); - if (isinlinableimage($bestlink)) { + if (! $noimagelink && isinlinableimage($bestlink)) { return ""; } return "$link"; -} +} #}}} -sub linkify ($$) { +sub linkify ($$) { #{{{ my $content=shift; my $file=shift; - $content =~ s/$link/htmllink(pagename($file), $1)/eg; + $content =~ s/$wiki_link_regexp/htmllink(pagename($file), $1)/eg; return $content; -} +} #}}} -sub htmlize ($$) { +sub htmlize ($$) { #{{{ my $type=shift; my $content=shift; @@ -193,10 +216,9 @@ sub htmlize ($$) { else { error("htmlization of $type not supported"); } -} +} #}}} -sub linkbacks ($$) { - my $content=shift; +sub backlinks ($) { #{{{ my $page=shift; my @links; @@ -214,48 +236,74 @@ sub linkbacks ($$) { $p_trimmed=~s/^\Q$dir\E// && $page_trimmed=~s/^\Q$dir\E//; - push @links, "$p_trimmed"; + push @links, { url => $href, page => $p_trimmed }; } } - $content.="

Links: ".join(" ", sort @links)."

\n" if @links; - return $content; -} - -sub finalize ($$) { - my $content=shift; + return @links; +} #}}} + +sub parentlinks ($) { #{{{ my $page=shift; - - my $title=basename($page); - $title=~s/_/ /g; + my @ret; my $pagelink=""; my $path=""; + my $skip=1; foreach my $dir (reverse split("/", $page)) { - if (length($pagelink)) { - $pagelink="$dir/ $pagelink"; + if (! $skip) { + unshift @ret, { url => "$path$dir.html", page => $dir }; } else { - $pagelink=$dir; + $skip=0; } $path.="../"; } - $path=~s/\.\.\/$/index.html/; - $pagelink="$wikiname/ $pagelink"; + return @ret; +} #}}} + +sub indexlink () { #{{{ + return "$wikiname/ "; +} #}}} - $content="\n$title\n\n". - "

$pagelink

\n". - $content. - "\n\n"; +sub finalize ($$) { #{{{ + my $content=shift; + my $page=shift; + + my $title=basename($page); + $title=~s/_/ /g; - return $content; -} + my $template=HTML::Template->new(blind_cache => 1, + filename => "$templatedir/page.tmpl"); + + if (length $cgiurl) { + $template->param(editurl => "$cgiurl?do=edit&page=$page"); + $template->param(recentchangesurl => "$cgiurl?do=recentchanges"); + } -sub render ($) { + if (length $historyurl) { + my $u=$historyurl; + $u=~s/\[\[\]\]/$pagesources{$page}/g; + $template->param(historyurl => $u); + } + + $template->param( + title => $title, + indexlink => $url, + wikiname => $wikiname, + parentlinks => [parentlinks($page)], + content => $content, + backlinks => [backlinks($page)], + ); + + return $template->output; +} #}}} + +sub render ($) { #{{{ my $file=shift; my $type=pagetype($file); - my $content=readpage($file); + my $content=readfile("$srcdir/$file"); if ($type ne 'unknown') { my $page=pagename($file); @@ -263,25 +311,24 @@ sub render ($) { $content=linkify($content, $file); $content=htmlize($type, $content); - $content=linkbacks($content, $page); $content=finalize($content, $page); - writepage(htmlpage($page), $content); + writefile("$destdir/".htmlpage($page), $content); $oldpagemtime{$page}=time; $renderedfiles{$page}=htmlpage($page); } else { $links{$file}=[]; - writepage($file, $content); + writefile("$destdir/$file", $content); $oldpagemtime{$file}=time; $renderedfiles{$file}=$file; } -} +} #}}} -sub loadindex () { - open (IN, "$srcdir/.index") || return; +sub loadindex () { #{{{ + open (IN, "$srcdir/.ikiwiki/index") || return; while () { - ($_)=/(.*)/; # untaint + $_=possibly_foolish_untaint($_); chomp; my ($mtime, $file, $rendered, @links)=split(' ', $_); my $page=pagename($file); @@ -292,19 +339,112 @@ sub loadindex () { $renderedfiles{$page}=$rendered; } close IN; -} +} #}}} -sub saveindex () { - open (OUT, ">$srcdir/.index") || error("cannot write to .index: $!"); +sub saveindex () { #{{{ + if (! -d "$srcdir/.ikiwiki") { + mkdir("$srcdir/.ikiwiki"); + } + open (OUT, ">$srcdir/.ikiwiki/index") || error("cannot write to index: $!"); foreach my $page (keys %oldpagemtime) { - print OUT "$oldpagemtime{$page} $pagesources{$page} $renderedfiles{$page} ". - join(" ", @{$links{$page}})."\n" - if $oldpagemtime{$page}; + print OUT "$oldpagemtime{$page} $pagesources{$page} $renderedfiles{$page} ". + join(" ", @{$links{$page}})."\n" + if $oldpagemtime{$page}; } close OUT; -} +} #}}} + +sub rcs_update () { #{{{ + if (-d "$srcdir/.svn") { + if (system("svn", "update", "--quiet", $srcdir) != 0) { + warn("svn update failed\n"); + } + } +} #}}} + +sub rcs_commit ($) { #{{{ + my $message=shift; + + if (-d "$srcdir/.svn") { + if (system("svn", "commit", "--quiet", "-m", + possibly_foolish_untaint($message), $srcdir) != 0) { + warn("svn commit failed\n"); + } + } +} #}}} + +sub rcs_add ($) { #{{{ + my $file=shift; + + if (-d "$srcdir/.svn") { + my $parent=dirname($file); + while (! -d "$srcdir/$parent/.svn") { + $file=$parent; + $parent=dirname($file); + } + + if (system("svn", "add", "--quiet", "$srcdir/$file") != 0) { + warn("svn add failed\n"); + } + } +} #}}} + +sub rcs_recentchanges ($) { #{{{ + my $num=shift; + my @ret; + + eval q{use Date::Parse}; + eval q{use Time::Duration}; + + if (-d "$srcdir/.svn") { + my $info=`LANG=C svn info $srcdir`; + my ($svn_url)=$info=~/^URL: (.*)$/m; + + # FIXME: currently assumes that the wiki is somewhere + # under trunk in svn, doesn't support other layouts. + my ($svn_base)=$svn_url=~m!(/trunk(?:/.*)?)$!; + + my $div=qr/^--------------------+$/; + my $infoline=qr/^r(\d+)\s+\|\s+([^\s]+)\s+\|\s+(\d+-\d+-\d+\s+\d+:\d+:\d+\s+[-+]?\d+).*/; + my $state='start'; + my ($rev, $user, $when, @pages, @message); + foreach (`LANG=C svn log -v '$svn_url'`) { + chomp; + if ($state eq 'start' && /$div/) { + $state='header'; + } + elsif ($state eq 'header' && /$infoline/) { + $rev=$1; + $user=$2; + $when=concise(ago(time - str2time($3))); + } + elsif ($state eq 'header' && /^\s+[A-Z]\s+\Q$svn_base\E\/(.+)$/) { + push @pages, { link => htmllink("", pagename($1), 1) } + if length $1; + } + elsif ($state eq 'header' && /^$/) { + $state='body'; + } + elsif ($state eq 'body' && /$div/) { + push @ret, { rev => $rev, user => $user, + when => $when, message => [@message], + pages => [@pages] } if @pages; + return @ret if @ret >= $num; + + $state='header'; + $rev=$user=$when=undef; + @pages=@message=(); + } + elsif ($state eq 'body') { + push @message, {line => $_}, + } + } + } + + return @ret; +} #}}} -sub prune ($) { +sub prune ($) { #{{{ my $file=shift; unlink($file); @@ -312,20 +452,20 @@ sub prune ($) { while (rmdir($dir)) { $dir=dirname($dir); } -} +} #}}} -sub refresh () { +sub refresh () { #{{{ # Find existing pages. my %exists; my @files; find({ no_chdir => 1, wanted => sub { - if (/\/\.svn\//) { + if (/$wiki_file_prune_regexp/) { $File::Find::prune=1; } - elsif (! -d $_ && ! /\.html$/ && ! /\/\./) { - my ($f)=/(^[-A-Za-z0-9_.:\/+]+$)/; # untaint + elsif (! -d $_) { + my ($f)=/$wiki_file_regexp/; # untaint if (! defined $f) { warn("skipping bad filename $_\n"); } @@ -397,10 +537,10 @@ FILE: foreach my $file (@files) { } } - # handle linkbacks; if a page has added/removed links, update the + # handle backlinks; if a page has added/removed links, update the # pages it links to # TODO: inefficient; pages may get rendered above and again here; - # problem is the linkbacks could be wrong in the first pass render + # problem is the backlinks could be wrong in the first pass render # above if (%rendered) { my %linkchanged; @@ -430,41 +570,81 @@ FILE: foreach my $file (@files) { foreach my $link (keys %linkchanged) { my $linkfile=$pagesources{$link}; if (defined $linkfile) { - debug("rendering $linkfile, to update its linkbacks"); + debug("rendering $linkfile, to update its backlinks"); render($linkfile); } } } -} +} #}}} # Generates a C wrapper program for running ikiwiki in a specific way. # The wrapper may be safely made suid. -sub gen_wrapper ($$) { - my ($offline, $rebuild)=@_; +sub gen_wrapper ($$) { #{{{ + my ($svn, $rebuild)=@_; - eval {use Cwd 'abs_path'}; + eval q{use Cwd 'abs_path'}; $srcdir=abs_path($srcdir); $destdir=abs_path($destdir); my $this=abs_path($0); if (! -x $this) { error("$this doesn't seem to be executable"); } + + my @params=($srcdir, $templatedir, $destdir, "--wikiname=$wikiname"); + push @params, "--verbose" if $verbose; + push @params, "--rebuild" if $rebuild; + push @params, "--nosvn" if !$svn; + push @params, "--cgi" if $cgi; + push @params, "--url=$url" if $url; + push @params, "--cgiurl=$cgiurl" if $cgiurl; + push @params, "--historyurl=$historyurl" if $historyurl; + push @params, "--anonok" if $anonok; + my $params=join(" ", @params); + my $call=''; + foreach my $p ($this, $this, @params) { + $call.=qq{"$p", }; + } + $call.="NULL"; - my $call=qq{"$this", "$this", "$srcdir", "$destdir", "--wikiname=$wikiname"}; - $call.=', "--verbose"' if $verbose; - $call.=', "--rebuild"' if $rebuild; - $call.=', "--offline"' if $offline; + my @envsave; + push @envsave, qw{REMOTE_ADDR QUERY_STRING REQUEST_METHOD REQUEST_URI + CONTENT_TYPE CONTENT_LENGTH GATEWAY_INTERFACE + HTTP_COOKIE} if $cgi; + my $envsave=""; + foreach my $var (@envsave) { + $envsave.=<<"EOF" + if ((s=getenv("$var"))) + asprintf(&newenviron[i++], "%s=%s", "$var", s); +EOF + } open(OUT, ">ikiwiki-wrap.c") || error("failed to write ikiwiki-wrap.c: $!");; print OUT <<"EOF"; -/* A suid wraper for ikiwiki */ +/* A wrapper for ikiwiki, can be safely made suid. */ +#define _GNU_SOURCE #include #include #include - -int main (void) { - clearenv(); - execl($call, NULL); +#include + +extern char **environ; + +int main (int argc, char **argv) { + /* Sanitize environment. */ + char *s; + char *newenviron[$#envsave+3]; + int i=0; +$envsave + newenviron[i++]="HOME=$ENV{HOME}"; + newenviron[i]=NULL; + environ=newenviron; + + if (argc == 2 && strcmp(argv[1], "--params") == 0) { + printf("$params\\n"); + exit(0); + } + + execl($call); perror("failed to run $this"); exit(1); } @@ -476,37 +656,309 @@ EOF unlink("ikiwiki-wrap.c"); print "successfully generated ikiwiki-wrap\n"; exit 0; -} +} #}}} -sub update () { - if (-d "$srcdir/.svn") { - if (system("svn", "update", "--quiet", $srcdir) != 0) { - warn("svn update failed\n"); +sub cgi_recentchanges ($) { #{{{ + my $q=shift; + + my $template=HTML::Template->new( + filename => "$templatedir/recentchanges.tmpl"); + $template->param( + title => "RecentChanges", + indexlink => $url, + wikiname => $wikiname, + changelog => [rcs_recentchanges(100)], + ); + print $q->header, $template->output; +} #}}} + +sub cgi_signin ($$) { #{{{ + my $q=shift; + my $session=shift; + + eval q{use CGI::FormBuilder}; + my $form = CGI::FormBuilder->new( + title => "$wikiname signin", + fields => [qw(do page name password confirm_password email)], + header => 1, + method => 'POST', + validate => { + name => '/^\w+$/', + confirm_password => { + perl => q{eq $form->field("password")}, + }, + email => 'EMAIL', + }, + required => 'NONE', + javascript => 0, + params => $q, + action => $q->request_uri, + template => (-e "$templatedir/signin.tmpl" ? "$templatedir/signin.tmpl" : "") + ); + + $form->sessionid($session->id); + $form->field(name => "name", required => 0); + $form->field(name => "do", type => "hidden"); + $form->field(name => "page", type => "hidden"); + $form->field(name => "password", type => "password", required => 0); + $form->field(name => "confirm_password", type => "password", required => 0); + $form->field(name => "email", required => 0); + if ($session->param("name")) { + $form->field(name => "name", value => $session->param("name")); + } + if ($q->param("do") ne "signin") { + $form->text("You need to log in before you can edit pages."); + } + + if ($form->submitted) { + # Set required fields based on how form was submitted. + my %required=( + "Login" => [qw(name password)], + "Register" => [qw(name password confirm_password email)], + "Mail Password" => [qw(name)], + ); + foreach my $opt (@{$required{$form->submitted}}) { + $form->field(name => $opt, required => 1); + } + + # Validate password differently depending on how form was + # submitted. + if ($form->submitted eq 'Login') { + $form->field( + name => "password", + validate => sub { + # TODO get real user password + shift eq "foo"; + }, + ); + } + else { + $form->field(name => "password", validate => 'VALUE'); } } -} + else { + # Comments only shown first time. + $form->field(name => "name", comment => "use FirstnameLastName"); + $form->field(name => "confirm_password", comment => "(only needed"); + $form->field(name => "email", comment => "for registration)"); + } + + if ($form->submitted && $form->validate) { + if ($form->submitted eq 'Login') { + $session->param("name", $form->field("name")); + if (defined $form->field("do") && + $form->field("do") ne 'signin') { + print $q->redirect( + "$cgiurl?do=".$form->field("do"). + "&page=".$form->field("page")); + } + else { + print $q->redirect($url); + } + } + elsif ($form->submitted eq 'Register') { + # TODO: save registration info + $form->field(name => "confirm_password", type => "hidden"); + $form->field(name => "email", type => "hidden"); + $form->text("Registration successful. Now you can Login."); + print $form->render(submit => ["Login"]);; + } + elsif ($form->submitted eq 'Mail Password') { + # TODO mail password + $form->text("Your password has been emailed to you."); + print $form->render(submit => ["Login", "Register", "Mail Password"]);; + } + } + else { + print $form->render(submit => ["Login", "Register", "Mail Password"]);; + } +} #}}} + +sub cgi () { #{{{ + eval q{use CGI}; + eval q{use CGI::Session}; + + my $q=CGI->new; + + my $do=$q->param('do'); + if (! defined $do || ! length $do) { + error("\"do\" parameter missing"); + } + + # This does not need a session. + if ($do eq 'recentchanges') { + cgi_recentchanges($q); + return; + } + + # session id has to be _sessionid for CGI::FormBuilder to work. + # TODO: stop having the formbuilder emit cookies and change session + # id to something else. + CGI::Session->name("_sessionid"); + my $session = CGI::Session->new(undef, $q, + { Directory=> "$srcdir/.ikiwiki/sessions" }); + + # Everything below this point needs the user to be signed in. + if ((! $anonok && ! defined $session->param("name")) || $do eq 'signin') { + cgi_signin($q, $session); + return; + } + + my ($page)=$q->param('page')=~/$wiki_file_regexp/; + if (! defined $page || ! length $page || $page ne $q->param('page') || + $page=~/$wiki_file_prune_regexp/ || $page=~/^\//) { + error("bad page name"); + } + $page=lc($page); + + my $action=$q->request_uri; + $action=~s/\?.*//; + + if ($do eq 'create') { + if (exists $pagesources{lc($page)}) { + # hmm, someone else made the page in the meantime? + print $q->redirect("$url/".htmlpage($page)); + } + + my @page_locs; + my ($from)=$q->param('from')=~/$wiki_file_regexp/; + if (! defined $from || ! length $from || + $from ne $q->param('from') || + $from=~/$wiki_file_prune_regexp/ || $from=~/^\//) { + @page_locs=$page; + } + else { + my $dir=$from."/"; + $dir=~s![^/]+/$!!; + push @page_locs, $dir.$page; + push @page_locs, "$from/$page"; + while (length $dir) { + $dir=~s![^/]+/$!!; + push @page_locs, $dir.$page; + } + } + + $q->param("do", "save"); + print $q->header, + $q->start_html("Creating $page"), + $q->h1(indexlink()." Creating $page"), + $q->start_form(-action => $action), + $q->hidden('do'), + "Select page location:", + $q->popup_menu('page', \@page_locs), + $q->textarea(-name => 'content', + -default => "", + -rows => 20, + -columns => 80), + $q->br, + "Optional comment about this change:", + $q->br, + $q->textfield(-name => "comments", -size => 80), + $q->br, + $q->submit("Save Page"), + $q->end_form, + $q->end_html; + } + elsif ($do eq 'edit') { + my $content=""; + if (exists $pagesources{lc($page)}) { + $content=readfile("$srcdir/$pagesources{lc($page)}"); + $content=~s/\n/\r\n/g; + } + $q->param("do", "save"); + print $q->header, + $q->start_html("Editing $page"), + $q->h1(indexlink()." Editing $page"), + $q->start_form(-action => $action), + $q->hidden('do'), + $q->hidden('page'), + $q->textarea(-name => 'content', + -default => $content, + -rows => 20, + -columns => 80), + $q->br, + "Optional comment about this change:", + $q->br, + $q->textfield(-name => "comments", -size => 80), + $q->br, + $q->submit("Save Page"), + $q->end_form, + $q->end_html; + } + elsif ($do eq 'save') { + my $file=$page.$default_pagetype; + my $newfile=1; + if (exists $pagesources{lc($page)}) { + $file=$pagesources{lc($page)}; + $newfile=0; + } + + my $content=$q->param('content'); + $content=~s/\r\n/\n/g; + $content=~s/\r/\n/g; + writefile("$srcdir/$file", $content); + + my $message="web commit from $ENV{REMOTE_ADDR}"; + if (defined $q->param('comments')) { + $message.=": ".$q->param('comments'); + } + + if ($svn) { + if ($newfile) { + rcs_add($file); + } + # presumably the commit will trigger an update + # of the wiki + rcs_commit($message); + } + else { + refresh(); + } + + print $q->redirect("$url/".htmlpage($page)); + } + else { + error("unknown do parameter"); + } +} #}}} +# main {{{ my $rebuild=0; -my $offline=0; -my $gen_wrapper=0; +my $wrapper=0; if (grep /^-/, @ARGV) { eval {use Getopt::Long}; GetOptions( "wikiname=s" => \$wikiname, "verbose|v" => \$verbose, "rebuild" => \$rebuild, - "gen-wrapper" => \$gen_wrapper, - "offline" => \$offline, + "wrapper" => \$wrapper, + "svn!" => \$svn, + "anonok!" => \$anonok, + "cgi" => \$cgi, + "url=s" => \$url, + "cgiurl=s" => \$cgiurl, + "historyurl=s" => \$historyurl, ) || usage(); } -usage() unless @ARGV == 2; -($srcdir) = shift =~ /(.*)/; # untaint -($destdir) = shift =~ /(.*)/; # untaint +usage() unless @ARGV == 3; +($srcdir) = possibly_foolish_untaint(shift); +($templatedir) = possibly_foolish_untaint(shift); +($destdir) = possibly_foolish_untaint(shift); + +if ($cgi && ! length $url) { + error("Must specify url to wiki with --url when using --cgi"); +} -gen_wrapper($offline, $rebuild) if $gen_wrapper; +gen_wrapper($svn, $rebuild) if $wrapper; memoize('pagename'); memoize('bestlink'); -update() unless $offline; loadindex() unless $rebuild; -refresh(); -saveindex(); +if ($cgi) { + cgi(); +} +else { + rcs_update() if $svn; + refresh(); + saveindex(); +} +#}}}