<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN">
<HTML><HEAD>
<META http-equiv=Content-Type content="text/html; charset=iso-8859-1">
<META content="MSHTML 6.00.2800.1400" name=GENERATOR></HEAD>
<BODY style="MARGIN-TOP: 2px; FONT: 10pt Book Antiqua; MARGIN-LEFT: 2px">
<DIV>The tool I've used in the past to map out an entire website is Visio
if you have access to a windows base pc.</DIV>
<DIV> </DIV>
<DIV>Jon L. Miller, MCNE, CNS, ASE<BR>Director/Sr Systems Consultant<BR>MMT
Networks Pty Ltd<BR><A
href="http://www.mmtnetworks.com.au">http://www.mmtnetworks.com.au</A></DIV>
<DIV> </DIV>
<DIV>"I don't know the key to success, but the key to failure<BR> is trying
to please everybody." -Bill Cosby</DIV>
<DIV> </DIV>
<DIV><BR><BR>>>> steveg@calm.wa.gov.au 4:39:32 pm 10/06/2004
>>><BR>Hi David,<BR><BR>A reliable site-mapping perl script I use is
tree.pl:<BR><A
href="http://www.danielnaber.de/tree/">http://www.danielnaber.de/tree/</A><BR><BR>HTH<BR>Steve<BR><BR>Quoting
David Buddrige <buddrige@wasp.net.au>:<BR><BR>> Hi all, <BR>>
<BR>> I have been asked to map out all the pages in a given intranet
website. So <BR>> for example, given website url: <BR>> <BR>> <A
href="http://abc.com/">http://abc.com/</A> <BR>> <BR>> They want a list of
every url that can be got at from the links on the <BR>> initial page, sort
of like this: <BR>> <BR>> <A
href="http://abc.com/">http://abc.com/</A><BR>> <A
href="http://abc.com/page1.html">http://abc.com/page1.html</A><BR>>
<A
href="http://abc.com/page2.html">http://abc.com/page2.html</A><BR>>
<A
href="http://abc.com/page2a.html">http://abc.com/page2a.html</A><BR>>
<A
href="http://abc.com/page2b.html">http://abc.com/page2b.html</A><BR>>
<A
href="http://abc.com/page2c.html">http://abc.com/page2c.html</A><BR>>
<A
href="http://abc.com/page3.html">http://abc.com/page3.html</A><BR>>
<A
href="http://abc.com/page4.html">http://abc.com/page4.html</A><BR>>
<A
href="http://abc.com/page4a.html">http://abc.com/page4a.html</A><BR>>
<A
href="http://abc.com/page4b.html">http://abc.com/page4b.html</A><BR>>
<A
href="http://abc.com/page4b1.html">http://abc.com/page4b1.html</A><BR>>
<A href="http://abc.com/page4c.html">http://abc.com/page4c.html</A> <BR>>
<BR>> And so on, mapping out the structure of links in the website. <BR>>
<BR>> It seemed to me that this ought to be something that is scriptable -
most <BR>> likely using wget or something... I have been experimenting with
wget, <BR>> however I have not been able to determine a way of just getting
the url's as<BR>> <BR>> opposed to actually downloading the entire page...
<BR>> <BR>> Does anyone know if wget can be used just to map out the tree
of url's in a <BR>> given website, as opposed to fully downloading and
mirroring the entire <BR>> website? <BR>> <BR>> I've been pouring over
the wget manual, but to no avail... is there a <BR>> similar command that is
more appropriate to what I am trying to do? <BR>> <BR>> thanks heaps guys
<BR>> <BR>> David. <BR>> <BR>>
_______________________________________________<BR>> PLUG discussion list:
plug@plug.linux.org.au<BR>> <A
href="http://mail.plug.linux.org.au/cgi-bin/mailman/listinfo/plug">http://mail.plug.linux.org.au/cgi-bin/mailman/listinfo/plug</A><BR>>
Committee e-mail: committee@plug.linux.org.au<BR>>
<BR><BR><BR><BR><BR><BR><BR>_______________________________________________<BR>PLUG
discussion list: plug@plug.linux.org.au<BR><A
href="http://mail.plug.linux.org.au/cgi-bin/mailman/listinfo/plug">http://mail.plug.linux.org.au/cgi-bin/mailman/listinfo/plug</A><BR>Committee
e-mail: committee@plug.linux.org.au<BR><BR></DIV></BODY></HTML>