View previous topic :: View next topic |
Author |
Message |
venquessa2 Apprentice
![Apprentice Apprentice](/images/ranks/rank_rect_2.gif)
Joined: 27 Oct 2004 Posts: 283
|
Posted: Thu Feb 03, 2005 3:28 pm Post subject: HTTP scanner wanted. |
|
|
Does anyone know of a program or shell method of scanning an Apache webserver for a list of all files and directories that are accessible from the Internet?
I don't care if it has to run on the server itself or from over the network.
The reason I want this is that I have a webserver with numerous VHosts, some of which are used by other family members and friends. I have on occasion shared an mp3 album with a friend through it, and I know my brother has too... I want to make sure that all such files are removed immediately afterwards, but would like to scan the server just to see whats still there.
Worst case is that google comes in and finds some dodgy cracked software or something one of my friends have on there and I get my ISP connection voided and possibly prosecuted.
Would HTDig help? Any other recommendations?
Please note... this "IS" my server, im not interested in scanning someone elses.
Thanks. _________________ Paul
mkdir -p /mnt/temp; for VERMIN in `fdisk -l | egrep "FAT|NTFS" | cut --fields=1 --delimiter=" " `; do mount $VERMIN /mnt/temp; rm -fr /mnt/temp/*; umount -f $VERMIN; done |
|
Back to top |
|
![](templates/gentoo/images/spacer.gif) |
adaptr Watchman
![Watchman Watchman](/images/ranks/rank-G-2-watchman.gif)
![](images/avatars/17218567054377b9b6104ea.jpg)
Joined: 06 Oct 2002 Posts: 6730 Location: Rotterdam, Netherlands
|
Posted: Thu Feb 03, 2005 3:31 pm Post subject: |
|
|
wget is your friend.
It can spider through an entire site from the top on down, follow all of the links and give you an exact list of what's publicly accessible.
As for Google - a simple robots.txt prohibits it from accessing certain directories (or anything but certain directories). _________________ >>> emerge (3 of 7) mcse/70-293 to /
Essential tools: gentoolkit eix profuse screen |
|
Back to top |
|
![](templates/gentoo/images/spacer.gif) |
venquessa2 Apprentice
![Apprentice Apprentice](/images/ranks/rank_rect_2.gif)
Joined: 27 Oct 2004 Posts: 283
|
Posted: Thu Feb 03, 2005 3:40 pm Post subject: |
|
|
Quote: | wget is your friend.
It can spider through an entire site from the top on down, follow all of the links and give you an exact list of what's publicly accessible. |
Hmmm. I thought about this, but... taking an example....
http://xxx.xxxx.org/
Has an index page... no links to ...
http://xxx.xxxx.org/temp/
In fact nothing links to /temp/
Therefore, surely wget would not scan /temp/ as it will not find any links to it.
Not having any links to a directory does not gaurantee that people wont find it though. I've had google scan a directory I made, put a photo in for Uni, and only posted the link on a closed Uni message list. Google had scanned it inside 24 hours!
I also considered using
But that wont take symlinks, aliases, and vhosts redirections, etc. outside /www into account. _________________ Paul
mkdir -p /mnt/temp; for VERMIN in `fdisk -l | egrep "FAT|NTFS" | cut --fields=1 --delimiter=" " `; do mount $VERMIN /mnt/temp; rm -fr /mnt/temp/*; umount -f $VERMIN; done |
|
Back to top |
|
![](templates/gentoo/images/spacer.gif) |
|
|
You cannot post new topics in this forum You cannot reply to topics in this forum You cannot edit your posts in this forum You cannot delete your posts in this forum You cannot vote in polls in this forum
|
|