I've been reading stackexchange for years now, and always managed to find an answer to the problems I had to deal with. This time however, I'm going to ask a question as I could not find a solution to the issue I'm facing now.
I'm currently working on a research project, and need to create a sitemap for a website. I have more than 10GB of php files, downloaded from an online forum (now deceased). The files are all in one folder, and therefore I can't reconstruct the original structure based on folders.
I was wondering if there is a tool/script out there (I don't really care about the language it is written in) to analyse the php files and output some sort of sitemap/structure (XML, HTML, or else)
The files were created with phpBB; I'm using Debian and Ubuntu (two different PCs). Any suggestions are welcome. Many thanks in advance
wgetor similar)? Because that means they're HTML, and the.phpextension is irrelevant.libwww-perlaka LWP and DBI for the database inserts. I like to use HTML::TokeParser withLWPas a reasonably simple way of extracting data from HTML pages or files.