Hi guys.
I hopes you can help me with this task.
This website is very simple if you just look at it. It is used for finding specific files, just like Windows Explore etc.
In the left side there is a normal folder structure where everything is indexed. So if I go to some folder, like “Furniture” and then click “Chairs” and then click “Legs for chair” then I have x number of documents for chair legs in this folder. This particular page is in XML, so I have the file name, version, description etc. in a table, where I can click on the file I want, and then it downloads it to the harddrive (Probably through the javascript). This document is already on the hard drive in a “files” folder for the website, but the filenames are just numbers, like 4325436453 for example, so I don’t know what is what, and there are 5000 documents.
What I want is to extract all these files in their respective folders on a hard drive, instead of having to access the webpage all the time. This way I can load it to portable devices etc., which is good, because this webpage is inly compatible with IE 11 with MSXML and java etc. so I cannot just download the website to my Android or windows phone and make it work.
A secondary solution, would be to make some kind of bot or script that opens all the XML files, finds the description/title for the specific file names and renames the files, so that I will have both the document number and the description in the file name. This way I would be able to search for it.
The document number, version, file name, title etc. is in the same XML file, and if there are 3 documents available for download on a specific page, there will also be 3 document references in that XML file of course.
I thought about getting a script to to pair the document number and the title so that I can make some Excel magic and put it all into a -bat-file where I mass rename all the files in the folder.
Any suggestions?
Thank you very much.