- Jun 2, 2009
- 488
- 13
- 81
I'm using Wget to download a 110MB XML product sheet from a vendor's server on a regular basis for inventory and other updates, but it keeps sanitizing the output file. Not a huge deal, but right now it's taking me ~15 mins every day to find/replace < and > back to < and >. If I call it from the browser it comes through unsanitized, but there's so much text it crashes Chrome when I try to copy-paste... and getting it through the browser isn't very amenable to the automation I'm working on anyway.
Is there a command to keep Wget from sanitizing the output file? And while I'm at it, I'm sure there's a command to specify the result file name and destination but I'm not finding it in the documentation.
The command I'm using is:
wget "url¶meters" HTTP/1.1
also tried wget -o, same result. I contacted the vendor about it and they said their files are all unsanitized and it's gotta be something on my end, but didn't get much more specific.
Is there a command to keep Wget from sanitizing the output file? And while I'm at it, I'm sure there's a command to specify the result file name and destination but I'm not finding it in the documentation.
The command I'm using is:
wget "url¶meters" HTTP/1.1
also tried wget -o, same result. I contacted the vendor about it and they said their files are all unsanitized and it's gotta be something on my end, but didn't get much more specific.