I'm no web expert, but two things to check--both of which depend on the web server:
1. Does the web server offer different pages based on the user-agent string supplied by the request? firefox and wget certainly provide different user-agents by default. I believe wget can be told to supply a different user-agent on the command line
2. Does the web server offer different pages for (ro)bots? As far as I understand it, wget honors site restrictions for robots. And as far as I know, that adherence is fixed--
not configurable via command line.
EDIT:
I just tried running the following command:
Code:
wget --user-agent="Mozilla/5.0 (Windows NT 6.2; rv:9.0.1) Gecko/20100101 Firefox/9.0.1" http://www.discoverygc.com/server/players_online.html"
The transfer halted after about 1 or 2 KB. I then ran:
Code:
wget -c --user-agent="Mozilla/5.0 (Windows NT 6.2; rv:9.0.1) Gecko/20100101 Firefox/9.0.1" http://www.discoverygc.com/server/players_online.html"
The wget output indicated that the second command did, indeed, continue the earlier transaction. However, after looking at the file, it was
still incomplete.
That would seem to say the problem is on the web server side. Perhaps Firefox has some error-recovery code that wget does not.
EDIT2:
The same problem occurs for curl as well.
EDIT3:
I just tried the following command:
Code:
curl --compressed http://www.discoverygc.com/server/players_online.html > players_online.html
I received the whole, correct page twice in a row now. I can't say that it's not a fluke, but perhaps the request for the web server to compress the page forces the web server to collect all the information at once before sending it--possibly preventing a mid-stream break in the transfer.
I do not know if wget has a similar option.