Questions tagged [wget]

GNU Wget (or just Wget, formerly Geturl) is a computer program that retrieves content from web servers, and is part of the GNU Project. Its name is derived from World Wide Web and get. It supports downloading via HTTP, HTTPS, and FTP protocols.

GNU Wget (or just Wget, formerly Geturl) is a computer program that retrieves content from web servers, and is part of the GNU Project. Its name is derived from World Wide Web and get. It supports downloading via HTTP, HTTPS, and FTP protocols.

Its features include recursive download, conversion of links for offline viewing of local HTML, support for proxies, and much more. It appeared in 1996, coinciding with the boom of popularity of the Web, causing its wide use among Unix users and distribution with most major GNU/Linux based distributions. Written in portable C, Wget can be easily installed on any Unix-like system.

Source: wikipedia

Man page

290 questions
1
vote
1 answer

How to block some robots and referer pages in Apache root level

When I was using Lighttpd I can easily achieve this by such entries. So all websites were protected. Wget robots: $HTTP["useragent"] =~ "Wget" { $HTTP["url"] =~ "^/tagi(.*)" { # $HTTP["url"] =~ "" { url.access-deny = (…
Spacedust
  • 568
  • 5
  • 13
  • 28
1
vote
1 answer

Wget - download all links from a http location (not recursivly)

I have a link to an http page that has a structure like this: Parent Directory - [DIR] _OLD/ 01-Feb-2012 06:05 - [DIR] _Jan/ 01-Feb-2012 06:05 …
Cris
  • 245
  • 1
  • 7
  • 13
1
vote
3 answers

Limiting specific process to just one per user

My scripts are using wget to retrieve data from Internet. When many users are using this script I got very high load (about 20.00), because of disk I/O. Wget is automatically started each hour by cron. I would like to limit one wget to one customer…
Spacedust
  • 568
  • 5
  • 13
  • 28
1
vote
1 answer

Wget from HTTPS url using user/password -still getting 400 error

I'm trying to run Wget for a certain HTTP url, also using http-user and http-password command line options (tried the --user and --password, same results). Using the --no-check-certificate options as well so basically running this line : wget…
akapulko2020
  • 111
  • 1
  • 1
  • 7
1
vote
1 answer

wget and connection errors / timeouts

I was using wget the in the last week to recursively download a whole website of html pages. I used it this way: wget --recursive --no-clobber --page-requisites --html-extension --convert-links --domains XXXX.com --no-parent http://www.XXXX.com the…
soulSurfer2010
  • 307
  • 4
  • 10
1
vote
3 answers

download only new files from ftp using WGET/cUrl

i have Windows 2008 Server on which i store DB backups on daily basis. I want to be able to download new files only using wget, curl or windows built-in FTP doesn't matter. Can you help me with command? I need only new files, ones that exists on…
eugeneK
  • 410
  • 2
  • 8
  • 18
1
vote
2 answers

Download file in Linux when file location is unknown

Sorry if this is unclear, but I'm trying to set up a script that downloads a file. Currently, my method of downloading the file is by clicking on a link like so: https://www.URL.com/view?downloadFile=AcctItemFiles\1234567890.txt I tried using a…
Nick
  • 315
  • 2
  • 7
  • 15
1
vote
1 answer

Translate web page from shell

I'm trying to do translate a web page automatically with a shell script. Here is the interesting part: wget http://www.google.com/translate?langpair=en%7Ces&u=http%3A%2F%2Fasdf.com I'm getting the "ERROR 403: Forbidden" even though I can get the…
facha
  • 1,368
  • 2
  • 18
  • 26
1
vote
4 answers

wget an remote website index directory

Is it possible to wget an entire index/parent directory with wget? Because I am trying to transfer images to a VPS of mine from my old shared host. When I say an index/parent directory, I meant like this: So basically I want wget to go through each…
MacMac
  • 2,061
  • 10
  • 31
  • 38
1
vote
2 answers

wget has different response times for same fetch

Any ideas why one fetch would fetch a file (big picture in this example) fast while another fetch would fetch same file slow? Not seeing this issue at home. Not seeing this issue through ip addresses they map to 192.xx.xx.xx. Only seeing this…
Felice
  • 11
  • 1
1
vote
4 answers

Internet speed in Linux via command

I am running several sites and I was trying wget and I noticed slow speed. So I suspect that my sites are taking considerable amount of traffic. Question is, how I can really detect the total consumption of traffic in and out to know the real speed…
TDSii
1
vote
3 answers

wget: download a website which is password protected using a form

I have a problem with downloading a website using wget. First I need to login to the website with a username and password: wget --save-cookies cookies.txt --post-data --cookies=on --keep-session-cookies …
Rama
  • 11
  • 1
  • 3
1
vote
1 answer

wget not getting all files

I have issues using wget, basically I use it like: wget -m ftp://username:password@hostname.com:/path/to/dirs After everything's complete (it says xxx files in xx MB for xx seconds), however, when I inspect the files, I notice it actually didn't…
Andreas Wong
  • 219
  • 3
  • 10
1
vote
3 answers

wget is working only when used with sudo

I'm having quite a strange behavior with wget since yesterday. I can download files by using sudo wget, but when I try the same file with only wget, I can get this error: yusufh@ubuntu-yuh:~$ wget http://www.kegel.com/wine/winetricks --2010-12-17…
Yusuf
  • 205
  • 1
  • 3
  • 12
1
vote
3 answers

File timestamp not changing even with touch

I have the following problem. I have a bash script that uses wget to get a few files from the build server, and then scp's them over to the production system: Here is the relevant snippet: #Several of these lines wget -nv -O birt_reports.zip…
Yishai
  • 708
  • 1
  • 6
  • 15