[yunqa.de] Re: XML Parser error while importing a wiki dump

  • From: ". ." <sl01ck@xxxxxxxxx>
  • To: yunqa@xxxxxxxxxxxxx
  • Date: Wed, 7 Apr 2010 20:20:08 +0530

Hi,

Here is the log when i use wget to download the latest dump.

c:\wiki>wget -c
http://download.wikimedia.org/enwiki/20100312/enwiki-20100312-pages-articles.xml.bz2
--20:16:15--
http://download.wikimedia.org/enwiki/20100312/enwiki-20100312-pages-articles.xml.bz2
           => `enwiki-20100312-pages-articles.xml.bz2'
Resolving download.wikimedia.org... done.
Connecting to download.wikimedia.org[208.80.152.183]:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: 1,822,913,845 [application/octet-stream]

Notice that is shows 5.7 GB on the webpage but 1.6 GB when i download
it(Length). This has happened twice.

So are all dumps corrupted ?

Wasnt sure.

Would you know?

--L

On Wed, Apr 7, 2010 at 3:39 PM, Delphi Inspiration <delphi@xxxxxxxx> wrote:

> At 20:47 06.04.2010, . . wrote:
>
> >I downloaded enwiki-latest-pages-articles.xml.bz2 (1.6 GB).
>
> Your download is corrupted. The current
> enwiki-latest-pages-articles.xml.bz2 is 5.7 GB instead of just the 1.6 GB
> you downloaded.
>
> Quote from http://dumps.wikimedia.org/enwiki/latest/, taken today:
>
> >enwiki-latest-pages-articles.xml.bz2
> >2010-Mar-16 08:44:40
> >5.7G
> >application/octet-stream
>
> Please download again, possibly using a download manager. If you download a
> dated version, you can compare the MD5 sums to verify your download
> integrity.
>
> Currently the latest download and MD5 locations for the English language
> Wikipedia dump are here:
>
> http://dumps.wikimedia.org/enwiki/20100312/
>
> http://download.wikimedia.org/enwiki/20100312/enwiki-20100312-md5sums.txt
>
> Ralf
>
> _______________________________________________
> Delphi Inspiration mailing list
> yunqa@xxxxxxxxxxxxx
> //www.freelists.org/list/yunqa
>
>
>
>

Other related posts: