tag:blogger.com,1999:blog-28555040.post6971286567520545037..comments2023-12-20T02:38:58.401+01:00Comments on Morfologik: Wikipedia history diff as a revision corpusMarcin Miłkowskihttp://www.blogger.com/profile/11617540925216664775noreply@blogger.comBlogger2125tag:blogger.com,1999:blog-28555040.post-84111919879840523732008-08-03T23:06:00.000+02:002008-08-03T23:06:00.000+02:00Actually, both :) I downloaded the 7bz file and de...Actually, both :) I downloaded the 7bz file and decompressed it to the standard output, feeding the gawk cleaning script with it. So of course you'd have to add the 7bz decompression time to the equation.<BR/><BR/>I also had a run on a decompressed XML corpus but I didn't notice much change in speed. Probably it was because I had it on a rather old drive and it was surely fragmented (it was large).<BR/><BR/>BTW, I have a publication forthcoming on this experiment in PALC 2007 proceedings. Here's a presentation: http://marcinmilkowski.pl/downloads/korpus_bledow.pdf<BR/>(in Polish, the publication will be in English).Marcin Miłkowskihttps://www.blogger.com/profile/11617540925216664775noreply@blogger.comtag:blogger.com,1999:blog-28555040.post-29142808302639717112008-07-30T11:11:00.000+02:002008-07-30T11:11:00.000+02:00Did you use the decompressed xml-File or the compr...Did you use the decompressed xml-File or the compressed 7z/bz2 Archive for your Experiment?Anonymousnoreply@blogger.com