[Corpora-List] language-specific harvesting of texts from the Web
    Stuart A Yeates 
    stuart.yeates at computing-services.oxford.ac.uk
       
    Wed Sep  1 07:55:11 UTC 2004
    
    
  
Marco Baroni wrote:
>>One situation where your approach may not work so well, is when a
>>language's websites use multiple character encodings.  Unfortunately,
>>this is quite common in languages that have non-Roman writing systems,
>
>
> At least for Japanese, our way to get around this problem in our
> web-mining scripts was to look for the charset declaration in the html
> code of each page, and then to convert (inside the script) the page from
> that charset to utf8.
>
> I would be interested in hearing about other ways to deal with multiple
> encodings.
textcat (http://odur.let.rug.nl/~vannoord/TextCat/) is a language and
encoding guesser which reliably guesses test language and encoding based
solely on examples and statistics. Knows 69 natural languages. Open source.
I've had good experiance using the built-in java encoding converters
(readers and writers shipped for ~100 encodings as standard) to convert
between languages. Freely avaliable.
cheers
stuart
--
Stuart Yeates            stuart.yeates at computing-services.oxford.ac.uk
OSS Watch                                  http://www.oss-watch.ac.uk/
Oxford Text Archive                             http://ota.ahds.ac.uk/
Humbul Humanities Hub                         http://www.humbul.ac.uk/
    
    
More information about the Corpora
mailing list