[Python-Dev] Encoding detection in the standard library?

M.-A. Lemburg mal at egenix.com
Wed Apr 23 12:01:18 CEST 2008


On 2008-04-23 07:26, Terry Reedy wrote:
> ""Martin v. Löwis"" <martin at v.loewis.de> wrote in message 
> news:480EC376.8070406 at v.loewis.de...
> |> I certainly agree that if the target set of documents is small enough it
> |
> | Ok. What advantage would you (or somebody working on a similar project)
> | gain if chardet was part of the standard library? What if it was not
> | chardet, but some other algorithm?
> 
> It seems to me that since there is not a 'correct' algorithm but only 
> competing heuristics, encoding detection modules should be made available 
> via PyPI and only be considered for stdlib after a best of breed emerges 
> with community support. 

+1

Though in practice, determining the "best of breed" often becomes a
problem (see e.g. the JSON implementation discussion).

-- 
Marc-Andre Lemburg
eGenix.com

Professional Python Services directly from the Source  (#1, Apr 23 2008)
 >>> Python/Zope Consulting and Support ...        http://www.egenix.com/
 >>> mxODBC.Zope.Database.Adapter ...             http://zope.egenix.com/
 >>> mxODBC, mxDateTime, mxTextTools ...        http://python.egenix.com/
________________________________________________________________________

:::: Try mxODBC.Zope.DA for Windows,Linux,Solaris,MacOSX for free ! ::::


    eGenix.com Software, Skills and Services GmbH  Pastor-Loeh-Str.48
     D-40764 Langenfeld, Germany. CEO Dipl.-Math. Marc-Andre Lemburg
            Registered at Amtsgericht Duesseldorf: HRB 46611


More information about the Python-Dev mailing list