确定C#中字符串的编码


127

有什么方法可以确定C#中的字符串编码吗?

说,我有一个文件名字符串,但是我不知道它是以Unicode UTF-16还是系统默认编码编码的,我如何找出来?


您不能使用Unicode“编码”。没有任何其他先验信息,就无法自动确定任何给定String 的编码。
Nicolas Dumazet,2009年

5
更清晰也许:你编码 Unicode代码点到使用“编码”方案(UTF-一个字符集的字节串,异,BIG5,移位JIS,等...),以及你解码从字节串字符集设置为Unicode。您不使用Unicode编码字节串。您不解码字节串中的Unicode。
Nicolas Dumazet

13
@NicDunZ-编码本身(特别是UTF-16)通常也称为“ Unicode”。是非是生命。即使在.NET中,也要查看Encoding.Unicode-表示UTF-16。
马克·格雷韦尔

2
哦,好吧,我不知道.NET如此具有误导性。这看起来像是一个可怕的学习习惯。抱歉@krebstar,这不是我的意图(我仍然认为您编辑的问题现在比以前有意义得多)
Nicolas Dumazet 2009年

1
@Nicdumz#1:有一种方法可以概率性地确定要使用的编码。查看IE(以及现在具有FF的View-Character Encoding-Auto-detect)的作用:尝试一种编码,看看它是否可能“写得很好<在此处输入语言名称>”或更改它然后再试一次。来吧,这很有趣!
SnippyHolloW 2009年

Answers:


31

签出Utf8Checker,这是一个简单的类,可以在纯托管代码中完成此任务。 http://utf8checker.codeplex.com

注意:正如已经指出的那样,“确定编码”仅对字节流有意义。如果您有一个字符串,则已经有人知道或猜测该编码是从某个人那里获取编码的。


如果该字符串是仅用8位编码完成的不正确解码,并且您使用了编码来对其进行解码,通常可以将字节取回而不会造成任何损坏。
Nyerguds

57

下面的代码具有以下功能:

  1. 检测或尝试检测UTF-7,UTF-8 / 16/32(BOM,无Bom,小端和大端)
  2. 如果未找到Unicode编码,则退回到本地默认代码页。
  3. 检测(很有可能)缺少BOM /签名的Unicode文件
  4. 在文件内搜索charset = xyz和encoding = xyz以帮助确定编码。
  5. 为了节省处理,您可以“品尝”文件(可定义的字节数)。
  6. 返回编码和解码的文本文件。
  7. 纯粹基于字节的解决方案,提高效率

就像其他人所说的那样,没有一种解决方案是完美的(当然不能轻易区分世界范围内使用的各种8位扩展ASCII编码),但是我们可以做到“足够好”,特别是如果开发人员还向用户展示替代编码列表,如下所示:每种语言最常见的编码是什么?

完整的编码列表可以使用 Encoding.GetEncodings();

// Function to detect the encoding for UTF-7, UTF-8/16/32 (bom, no bom, little
// & big endian), and local default codepage, and potentially other codepages.
// 'taster' = number of bytes to check of the file (to save processing). Higher
// value is slower, but more reliable (especially UTF-8 with special characters
// later on may appear to be ASCII initially). If taster = 0, then taster
// becomes the length of the file (for maximum reliability). 'text' is simply
// the string with the discovered encoding applied to the file.
public Encoding detectTextEncoding(string filename, out String text, int taster = 1000)
{
    byte[] b = File.ReadAllBytes(filename);

    //////////////// First check the low hanging fruit by checking if a
    //////////////// BOM/signature exists (sourced from http://www.unicode.org/faq/utf_bom.html#bom4)
    if (b.Length >= 4 && b[0] == 0x00 && b[1] == 0x00 && b[2] == 0xFE && b[3] == 0xFF) { text = Encoding.GetEncoding("utf-32BE").GetString(b, 4, b.Length - 4); return Encoding.GetEncoding("utf-32BE"); }  // UTF-32, big-endian 
    else if (b.Length >= 4 && b[0] == 0xFF && b[1] == 0xFE && b[2] == 0x00 && b[3] == 0x00) { text = Encoding.UTF32.GetString(b, 4, b.Length - 4); return Encoding.UTF32; }    // UTF-32, little-endian
    else if (b.Length >= 2 && b[0] == 0xFE && b[1] == 0xFF) { text = Encoding.BigEndianUnicode.GetString(b, 2, b.Length - 2); return Encoding.BigEndianUnicode; }     // UTF-16, big-endian
    else if (b.Length >= 2 && b[0] == 0xFF && b[1] == 0xFE) { text = Encoding.Unicode.GetString(b, 2, b.Length - 2); return Encoding.Unicode; }              // UTF-16, little-endian
    else if (b.Length >= 3 && b[0] == 0xEF && b[1] == 0xBB && b[2] == 0xBF) { text = Encoding.UTF8.GetString(b, 3, b.Length - 3); return Encoding.UTF8; } // UTF-8
    else if (b.Length >= 3 && b[0] == 0x2b && b[1] == 0x2f && b[2] == 0x76) { text = Encoding.UTF7.GetString(b,3,b.Length-3); return Encoding.UTF7; } // UTF-7


    //////////// If the code reaches here, no BOM/signature was found, so now
    //////////// we need to 'taste' the file to see if can manually discover
    //////////// the encoding. A high taster value is desired for UTF-8
    if (taster == 0 || taster > b.Length) taster = b.Length;    // Taster size can't be bigger than the filesize obviously.


    // Some text files are encoded in UTF8, but have no BOM/signature. Hence
    // the below manually checks for a UTF8 pattern. This code is based off
    // the top answer at: /programming/6555015/check-for-invalid-utf8
    // For our purposes, an unnecessarily strict (and terser/slower)
    // implementation is shown at: /programming/1031645/how-to-detect-utf-8-in-plain-c
    // For the below, false positives should be exceedingly rare (and would
    // be either slightly malformed UTF-8 (which would suit our purposes
    // anyway) or 8-bit extended ASCII/UTF-16/32 at a vanishingly long shot).
    int i = 0;
    bool utf8 = false;
    while (i < taster - 4)
    {
        if (b[i] <= 0x7F) { i += 1; continue; }     // If all characters are below 0x80, then it is valid UTF8, but UTF8 is not 'required' (and therefore the text is more desirable to be treated as the default codepage of the computer). Hence, there's no "utf8 = true;" code unlike the next three checks.
        if (b[i] >= 0xC2 && b[i] <= 0xDF && b[i + 1] >= 0x80 && b[i + 1] < 0xC0) { i += 2; utf8 = true; continue; }
        if (b[i] >= 0xE0 && b[i] <= 0xF0 && b[i + 1] >= 0x80 && b[i + 1] < 0xC0 && b[i + 2] >= 0x80 && b[i + 2] < 0xC0) { i += 3; utf8 = true; continue; }
        if (b[i] >= 0xF0 && b[i] <= 0xF4 && b[i + 1] >= 0x80 && b[i + 1] < 0xC0 && b[i + 2] >= 0x80 && b[i + 2] < 0xC0 && b[i + 3] >= 0x80 && b[i + 3] < 0xC0) { i += 4; utf8 = true; continue; }
        utf8 = false; break;
    }
    if (utf8 == true) {
        text = Encoding.UTF8.GetString(b);
        return Encoding.UTF8;
    }


    // The next check is a heuristic attempt to detect UTF-16 without a BOM.
    // We simply look for zeroes in odd or even byte places, and if a certain
    // threshold is reached, the code is 'probably' UF-16.          
    double threshold = 0.1; // proportion of chars step 2 which must be zeroed to be diagnosed as utf-16. 0.1 = 10%
    int count = 0;
    for (int n = 0; n < taster; n += 2) if (b[n] == 0) count++;
    if (((double)count) / taster > threshold) { text = Encoding.BigEndianUnicode.GetString(b); return Encoding.BigEndianUnicode; }
    count = 0;
    for (int n = 1; n < taster; n += 2) if (b[n] == 0) count++;
    if (((double)count) / taster > threshold) { text = Encoding.Unicode.GetString(b); return Encoding.Unicode; } // (little-endian)


    // Finally, a long shot - let's see if we can find "charset=xyz" or
    // "encoding=xyz" to identify the encoding:
    for (int n = 0; n < taster-9; n++)
    {
        if (
            ((b[n + 0] == 'c' || b[n + 0] == 'C') && (b[n + 1] == 'h' || b[n + 1] == 'H') && (b[n + 2] == 'a' || b[n + 2] == 'A') && (b[n + 3] == 'r' || b[n + 3] == 'R') && (b[n + 4] == 's' || b[n + 4] == 'S') && (b[n + 5] == 'e' || b[n + 5] == 'E') && (b[n + 6] == 't' || b[n + 6] == 'T') && (b[n + 7] == '=')) ||
            ((b[n + 0] == 'e' || b[n + 0] == 'E') && (b[n + 1] == 'n' || b[n + 1] == 'N') && (b[n + 2] == 'c' || b[n + 2] == 'C') && (b[n + 3] == 'o' || b[n + 3] == 'O') && (b[n + 4] == 'd' || b[n + 4] == 'D') && (b[n + 5] == 'i' || b[n + 5] == 'I') && (b[n + 6] == 'n' || b[n + 6] == 'N') && (b[n + 7] == 'g' || b[n + 7] == 'G') && (b[n + 8] == '='))
            )
        {
            if (b[n + 0] == 'c' || b[n + 0] == 'C') n += 8; else n += 9;
            if (b[n] == '"' || b[n] == '\'') n++;
            int oldn = n;
            while (n < taster && (b[n] == '_' || b[n] == '-' || (b[n] >= '0' && b[n] <= '9') || (b[n] >= 'a' && b[n] <= 'z') || (b[n] >= 'A' && b[n] <= 'Z')))
            { n++; }
            byte[] nb = new byte[n-oldn];
            Array.Copy(b, oldn, nb, 0, n-oldn);
            try {
                string internalEnc = Encoding.ASCII.GetString(nb);
                text = Encoding.GetEncoding(internalEnc).GetString(b);
                return Encoding.GetEncoding(internalEnc);
            }
            catch { break; }    // If C# doesn't recognize the name of the encoding, break.
        }
    }


    // If all else fails, the encoding is probably (though certainly not
    // definitely) the user's local codepage! One might present to the user a
    // list of alternative encodings as shown here: /programming/8509339/what-is-the-most-common-encoding-of-each-language
    // A full list can be found using Encoding.GetEncodings();
    text = Encoding.Default.GetString(b);
    return Encoding.Default;
}

这适用于西里尔字母(可能还有其他所有).eml文件(来自邮件的字符集标头)
Nime Cloud

实际上,UTF-7不能天真地解码。它的完整前导较长,并且包含第一个字符的两位。.Net系统似乎完全不支持UTF7的前导系统。
Nyerguds

当我检查的其他方法均无济于事时,为我工作!谢谢丹。
Tejasvi Hegde

感谢您的解决方案。我用它来确定来自完全不同来源的文件的编码。但是我发现,如果我使用的品尝者值太低,结果可能是错误的。(例如,即使我使用b.Length / 10作为我的品尝者,代码也会为UTF8文件返回Encoding.Default。)所以我很想知道,使用小于b.Length的品尝者有什么理由?看来我只能得出结论,即且仅当我扫描了整个文件时,Encoding.Default才可接受。
肖恩

@Sean:它适用于速度比精度更重要的情况,特别是对于可能数十或数百兆字节的文件。以我的经验,即使是较低的品尝价值也可以在大约99.9%的时间内产生正确的结果。您的经验可能有所不同。
Dan W

33

这取决于字符串从何而来。.NET字符串是Unicode(UTF-16)。例如,如果您从数据库中将数据读取到字节数组中,则唯一的方式可能会有所不同。

这篇CodeProject文章可能很有趣:检测输入和输出文本的编码

乔恩·斯基特(Jon Skeet)的C#和.NET中的字符串是.NET字符串的绝佳解释。


它来自非Unicode C ++应用程序。该CodeProject上的文章似乎有点太复杂了,但是似乎做我想做的事..谢谢..
krebstar

18

我知道这有点晚-但要明确:

字符串实际上没有编码...在.NET中,字符串是char对象的集合。本质上,如果它是字符串,则它已经被解码。

但是,如果您正在读取由字节组成的文件内容,并希望将其转换为字符串,则必须使用文件的编码。

.NET包括用于ASCII,UTF7,UTF8,UTF32等的编码和解码类。

这些编码中的大多数都包含某些字节顺序标记,这些标记可用于区分使用了哪种编码类型。

.NET类System.IO.StreamReader能够通过读取这些字节顺序标记来确定流中使用的编码。

这是一个例子:

    /// <summary>
    /// return the detected encoding and the contents of the file.
    /// </summary>
    /// <param name="fileName"></param>
    /// <param name="contents"></param>
    /// <returns></returns>
    public static Encoding DetectEncoding(String fileName, out String contents)
    {
        // open the file with the stream-reader:
        using (StreamReader reader = new StreamReader(fileName, true))
        {
            // read the contents of the file into a string
            contents = reader.ReadToEnd();

            // return the encoding.
            return reader.CurrentEncoding;
        }
    }

3
如果没有BOM表,这对于检测UTF 16将不起作用。如果无法检测到任何Unicode编码,它也不会退回到用户的本地默认代码页。您可以通过添加Encoding.Default作为StreamReader参数来修复后者,但是如果没有BOM表,代码将无法检测到UTF8。
丹·W

1
@DanW:不过,实际上没有做过没有BOM的UTF-16吗?我永远不会使用它;打开几乎所有东西注定是一场灾难。
Nyerguds

11

另一个选择,很晚才来,对不起:

http://www.architectshack.com/TextFileEncodingDetector.ashx

这个仅C#的小型类使用BOMS(如果存在),否则尝试自动检测可能的unicode编码,如果没有或不可能使用Unicode编码,则回退。

听起来上面提到的UTF8Checker做了类似的事情,但是我认为它的范围略宽-不仅是UTF8,它还会检查可能缺少BOM的其他可能的Unicode编码(UTF-16 LE或BE)。

希望这对某人有帮助!


很好的代码,它解决了我的编码检测问题:)
CARLOS LOTH'April

9

SimpleHelpers.FileEncoding NuGet包包装了一个Mozilla的通用字符检测器的C#端口到死简单的API:

var encoding = FileEncoding.DetectFileEncoding(txtFile);

这应该更高一些,它提供了一个非常简单的解决方案:让其他人来做:D
buddybubble

该库是GPL
的BR

是吗?我看到了MIT许可证,它使用了三重许可组件(UDE),其中之一是MPL。我一直在尝试确定UDE对于专有产品是否有问题,因此,如果您有更多信息,将不胜感激。
西蒙·伍兹

5

我的解决方案是使用内置的东西,并具有一些后备功能。

我从另一个类似stackoverflow问题的答案中选择了该策略,但现在找不到。

它首先使用StreamReader中的内置逻辑检查BOM,如果存在BOM,则编码将不是Encoding.Default,我们应该相信该结果。

如果不是,则检查字节序列是否为有效的UTF-8序列。如果是,它将猜测为UTF-8编码,如果不是,则再次使用默认的ASCII编码。

static Encoding getEncoding(string path) {
    var stream = new FileStream(path, FileMode.Open);
    var reader = new StreamReader(stream, Encoding.Default, true);
    reader.Read();

    if (reader.CurrentEncoding != Encoding.Default) {
        reader.Close();
        return reader.CurrentEncoding;
    }

    stream.Position = 0;

    reader = new StreamReader(stream, new UTF8Encoding(false, true));
    try {
        reader.ReadToEnd();
        reader.Close();
        return Encoding.UTF8;
    }
    catch (Exception) {
        reader.Close();
        return Encoding.Default;
    }
}

3

注意:这是一个实验,以了解UTF-8编码在内部如何工作。由vilicvane提供的解决方案,使用UTF8Encoding初始化为在解码失败时引发异常的对象,要简单得多,并且基本上可以完成相同的工作。


我写了这段代码来区分UTF-8和Windows-1252。但是,不应将其用于巨大的文本文件,因为它将整个内容加载到内存中并对其进行完全扫描。我将它用于.srt字幕文件,只是为了能够将它们保存回加载它们的编码中。

给该函数作为ref的编码应该是8位回退编码,以防检测到该文件不是有效的UTF-8。通常,在Windows系统上,它将是Windows-1252。但这并没有像检查实际的有效ascii范围那样花哨的事情,即使在字节顺序标记上也不会检测到UTF-16。

可以在此处找到按位检测的原理:https : //ianthehenry.com/2015/1/17/decoding-utf-8/

基本上,第一个字节的位范围确定其后是UTF-8实体的一部分。其后的这些字节始终在相同的位范围内。

/// <summary>
/// Reads a text file, and detects whether its encoding is valid UTF-8 or ascii.
/// If not, decodes the text using the given fallback encoding.
/// Bit-wise mechanism for detecting valid UTF-8 based on
/// https://ianthehenry.com/2015/1/17/decoding-utf-8/
/// </summary>
/// <param name="docBytes">The bytes read from the file.</param>
/// <param name="encoding">The default encoding to use as fallback if the text is detected not to be pure ascii or UTF-8 compliant. This ref parameter is changed to the detected encoding.</param>
/// <returns>The contents of the read file, as String.</returns>
public static String ReadFileAndGetEncoding(Byte[] docBytes, ref Encoding encoding)
{
    if (encoding == null)
        encoding = Encoding.GetEncoding(1252);
    Int32 len = docBytes.Length;
    // byte order mark for utf-8. Easiest way of detecting encoding.
    if (len > 3 && docBytes[0] == 0xEF && docBytes[1] == 0xBB && docBytes[2] == 0xBF)
    {
        encoding = new UTF8Encoding(true);
        // Note that even when initialising an encoding to have
        // a BOM, it does not cut it off the front of the input.
        return encoding.GetString(docBytes, 3, len - 3);
    }
    Boolean isPureAscii = true;
    Boolean isUtf8Valid = true;
    for (Int32 i = 0; i < len; ++i)
    {
        Int32 skip = TestUtf8(docBytes, i);
        if (skip == 0)
            continue;
        if (isPureAscii)
            isPureAscii = false;
        if (skip < 0)
        {
            isUtf8Valid = false;
            // if invalid utf8 is detected, there's no sense in going on.
            break;
        }
        i += skip;
    }
    if (isPureAscii)
        encoding = new ASCIIEncoding(); // pure 7-bit ascii.
    else if (isUtf8Valid)
        encoding = new UTF8Encoding(false);
    // else, retain given encoding. This should be an 8-bit encoding like Windows-1252.
    return encoding.GetString(docBytes);
}

/// <summary>
/// Tests if the bytes following the given offset are UTF-8 valid, and
/// returns the amount of bytes to skip ahead to do the next read if it is.
/// If the text is not UTF-8 valid it returns -1.
/// </summary>
/// <param name="binFile">Byte array to test</param>
/// <param name="offset">Offset in the byte array to test.</param>
/// <returns>The amount of bytes to skip ahead for the next read, or -1 if the byte sequence wasn't valid UTF-8</returns>
public static Int32 TestUtf8(Byte[] binFile, Int32 offset)
{
    // 7 bytes (so 6 added bytes) is the maximum the UTF-8 design could support,
    // but in reality it only goes up to 3, meaning the full amount is 4.
    const Int32 maxUtf8Length = 4;
    Byte current = binFile[offset];
    if ((current & 0x80) == 0)
        return 0; // valid 7-bit ascii. Added length is 0 bytes.
    Int32 len = binFile.Length;
    for (Int32 addedlength = 1; addedlength < maxUtf8Length; ++addedlength)
    {
        Int32 fullmask = 0x80;
        Int32 testmask = 0;
        // This code adds shifted bits to get the desired full mask.
        // If the full mask is [111]0 0000, then test mask will be [110]0 0000. Since this is
        // effectively always the previous step in the iteration I just store it each time.
        for (Int32 i = 0; i <= addedlength; ++i)
        {
            testmask = fullmask;
            fullmask += (0x80 >> (i+1));
        }
        // figure out bit masks from level
        if ((current & fullmask) == testmask)
        {
            if (offset + addedlength >= len)
                return -1;
            // Lookahead. Pattern of any following bytes is always 10xxxxxx
            for (Int32 i = 1; i <= addedlength; ++i)
            {
                if ((binFile[offset + i] & 0xC0) != 0x80)
                    return -1;
            }
            return addedlength;
        }
    }
    // Value is greater than the maximum allowed for utf8. Deemed invalid.
    return -1;
}

也没有else在之后的最后声明if ((current & 0xE0) == 0xC0) { ... } else if ((current & 0xF0) == 0xE0) { ... } else if ((current & 0xF0) == 0xE0) { ... } else if ((current & 0xF8) == 0xF0) { ... }。我想这else情况是无效的UTF8: isUtf8Valid = false;。你会?
hal

@hal Ah,是的。。。自那以后,我用一个更通用(更高级)的系统更新了自己的代码,该系统使用的循环可达3,但从技术上讲可以更改为进一步循环(对此尚不明确。 ;我认为可以将UTF-8扩展为最多增加6个字节,但是在当前实现中仅使用3个字节),因此我没有更新此代码。
Nyerguds

@hal将其更新为我的新解决方案。原理保持不变,但是位掩码是在循环中创建和检查的,而不是全部用代码明确写出的。
Nyerguds

1

我在GitHub上找到了新库:CharsetDetector / UTF-unknown

Charset检测器内置在C#中-.NET Core 2-3,.NET标准1-2和.NET 4+

它也是基于其他存储库的Mozilla通用字符集检测器的端口。

CharsetDetector / UTF-unknown有一个名为的类CharsetDetector

CharsetDetector 包含一些静态编码检测方法:

  • CharsetDetector.DetectFromFile()
  • CharsetDetector.DetectFromStream()
  • CharsetDetector.DetectFromBytes()

检测到的结果在class中DetectionResult具有属性Detected,该属性是DetectionDetail具有以下属性的类的实例:

  • EncodingName
  • Encoding
  • Confidence

以下是显示用法的示例:

// Program.cs
using System;
using System.Text;
using UtfUnknown;

namespace ConsoleExample
{
    public class Program
    {
        public static void Main(string[] args)
        {
            string filename = @"E:\new-file.txt";
            DetectDemo(filename);
        }

        /// <summary>
        /// Command line example: detect the encoding of the given file.
        /// </summary>
        /// <param name="filename">a filename</param>
        public static void DetectDemo(string filename)
        {
            // Detect from File
            DetectionResult result = CharsetDetector.DetectFromFile(filename);
            // Get the best Detection
            DetectionDetail resultDetected = result.Detected;

            // detected result may be null.
            if (resultDetected != null)
            {
                // Get the alias of the found encoding
                string encodingName = resultDetected.EncodingName;
                // Get the System.Text.Encoding of the found encoding (can be null if not available)
                Encoding encoding = resultDetected.Encoding;
                // Get the confidence of the found encoding (between 0 and 1)
                float confidence = resultDetected.Confidence;

                if (encoding != null)
                {
                    Console.WriteLine($"Detection completed: {filename}");
                    Console.WriteLine($"EncodingWebName: {encoding.WebName}{Environment.NewLine}Confidence: {confidence}");
                }
                else
                {
                    Console.WriteLine($"Detection completed: {filename}");
                    Console.WriteLine($"(Encoding is null){Environment.NewLine}EncodingName: {encodingName}{Environment.NewLine}Confidence: {confidence}");
                }
            }
            else
            {
                Console.WriteLine($"Detection failed: {filename}");
            }
        }
    }
}

示例结果截图: 在此处输入图片说明

By using our site, you acknowledge that you have read and understand our Cookie Policy and Privacy Policy.
Licensed under cc by-sa 3.0 with attribution required.