mirror of
https://github.com/rfivet/uemacs.git
synced 2024-11-17 18:16:22 -05:00
Renaud Fivet
e13bc9ca7e
Consistent display of µEMACS as program name among - ue --version - on status bar - insert-string $progname - write-message $progname (FIX).
112 lines
2.7 KiB
C
112 lines
2.7 KiB
C
#include "utf8.h"
|
|
|
|
#include <assert.h>
|
|
|
|
/*
|
|
* utf8_to_unicode()
|
|
*
|
|
* Convert a UTF-8 sequence to its unicode value, and return the length of
|
|
* the sequence in bytes.
|
|
*
|
|
* NOTE! Invalid UTF-8 will be converted to a one-byte sequence, so you can
|
|
* either use it as-is (ie as Latin1) or you can check for invalid UTF-8
|
|
* by checking for a length of 1 and a result > 127.
|
|
*
|
|
* NOTE 2! This does *not* verify things like minimality. So overlong forms
|
|
* are happily accepted and decoded, as are the various "invalid values".
|
|
*/
|
|
unsigned utf8_to_unicode(char *line, unsigned index, unsigned len, unicode_t *res)
|
|
{
|
|
unicode_t value ;
|
|
unsigned char c = line[index];
|
|
unsigned bytes, mask, i;
|
|
|
|
*res = c;
|
|
|
|
/*
|
|
* 0xxxxxxx is valid one byte utf8
|
|
* 10xxxxxx is invalid UTF-8 start byte, we assume it is Latin1
|
|
* 1100000x is start of overlong encoding sequence
|
|
* Sequence longer than 4 bytes are invalid
|
|
* Last valid code is 0x10FFFF, encoding start with 0xF4
|
|
*/
|
|
if( c <= 0xC1 || c > 0xF4)
|
|
return 1;
|
|
|
|
/* Ok, it's 11xxxxxx, do a stupid decode */
|
|
mask = 0x20;
|
|
bytes = 2;
|
|
while (c & mask) {
|
|
bytes++;
|
|
mask >>= 1;
|
|
}
|
|
|
|
/* bytes is in range [2..4] */
|
|
len -= index;
|
|
if (bytes > len)
|
|
return 1;
|
|
|
|
value = c & (mask-1);
|
|
|
|
/* Ok, do the bytes */
|
|
line += index;
|
|
for (i = 1; i < bytes; i++) {
|
|
c = line[i];
|
|
if ((c & 0xc0) != 0x80)
|
|
return 1;
|
|
value = (value << 6) | (c & 0x3f);
|
|
}
|
|
|
|
if( value > 0x10FFFF) /* Avoid 110000 - 13FFFF */
|
|
return 1 ;
|
|
|
|
*res = value;
|
|
return bytes;
|
|
}
|
|
|
|
static void reverse_string(char *begin, char *end)
|
|
{
|
|
do {
|
|
char a = *begin, b = *end;
|
|
*end = a; *begin = b;
|
|
begin++; end--;
|
|
} while (begin < end);
|
|
}
|
|
|
|
/*
|
|
* unicode_to_utf8()
|
|
*
|
|
* Convert a unicode value to its canonical utf-8 sequence.
|
|
*
|
|
* NOTE! This does not check for - or care about - the "invalid" unicode
|
|
* values. Also, converting a utf-8 sequence to unicode and back does
|
|
* *not* guarantee the same sequence, since this generates the shortest
|
|
* possible sequence, while utf8_to_unicode() accepts both Latin1 and
|
|
* overlong utf-8 sequences.
|
|
*/
|
|
unsigned unicode_to_utf8( unicode_t c, char *utf8) {
|
|
int bytes = 1 ;
|
|
|
|
assert( c <= 0x10FFFF) ;
|
|
|
|
#ifdef NDEBUG
|
|
if( c > 0x10FFFF) /* Let's assume this is due to sign extension */
|
|
c &= 0xFF ;
|
|
#endif
|
|
|
|
*utf8 = c ;
|
|
if (c > 0x7f) {
|
|
int prefix = 0x40;
|
|
char *p = utf8;
|
|
do {
|
|
*p++ = 0x80 + (c & 0x3f);
|
|
bytes++;
|
|
prefix >>= 1;
|
|
c >>= 6;
|
|
} while( c >= prefix) ;
|
|
*p = c - 2*prefix;
|
|
reverse_string(utf8, p);
|
|
}
|
|
return bytes;
|
|
}
|