1 /* utf8.h 2 * 3 * This file contains definitions for use with the UTF-8 encoding. It 4 * actually also works with the variant UTF-8 encoding called UTF-EBCDIC, and 5 * hides almost all of the differences between these from the caller. In other 6 * words, someone should #include this file, and if the code is being compiled 7 * on an EBCDIC platform, things should mostly just work. 8 * 9 * Copyright (C) 2000, 2001, 2002, 2005, 2006, 2007, 2009, 10 * 2010, 2011 by Larry Wall and others 11 * 12 * You may distribute under the terms of either the GNU General Public 13 * License or the Artistic License, as specified in the README file. 14 * 15 * A note on nomenclature: The term UTF-8 is used loosely and inconsistently 16 * in Perl documentation. For one, perl uses an extension of UTF-8 to 17 * represent code points that Unicode considers illegal. For another, ASCII 18 * platform UTF-8 is usually conflated with EBCDIC platform UTF-EBCDIC, because 19 * outside some of the macros in this this file, the differences are hopefully 20 * invisible at the semantic level. 21 * 22 * UTF-EBCDIC has an isomorphic translation named I8 (for "Intermediate eight") 23 * which differs from UTF-8 only in a few details. It is often useful to 24 * translate UTF-EBCDIC into this form for processing. In general, macros and 25 * functions that are expecting their inputs to be either in I8 or UTF-8 are 26 * named UTF_foo (without an '8'), to indicate this. 27 * 28 * Unfortunately there are inconsistencies. 29 * 30 */ 31 32 #ifndef PERL_UTF8_H_ /* Guard against recursive inclusion */ 33 #define PERL_UTF8_H_ 1 34 35 /* 36 =for apidoc Ay||utf8ness_t 37 38 This typedef is used by several core functions that return PV strings, to 39 indicate the UTF-8ness of those strings. 40 41 (If you write a new function, you probably should instead return the PV in an 42 SV with the UTF-8 flag of the SV properly set, rather than use this mechanism.) 43 44 The possible values this can be are: 45 46 =over 47 48 =item C<UTF8NESS_YES> 49 50 This means the string definitely should be treated as a sequence of 51 UTF-8-encoded characters. 52 53 Most code that needs to handle this typedef should be of the form: 54 55 if (utf8ness_flag == UTF8NESS_YES) { 56 treat as utf8; // like turning on an SV UTF-8 flag 57 } 58 59 =item C<UTF8NESS_NO> 60 61 This means the string definitely should be treated as a sequence of bytes, not 62 encoded as UTF-8. 63 64 =item C<UTF8NESS_IMMATERIAL> 65 66 This means it is equally valid to treat the string as bytes, or as UTF-8 67 characters; use whichever way you want. This happens when the string consists 68 entirely of characters which have the same representation whether encoded in 69 UTF-8 or not. 70 71 =item C<UTF8NESS_UNKNOWN> 72 73 This means it is unknown how the string should be treated. No core function 74 will ever return this value to a non-core caller. Instead, it is used by the 75 caller to initialize a variable to a non-legal value. A typical call will look like: 76 77 utf8ness_t string_is_utf8 = UTF8NESS_UNKNOWN 78 const char * string = foo(arg1, arg2, ..., &string_is_utf8); 79 if (string_is_utf8 == UTF8NESS_YES) { 80 do something for UTF-8; 81 } 82 83 =back 84 85 The following relationships hold between the enum values: 86 87 =over 88 89 =item S<C<0 E<lt>= I<enum value> E<lt>= UTF8NESS_IMMATERIAL>> 90 91 the string may be treated in code as non-UTF8 92 93 =item S<C<UTF8NESS_IMMATERIAL E<lt>= <I<enum value>>> 94 95 the string may be treated in code as encoded in UTF-8 96 97 =back 98 99 =cut 100 */ 101 102 typedef enum { 103 UTF8NESS_NO = 0, /* Definitely not UTF-8 */ 104 UTF8NESS_IMMATERIAL = 1, /* Representation is the same in UTF-8 as 105 not, so the UTF8ness doesn't actually 106 matter */ 107 UTF8NESS_YES = 2, /* Defintely is UTF-8, wideness 108 unspecified */ 109 UTF8NESS_UNKNOWN = -1, /* Undetermined so far */ 110 } utf8ness_t; 111 112 /* Use UTF-8 as the default script encoding? 113 * Turning this on will break scripts having non-UTF-8 binary 114 * data (such as Latin-1) in string literals. */ 115 #ifdef USE_UTF8_SCRIPTS 116 # define USE_UTF8_IN_NAMES (!IN_BYTES) 117 #else 118 # define USE_UTF8_IN_NAMES (PL_hints & HINT_UTF8) 119 #endif 120 121 #include "regcharclass.h" 122 #include "unicode_constants.h" 123 124 /* For to_utf8_fold_flags, q.v. */ 125 #define FOLD_FLAGS_LOCALE 0x1 126 #define FOLD_FLAGS_FULL 0x2 127 #define FOLD_FLAGS_NOMIX_ASCII 0x4 128 129 /* 130 =for apidoc is_ascii_string 131 132 This is a misleadingly-named synonym for L</is_utf8_invariant_string>. 133 On ASCII-ish platforms, the name isn't misleading: the ASCII-range characters 134 are exactly the UTF-8 invariants. But EBCDIC machines have more invariants 135 than just the ASCII characters, so C<is_utf8_invariant_string> is preferred. 136 137 =for apidoc is_invariant_string 138 139 This is a somewhat misleadingly-named synonym for L</is_utf8_invariant_string>. 140 C<is_utf8_invariant_string> is preferred, as it indicates under what conditions 141 the string is invariant. 142 143 =cut 144 */ 145 #define is_ascii_string(s, len) is_utf8_invariant_string(s, len) 146 #define is_invariant_string(s, len) is_utf8_invariant_string(s, len) 147 148 #define uvoffuni_to_utf8_flags(d,uv,flags) \ 149 uvoffuni_to_utf8_flags_msgs(d, uv, flags, 0) 150 #define uvchr_to_utf8(a,b) uvchr_to_utf8_flags(a,b,0) 151 #define uvchr_to_utf8_flags(d,uv,flags) \ 152 uvchr_to_utf8_flags_msgs(d,uv,flags, 0) 153 #define uvchr_to_utf8_flags_msgs(d,uv,flags,msgs) \ 154 uvoffuni_to_utf8_flags_msgs(d,NATIVE_TO_UNI(uv),flags, msgs) 155 #define utf8_to_uvchr_buf(s, e, lenp) \ 156 utf8_to_uvchr_buf_helper((const U8 *) (s), (const U8 *) e, lenp) 157 #define utf8n_to_uvchr(s, len, lenp, flags) \ 158 utf8n_to_uvchr_error(s, len, lenp, flags, 0) 159 #define utf8n_to_uvchr_error(s, len, lenp, flags, errors) \ 160 utf8n_to_uvchr_msgs(s, len, lenp, flags, errors, 0) 161 162 #define utf16_to_utf8(p, d, bytelen, newlen) \ 163 utf16_to_utf8_base(p, d, bytelen, newlen, 0, 1) 164 #define utf16_to_utf8_reversed(p, d, bytelen, newlen) \ 165 utf16_to_utf8_base(p, d, bytelen, newlen, 1, 0) 166 #define utf8_to_utf16(p, d, bytelen, newlen) \ 167 utf8_to_utf16_base(p, d, bytelen, newlen, 0, 1) 168 #define utf8_to_utf16_reversed(p, d, bytelen, newlen) \ 169 utf8_to_utf16_base(p, d, bytelen, newlen, 1, 0) 170 171 #define to_uni_fold(c, p, lenp) _to_uni_fold_flags(c, p, lenp, FOLD_FLAGS_FULL) 172 173 #define foldEQ_utf8(s1, pe1, l1, u1, s2, pe2, l2, u2) \ 174 foldEQ_utf8_flags(s1, pe1, l1, u1, s2, pe2, l2, u2, 0) 175 #define FOLDEQ_UTF8_NOMIX_ASCII (1 << 0) 176 #define FOLDEQ_LOCALE (1 << 1) 177 #define FOLDEQ_S1_ALREADY_FOLDED (1 << 2) 178 #define FOLDEQ_S2_ALREADY_FOLDED (1 << 3) 179 #define FOLDEQ_S1_FOLDS_SANE (1 << 4) 180 #define FOLDEQ_S2_FOLDS_SANE (1 << 5) 181 182 /* This will be described more fully below, but it turns out that the 183 * fundamental difference between UTF-8 and UTF-EBCDIC is that the former has 184 * the upper 2 bits of a continuation byte be '10', and the latter has the 185 * upper 3 bits be '101', leaving 6 and 5 significant bits respectively. 186 * 187 * It is helpful to know the EBCDIC value on ASCII platforms, mainly to avoid 188 * some #ifdef's */ 189 #define UTF_EBCDIC_CONTINUATION_BYTE_INFO_BITS 5 190 191 /* See explanation below at 'UTF8_MAXBYTES' */ 192 #define ASCII_PLATFORM_UTF8_MAXBYTES 13 193 194 #ifdef EBCDIC 195 196 /* The equivalent of the next few macros but implementing UTF-EBCDIC are in the 197 * following header file: */ 198 # include "utfebcdic.h" 199 200 # else /* ! EBCDIC */ 201 202 START_EXTERN_C 203 204 # ifndef DOINIT 205 EXTCONST unsigned char PL_utf8skip[]; 206 # else 207 EXTCONST unsigned char PL_utf8skip[] = { 208 /* 0x00 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 209 /* 0x10 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 210 /* 0x20 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 211 /* 0x30 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 212 /* 0x40 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 213 /* 0x50 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 214 /* 0x60 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 215 /* 0x70 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* ascii */ 216 /* 0x80 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* bogus: continuation byte */ 217 /* 0x90 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* bogus: continuation byte */ 218 /* 0xA0 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* bogus: continuation byte */ 219 /* 0xB0 */ 1,1,1,1,1,1,1,1,1,1,1,1,1,1,1,1, /* bogus: continuation byte */ 220 /* 0xC0 */ 2,2, /* overlong */ 221 /* 0xC2 */ 2,2,2,2,2,2,2,2,2,2,2,2,2,2, /* U+0080 to U+03FF */ 222 /* 0xD0 */ 2,2,2,2,2,2,2,2,2,2,2,2,2,2,2,2, /* U+0400 to U+07FF */ 223 /* 0xE0 */ 3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3, /* U+0800 to U+FFFF */ 224 /* 0xF0 */ 4,4,4,4,4,4,4,4,5,5,5,5,6,6, /* above BMP to 2**31 - 1 */ 225 /* Perl extended (never was official UTF-8). Up to 36 bit */ 226 /* 0xFE */ 7, 227 /* More extended, Up to 72 bits (64-bit + reserved) */ 228 /* 0xFF */ ASCII_PLATFORM_UTF8_MAXBYTES 229 }; 230 # endif 231 232 END_EXTERN_C 233 234 /* 235 236 =for apidoc Am|U8|NATIVE_TO_LATIN1|U8 ch 237 238 Returns the Latin-1 (including ASCII and control characters) equivalent of the 239 input native code point given by C<ch>. Thus, C<NATIVE_TO_LATIN1(193)> on 240 EBCDIC platforms returns 65. These each represent the character C<"A"> on 241 their respective platforms. On ASCII platforms no conversion is needed, so 242 this macro expands to just its input, adding no time nor space requirements to 243 the implementation. 244 245 For conversion of code points potentially larger than will fit in a character, 246 use L</NATIVE_TO_UNI>. 247 248 =for apidoc Am|U8|LATIN1_TO_NATIVE|U8 ch 249 250 Returns the native equivalent of the input Latin-1 code point (including ASCII 251 and control characters) given by C<ch>. Thus, C<LATIN1_TO_NATIVE(66)> on 252 EBCDIC platforms returns 194. These each represent the character C<"B"> on 253 their respective platforms. On ASCII platforms no conversion is needed, so 254 this macro expands to just its input, adding no time nor space requirements to 255 the implementation. 256 257 For conversion of code points potentially larger than will fit in a character, 258 use L</UNI_TO_NATIVE>. 259 260 =for apidoc Am|UV|NATIVE_TO_UNI|UV ch 261 262 Returns the Unicode equivalent of the input native code point given by C<ch>. 263 Thus, C<NATIVE_TO_UNI(195)> on EBCDIC platforms returns 67. These each 264 represent the character C<"C"> on their respective platforms. On ASCII 265 platforms no conversion is needed, so this macro expands to just its input, 266 adding no time nor space requirements to the implementation. 267 268 =for apidoc Am|UV|UNI_TO_NATIVE|UV ch 269 270 Returns the native equivalent of the input Unicode code point given by C<ch>. 271 Thus, C<UNI_TO_NATIVE(68)> on EBCDIC platforms returns 196. These each 272 represent the character C<"D"> on their respective platforms. On ASCII 273 platforms no conversion is needed, so this macro expands to just its input, 274 adding no time nor space requirements to the implementation. 275 276 =cut 277 */ 278 279 #define NATIVE_TO_LATIN1(ch) (__ASSERT_(FITS_IN_8_BITS(ch)) ((U8) (ch))) 280 #define LATIN1_TO_NATIVE(ch) (__ASSERT_(FITS_IN_8_BITS(ch)) ((U8) (ch))) 281 282 /* I8 is an intermediate version of UTF-8 used only in UTF-EBCDIC. We thus 283 * consider it to be identical to UTF-8 on ASCII platforms. Strictly speaking 284 * UTF-8 and UTF-EBCDIC are two different things, but we often conflate them 285 * because they are 8-bit encodings that serve the same purpose in Perl, and 286 * rarely do we need to distinguish them. The term "NATIVE_UTF8" applies to 287 * whichever one is applicable on the current platform */ 288 #define NATIVE_UTF8_TO_I8(ch) (__ASSERT_(FITS_IN_8_BITS(ch)) ((U8) (ch))) 289 #define I8_TO_NATIVE_UTF8(ch) (__ASSERT_(FITS_IN_8_BITS(ch)) ((U8) (ch))) 290 291 #define UNI_TO_NATIVE(ch) ((UV) ASSERT_NOT_PTR(ch)) 292 #define NATIVE_TO_UNI(ch) ((UV) ASSERT_NOT_PTR(ch)) 293 294 /* 295 296 The following table is from Unicode 3.2, plus the Perl extensions for above 297 U+10FFFF 298 299 Code Points 1st Byte 2nd Byte 3rd 4th 5th 6th 7th 8th-13th 300 301 U+0000..U+007F 00..7F 302 U+0080..U+07FF * C2..DF 80..BF 303 U+0800..U+0FFF E0 * A0..BF 80..BF 304 U+1000..U+CFFF E1..EC 80..BF 80..BF 305 U+D000..U+D7FF ED 80..9F 80..BF 306 U+D800..U+DFFF ED A0..BF 80..BF (surrogates) 307 U+E000..U+FFFF EE..EF 80..BF 80..BF 308 U+10000..U+3FFFF F0 * 90..BF 80..BF 80..BF 309 U+40000..U+FFFFF F1..F3 80..BF 80..BF 80..BF 310 U+100000..U+10FFFF F4 80..8F 80..BF 80..BF 311 Below are above-Unicode code points 312 U+110000..U+13FFFF F4 90..BF 80..BF 80..BF 313 U+110000..U+1FFFFF F5..F7 80..BF 80..BF 80..BF 314 U+200000..U+FFFFFF F8 * 88..BF 80..BF 80..BF 80..BF 315 U+1000000..U+3FFFFFF F9..FB 80..BF 80..BF 80..BF 80..BF 316 U+4000000..U+3FFFFFFF FC * 84..BF 80..BF 80..BF 80..BF 80..BF 317 U+40000000..U+7FFFFFFF FD 80..BF 80..BF 80..BF 80..BF 80..BF 318 U+80000000..U+FFFFFFFFF FE * 82..BF 80..BF 80..BF 80..BF 80..BF 80..BF 319 U+1000000000.. FF 80..BF 80..BF 80..BF 80..BF 80..BF * 81..BF 80..BF 320 321 Note the gaps before several of the byte entries above marked by '*'. These are 322 caused by legal UTF-8 avoiding non-shortest encodings: it is technically 323 possible to UTF-8-encode a single code point in different ways, but that is 324 explicitly forbidden, and the shortest possible encoding should always be used 325 (and that is what Perl does). The non-shortest ones are called 'overlongs'. 326 327 Another way to look at it, as bits: 328 329 Code Points 1st Byte 2nd Byte 3rd Byte 4th Byte 330 331 0aaa aaaa 0aaa aaaa 332 0000 0bbb bbaa aaaa 110b bbbb 10aa aaaa 333 cccc bbbb bbaa aaaa 1110 cccc 10bb bbbb 10aa aaaa 334 00 000d ddcc cccc bbbb bbaa aaaa 1111 0ddd 10cc cccc 10bb bbbb 10aa aaaa 335 336 As you can see, the continuation bytes all begin with C<10>, and the 337 leading bits of the start byte tell how many bytes there are in the 338 encoded character. 339 340 Perl's extended UTF-8 means we can have start bytes up through FF, though any 341 beginning with FF yields a code point that is too large for 32-bit ASCII 342 platforms. FF signals to use 13 bytes for the encoded character. This breaks 343 the paradigm that the number of leading bits gives how many total bytes there 344 are in the character. */ 345 346 /* This is the number of low-order bits a continuation byte in a UTF-8 encoded 347 * sequence contributes to the specification of the code point. In the bit 348 * maps above, you see that the first 2 bits are a constant '10', leaving 6 of 349 * real information */ 350 # define UTF_CONTINUATION_BYTE_INFO_BITS 6 351 352 /* ^? is defined to be DEL on ASCII systems. See the definition of toCTRL() 353 * for more */ 354 # define QUESTION_MARK_CTRL DEL_NATIVE 355 356 #endif /* EBCDIC vs ASCII */ 357 358 /* It turns out that in a number of cases, that handling ASCII vs EBCDIC is a 359 * matter of being off-by-one. So this is a convenience macro, used to avoid 360 * some #ifdefs. */ 361 #define ONE_IF_EBCDIC_ZERO_IF_NOT \ 362 (UTF_CONTINUATION_BYTE_INFO_BITS == UTF_EBCDIC_CONTINUATION_BYTE_INFO_BITS) 363 364 /* Since the significant bits in a continuation byte are stored in the 365 * least-significant positions, we often find ourselves shifting by that 366 * amount. This is a clearer name in such situations */ 367 #define UTF_ACCUMULATION_SHIFT UTF_CONTINUATION_BYTE_INFO_BITS 368 369 /* 2**info_bits - 1. This masks out all but the bits that carry real 370 * information in a continuation byte. This turns out to be 0x3F in UTF-8, 371 * 0x1F in UTF-EBCDIC. */ 372 #define UTF_CONTINUATION_MASK \ 373 ((U8) nBIT_MASK(UTF_CONTINUATION_BYTE_INFO_BITS)) 374 375 /* For use in UTF8_IS_CONTINUATION(). This turns out to be 0xC0 in UTF-8, 376 * E0 in UTF-EBCDIC */ 377 #define UTF_IS_CONTINUATION_MASK \ 378 ((U8) ((0xFF << UTF_ACCUMULATION_SHIFT) & 0xFF)) 379 380 /* This defines the bits that are to be in the continuation bytes of a 381 * multi-byte UTF-8 encoded character that mark it is a continuation byte. 382 * This turns out to be 0x80 in UTF-8, 0xA0 in UTF-EBCDIC. (khw doesn't know 383 * the underlying reason that B0 works here, except it just happens to work. 384 * One could solve for two linear equations and come up with it.) */ 385 #define UTF_CONTINUATION_MARK (UTF_IS_CONTINUATION_MASK & 0xB0) 386 387 /* This value is clearer in some contexts */ 388 #define UTF_MIN_CONTINUATION_BYTE UTF_CONTINUATION_MARK 389 390 /* Is the byte 'c' part of a multi-byte UTF8-8 encoded sequence, and not the 391 * first byte thereof? */ 392 #define UTF8_IS_CONTINUATION(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 393 (((NATIVE_UTF8_TO_I8(c) & UTF_IS_CONTINUATION_MASK) \ 394 == UTF_CONTINUATION_MARK))) 395 396 /* Is the representation of the Unicode code point 'cp' the same regardless of 397 * being encoded in UTF-8 or not? This is a fundamental property of 398 * UTF-8,EBCDIC */ 399 #define OFFUNI_IS_INVARIANT(c) \ 400 (((WIDEST_UTYPE)(c)) < UTF_MIN_CONTINUATION_BYTE) 401 402 /* 403 =for apidoc Am|bool|UVCHR_IS_INVARIANT|UV cp 404 405 Evaluates to 1 if the representation of code point C<cp> is the same whether or 406 not it is encoded in UTF-8; otherwise evaluates to 0. UTF-8 invariant 407 characters can be copied as-is when converting to/from UTF-8, saving time. 408 C<cp> is Unicode if above 255; otherwise is platform-native. 409 410 =cut 411 */ 412 #if defined(__m88k__) 413 /* XXX workaround: m88k gcc3 produces wrong code with NATIVE_TO_UNI() */ 414 #define UVCHR_IS_INVARIANT(cp) (OFFUNI_IS_INVARIANT(cp)) 415 #else /* the original one */ 416 #define UVCHR_IS_INVARIANT(cp) (OFFUNI_IS_INVARIANT(NATIVE_TO_UNI(cp))) 417 #endif 418 419 /* This defines the 1-bits that are to be in the first byte of a multi-byte 420 * UTF-8 encoded character that mark it as a start byte and give the number of 421 * bytes that comprise the character. 'len' is that number. 422 * 423 * To illustrate: len = 2 => ((U8) ~ 0b0011_1111) or 1100_0000 424 * 7 => ((U8) ~ 0b0000_0001) or 1111_1110 425 * > 7 => 0xFF 426 * 427 * This is not to be used on a single-byte character. As in many places in 428 * perl, U8 must be 8 bits 429 */ 430 #define UTF_START_MARK(len) ((U8) ~(0xFF >> (len))) 431 432 /* Masks out the initial one bits in a start byte, leaving the following 0 bit 433 * and the real data bits. 'len' is the number of bytes in the multi-byte 434 * sequence that comprises the character. 435 * 436 * To illustrate: len = 2 => 0b0011_1111 works on start byte 110xxxxx 437 * 6 => 0b0000_0011 works on start byte 1111110x 438 * >= 7 => There are no data bits in the start byte 439 * Note that on ASCII platforms, this can be passed a len=1 byte; and all the 440 * real data bits will be returned: 441 len = 1 => 0b0111_1111 442 * This isn't true on EBCDIC platforms, where some len=1 bytes are of the form 443 * 0b101x_xxxx, so this can't be used there on single-byte characters. */ 444 #define UTF_START_MASK(len) (0xFF >> (len)) 445 446 /* 447 448 =for apidoc AmnU|STRLEN|UTF8_MAXBYTES 449 450 The maximum width of a single UTF-8 encoded character, in bytes. 451 452 NOTE: Strictly speaking Perl's UTF-8 should not be called UTF-8 since UTF-8 453 is an encoding of Unicode, and Unicode's upper limit, 0x10FFFF, can be 454 expressed with 4 bytes. However, Perl thinks of UTF-8 as a way to encode 455 non-negative integers in a binary format, even those above Unicode. 456 457 =cut 458 459 The start byte 0xFE, never used in any ASCII platform UTF-8 specification, has 460 an obvious meaning, namely it has its upper 7 bits set, so it should start a 461 sequence of 7 bytes. And in fact, this is exactly what standard UTF-EBCDIC 462 does. 463 464 The start byte FF, on the other hand could have several different plausible 465 meanings: 466 1) The meaning in standard UTF-EBCDIC, namely as an FE start byte, with the 467 bottom bit that should be a fixed '0' to form FE, instead acting as an 468 info bit, 0 or 1. 469 2) That the sequence should have exactly 8 bytes. 470 3) That the next byte is to be treated as a sort of extended start byte, 471 which in combination with this one gives the total length of the sequence. 472 There are published UTF-8 extensions that do this, some string together 473 multiple initial FF start bytes to achieve arbitrary precision. 474 4) That the sequence has exactly n bytes, where n is what the implementation 475 chooses. 476 477 Perl has chosen 4). 478 The goal is to be able to represent 64-bit values in UTF-8 or UTF-EBCDIC. That 479 rules out items 1) and 2). Item 3) has the deal-breaking disadvantage of 480 requiring one to read more than one byte to determine the total length of the 481 sequence. So in Perl, a start byte of FF indicates a UTF-8 string consisting 482 of the start byte, plus enough continuation bytes to encode a 64 bit value. 483 This turns out to be 13 total bytes in UTF-8 and 14 in UTF-EBCDIC. This is 484 because we get zero info bits from the start byte, plus 485 12 * 6 bits of info per continuation byte (could encode 72-bit numbers) on 486 UTF-8 (khw knows not why 11, which would encode 66 bits wasn't 487 chosen instead); and 488 13 * 5 bits of info per byte (could encode 65-bit numbers) on UTF-EBCDIC 489 490 The disadvantages of this method are: 491 1) There's potentially a lot of wasted bytes for all but the largest values. 492 For example, something that could be represented by 7 continuation bytes, 493 instead requires the full 12 or 13. 494 2) There would be problems should larger values, 128-bit say, ever need to be 495 represented. 496 497 WARNING: This number must be in sync with the value in 498 regen/charset_translations.pl. */ 499 #define UTF8_MAXBYTES \ 500 (ASCII_PLATFORM_UTF8_MAXBYTES + ONE_IF_EBCDIC_ZERO_IF_NOT) 501 502 /* Calculate how many bytes are necessary to represent a value whose most 503 * significant 1 bit is in bit position 'pos' of the word. For 0x1, 'pos would 504 * be 0; and for 0x400, 'pos' would be 10, and the result would be: 505 * EBCDIC floor((-1 + (10 + 5 - 1 - 1)) / (5 - 1)) 506 * = floor((-1 + (13)) / 4) 507 * = floor(12 / 4) 508 * = 3 509 * ASCII floor(( 0 + (10 + 6 - 1 - 1)) / (6 - 1)) 510 * = floor(14 / 5) 511 * = 2 512 * The reason this works is because the number of bits needed to represent a 513 * value is proportional to (UTF_CONTINUATION_BYTE_INFO_BITS - 1). The -1 is 514 * because each new continuation byte removes one bit of information from the 515 * start byte. 516 * 517 * This is a step function (we need to allocate a full extra byte if we 518 * overflow by just a single bit) 519 * 520 * The caller is responsible for making sure 'pos' is at least 8 (occupies 9 521 * bits), as it breaks down at the lower edge. At the high end, if it returns 522 * 8 or more, Perl instead anomalously uses MAX_BYTES, so this would be wrong. 523 * */ 524 #define UNISKIP_BY_MSB_(pos) \ 525 ( ( -ONE_IF_EBCDIC_ZERO_IF_NOT /* platform break pos's are off-by-one */ \ 526 + (pos) + ((UTF_CONTINUATION_BYTE_INFO_BITS - 1) - 1)) /* Step fcn */ \ 527 / (UTF_CONTINUATION_BYTE_INFO_BITS - 1)) /* take floor of */ 528 529 /* Compute the number of UTF-8 bytes required for representing the input uv, 530 * which must be a Unicode, not native value. 531 * 532 * This uses msbit_pos() which doesn't work on NUL, and UNISKIP_BY_MSB_ breaks 533 * down for small code points. So first check if the input is invariant to get 534 * around that, and use a helper for high code points to accommodate the fact 535 * that above 7 btyes, the value is anomalous. The helper is empty on 536 * platforms that don't go that high */ 537 #define OFFUNISKIP(uv) \ 538 ((OFFUNI_IS_INVARIANT(uv)) \ 539 ? 1 \ 540 : (OFFUNISKIP_helper_(uv) UNISKIP_BY_MSB_(msbit_pos(uv)))) 541 542 /* We need to go to MAX_BYTES when we can't represent 'uv' by the number of 543 * information bits in 6 continuation bytes (when we get to 6, the start byte 544 * has no information bits to add to the total). But on 32-bit ASCII 545 * platforms, that doesn't happen until 6*6 bits, so on those platforms, this 546 * will always be false */ 547 #if UVSIZE * CHARBITS > (6 * UTF_CONTINUATION_BYTE_INFO_BITS) 548 # define HAS_EXTRA_LONG_UTF8 549 # define OFFUNISKIP_helper_(uv) \ 550 UNLIKELY(uv > nBIT_UMAX(6 * UTF_CONTINUATION_BYTE_INFO_BITS)) \ 551 ? UTF8_MAXBYTES : 552 #else 553 # define OFFUNISKIP_helper_(uv) 554 #endif 555 556 /* 557 558 =for apidoc Am|STRLEN|UVCHR_SKIP|UV cp 559 returns the number of bytes required to represent the code point C<cp> when 560 encoded as UTF-8. C<cp> is a native (ASCII or EBCDIC) code point if less than 561 255; a Unicode code point otherwise. 562 563 =cut 564 */ 565 #define UVCHR_SKIP(uv) OFFUNISKIP(NATIVE_TO_UNI(uv)) 566 567 #define NATIVE_SKIP(uv) UVCHR_SKIP(uv) /* Old terminology */ 568 569 /* Most code which says UNISKIP is really thinking in terms of native code 570 * points (0-255) plus all those beyond. This is an imprecise term, but having 571 * it means existing code continues to work. For precision, use UVCHR_SKIP, 572 * NATIVE_SKIP, or OFFUNISKIP */ 573 #define UNISKIP(uv) UVCHR_SKIP(uv) 574 575 /* Compute the start byte for a given code point. This requires the log2 of 576 * the code point, which is hard to compute at compile time, which this macro 577 * wants to be. (Perhaps deBruijn sequences could be used.) So a parameter 578 * for the number of bits the value occupies is passed in, which the programmer 579 * has had to figure out to get compile-time effect. And asserts are used to 580 * make sure the value is correct. 581 * 582 * Since we are interested only in the start byte, we ignore the lower bits 583 * accounted for by the continuation bytes. Each continuation byte eats up 584 * UTF_CONTINUATION_BYTE_INFO_BITS bits, so the number of continuation bytes 585 * needed is floor(bits / UTF_CONTINUATION_BYTE_INFO_BITS). That number is fed 586 * to UTF_START_MARK() to get the upper part of the start byte. The left over 587 * bits form the lower part which is OR'd with the mark 588 * 589 * Note that on EBCDIC platforms, this is actually the I8 */ 590 #define UTF_START_BYTE(uv, bits) \ 591 (__ASSERT_((uv) >> ((bits) - 1)) /* At least 'bits' */ \ 592 __ASSERT_(((uv) & ~nBIT_MASK(bits)) == 0) /* No extra bits */ \ 593 UTF_START_MARK(UNISKIP_BY_MSB_((bits) - 1)) \ 594 | ((uv) >> (((bits) / UTF_CONTINUATION_BYTE_INFO_BITS) \ 595 * UTF_CONTINUATION_BYTE_INFO_BITS))) 596 597 /* Compute the first continuation byte for a given code point. This is mostly 598 * for compile-time, so how many bits it occupies is also passed in). 599 * 600 * We are interested in the first continuation byte, so we ignore the lower 601 * bits accounted for by the rest of the continuation bytes by right shifting 602 * out their info bit, and mask out the higher bits that will go into the start 603 * byte. 604 * 605 * Note that on EBCDIC platforms, this is actually the I8 */ 606 #define UTF_FIRST_CONT_BYTE(uv, bits) \ 607 (__ASSERT_((uv) >> ((bits) - 1)) /* At least 'bits' */ \ 608 __ASSERT_(((uv) & ~nBIT_MASK(bits)) == 0) /* No extra bits */ \ 609 UTF_CONTINUATION_MARK \ 610 | ( UTF_CONTINUATION_MASK \ 611 & ((uv) >> ((((bits) / UTF_CONTINUATION_BYTE_INFO_BITS) - 1) \ 612 * UTF_CONTINUATION_BYTE_INFO_BITS)))) 613 614 #define UTF_MIN_START_BYTE UTF_START_BYTE(UTF_MIN_CONTINUATION_BYTE, 8) 615 616 /* Is the byte 'c' the first byte of a multi-byte UTF8-8 encoded sequence? 617 * This excludes invariants (they are single-byte). It also excludes the 618 * illegal overlong sequences that begin with C0 and C1 on ASCII platforms, and 619 * C0-C4 I8 start bytes on EBCDIC ones. On EBCDIC E0 can't start a 620 * non-overlong sequence, so we define a base macro and for those platforms, 621 * extend it to also exclude E0 */ 622 #define UTF8_IS_START_base(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 623 (NATIVE_UTF8_TO_I8(c) >= UTF_MIN_START_BYTE)) 624 #ifdef EBCDIC 625 # define UTF8_IS_START(c) \ 626 (UTF8_IS_START_base(c) && (c) != I8_TO_NATIVE_UTF8(0xE0)) 627 #else 628 # define UTF8_IS_START(c) UTF8_IS_START_base(c) 629 #endif 630 631 #define UTF_MIN_ABOVE_LATIN1_BYTE UTF_START_BYTE(0x100, 9) 632 633 /* Is the UTF8-encoded byte 'c' the first byte of a sequence of bytes that 634 * represent a code point > 255? */ 635 #define UTF8_IS_ABOVE_LATIN1(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 636 (NATIVE_UTF8_TO_I8(c) >= UTF_MIN_ABOVE_LATIN1_BYTE)) 637 638 /* Is the UTF8-encoded byte 'c' the first byte of a two byte sequence? Use 639 * UTF8_IS_NEXT_CHAR_DOWNGRADEABLE() instead if the input isn't known to 640 * be well-formed. */ 641 #define UTF8_IS_DOWNGRADEABLE_START(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 642 inRANGE_helper_(U8, NATIVE_UTF8_TO_I8(c), \ 643 UTF_MIN_START_BYTE, UTF_MIN_ABOVE_LATIN1_BYTE - 1)) 644 645 /* The largest code point representable by two UTF-8 bytes on this platform. 646 * The binary for that code point is: 647 * 1101_1111 10xx_xxxx in UTF-8, and 648 * 1101_1111 101y_yyyy in UTF-EBCDIC I8. 649 * where both x and y are 1, and shown this way to indicate there is one more x 650 * than there is y. The number of x and y bits are their platform's respective 651 * UTF_CONTINUATION_BYTE_INFO_BITS. Squeezing out the bits that don't 652 * contribute to the value, these evaluate to: 653 * 1_1111 xx_xxxx in UTF-8, and 654 * 1_1111 y_yyyy in UTF-EBCDIC I8. 655 * or, the maximum value of an unsigned with (5 + info_bit_count) bits */ 656 #define MAX_UTF8_TWO_BYTE nBIT_UMAX(5 + UTF_CONTINUATION_BYTE_INFO_BITS) 657 658 /* The largest code point representable by two UTF-8 bytes on any platform that 659 * Perl runs on. */ 660 #define MAX_PORTABLE_UTF8_TWO_BYTE \ 661 nBIT_UMAX(5 + MIN( UTF_CONTINUATION_BYTE_INFO_BITS, \ 662 UTF_EBCDIC_CONTINUATION_BYTE_INFO_BITS)) 663 664 /* 665 666 =for apidoc AmnU|STRLEN|UTF8_MAXBYTES_CASE 667 668 The maximum number of UTF-8 bytes a single Unicode character can 669 uppercase/lowercase/titlecase/fold into. 670 671 =cut 672 673 * Unicode guarantees that the maximum expansion is UTF8_MAX_FOLD_CHAR_EXPAND 674 * characters, but any above-Unicode code point will fold to itself, so we only 675 * have to look at the expansion of the maximum Unicode code point. But this 676 * number may be less than the space occupied by a very large code point under 677 * Perl's extended UTF-8. We have to make it large enough to fit any single 678 * character. (It turns out that ASCII and EBCDIC differ in which is larger) 679 * 680 =cut 681 */ 682 #define UTF8_MAXBYTES_CASE \ 683 MAX(UTF8_MAXBYTES, UTF8_MAX_FOLD_CHAR_EXPAND * UNISKIP_BY_MSB_(20)) 684 685 /* Rest of these are attributes of Unicode and perl's internals rather than the 686 * encoding, or happen to be the same in both ASCII and EBCDIC (at least at 687 * this level; the macros that some of these call may have different 688 * definitions in the two encodings */ 689 690 /* In domain restricted to ASCII, these may make more sense to the reader than 691 * the ones with Latin1 in the name */ 692 #define NATIVE_TO_ASCII(ch) NATIVE_TO_LATIN1(ch) 693 #define ASCII_TO_NATIVE(ch) LATIN1_TO_NATIVE(ch) 694 695 /* More or less misleadingly-named defines, retained for back compat */ 696 #define NATIVE_TO_UTF(ch) NATIVE_UTF8_TO_I8(ch) 697 #define NATIVE_TO_I8(ch) NATIVE_UTF8_TO_I8(ch) 698 #define UTF_TO_NATIVE(ch) I8_TO_NATIVE_UTF8(ch) 699 #define I8_TO_NATIVE(ch) I8_TO_NATIVE_UTF8(ch) 700 #define NATIVE8_TO_UNI(ch) NATIVE_TO_LATIN1(ch) 701 702 /* Adds a UTF8 continuation byte 'new' of information to a running total code 703 * point 'old' of all the continuation bytes so far. This is designed to be 704 * used in a loop to convert from UTF-8 to the code point represented. Note 705 * that this is asymmetric on EBCDIC platforms, in that the 'new' parameter is 706 * the UTF-EBCDIC byte, whereas the 'old' parameter is a Unicode (not EBCDIC) 707 * code point in process of being generated */ 708 #define UTF8_ACCUMULATE(old, new) (__ASSERT_(FITS_IN_8_BITS(new)) \ 709 ((old) << UTF_ACCUMULATION_SHIFT) \ 710 | ((NATIVE_UTF8_TO_I8(new)) \ 711 & UTF_CONTINUATION_MASK)) 712 713 /* This works in the face of malformed UTF-8. */ 714 #define UTF8_IS_NEXT_CHAR_DOWNGRADEABLE(s, e) \ 715 ( ( (e) - (s) > 1) \ 716 && UTF8_IS_DOWNGRADEABLE_START(*(s)) \ 717 && UTF8_IS_CONTINUATION(*((s)+1))) 718 719 /* Longer, but more accurate name */ 720 #define UTF8_IS_ABOVE_LATIN1_START(c) UTF8_IS_ABOVE_LATIN1(c) 721 722 /* Convert a UTF-8 variant Latin1 character to a native code point value. 723 * Needs just one iteration of accumulate. Should be used only if it is known 724 * that the code point is < 256, and is not UTF-8 invariant. Use the slower 725 * but more general TWO_BYTE_UTF8_TO_NATIVE() which handles any code point 726 * representable by two bytes (which turns out to be up through 727 * MAX_PORTABLE_UTF8_TWO_BYTE). The two parameters are: 728 * HI: a downgradable start byte; 729 * LO: continuation. 730 * */ 731 #define EIGHT_BIT_UTF8_TO_NATIVE(HI, LO) \ 732 ( __ASSERT_(UTF8_IS_DOWNGRADEABLE_START(HI)) \ 733 __ASSERT_(UTF8_IS_CONTINUATION(LO)) \ 734 LATIN1_TO_NATIVE(UTF8_ACCUMULATE(( \ 735 NATIVE_UTF8_TO_I8(HI) & UTF_START_MASK(2)), (LO)))) 736 737 /* Convert a two (not one) byte utf8 character to a native code point value. 738 * Needs just one iteration of accumulate. Should not be used unless it is 739 * known that the two bytes are legal: 1) two-byte start, and 2) continuation. 740 * Note that the result can be larger than 255 if the input character is not 741 * downgradable */ 742 #define TWO_BYTE_UTF8_TO_NATIVE(HI, LO) \ 743 (__ASSERT_(FITS_IN_8_BITS(HI)) \ 744 __ASSERT_(FITS_IN_8_BITS(LO)) \ 745 __ASSERT_(PL_utf8skip[(U8) HI] == 2) \ 746 __ASSERT_(UTF8_IS_CONTINUATION(LO)) \ 747 UNI_TO_NATIVE(UTF8_ACCUMULATE((NATIVE_UTF8_TO_I8(HI) & UTF_START_MASK(2)), \ 748 (LO)))) 749 750 /* Should never be used, and be deprecated */ 751 #define TWO_BYTE_UTF8_TO_UNI(HI, LO) NATIVE_TO_UNI(TWO_BYTE_UTF8_TO_NATIVE(HI, LO)) 752 753 /* 754 755 =for apidoc Am|STRLEN|UTF8SKIP|char* s 756 returns the number of bytes a non-malformed UTF-8 encoded character whose first 757 (perhaps only) byte is pointed to by C<s>. 758 759 If there is a possibility of malformed input, use instead: 760 761 =over 762 763 =item C<L</UTF8_SAFE_SKIP>> if you know the maximum ending pointer in the 764 buffer pointed to by C<s>; or 765 766 =item C<L</UTF8_CHK_SKIP>> if you don't know it. 767 768 =back 769 770 It is better to restructure your code so the end pointer is passed down so that 771 you know what it actually is at the point of this call, but if that isn't 772 possible, C<L</UTF8_CHK_SKIP>> can minimize the chance of accessing beyond the end 773 of the input buffer. 774 775 =cut 776 */ 777 #define UTF8SKIP(s) PL_utf8skip[*(const U8*)(ASSERT_IS_PTR(s))] 778 779 /* 780 =for apidoc Am|STRLEN|UTF8_SKIP|char* s 781 This is a synonym for C<L</UTF8SKIP>> 782 783 =cut 784 */ 785 786 #define UTF8_SKIP(s) UTF8SKIP(s) 787 788 /* 789 =for apidoc Am|STRLEN|UTF8_CHK_SKIP|char* s 790 791 This is a safer version of C<L</UTF8SKIP>>, but still not as safe as 792 C<L</UTF8_SAFE_SKIP>>. This version doesn't blindly assume that the input 793 string pointed to by C<s> is well-formed, but verifies that there isn't a NUL 794 terminating character before the expected end of the next character in C<s>. 795 The length C<UTF8_CHK_SKIP> returns stops just before any such NUL. 796 797 Perl tends to add NULs, as an insurance policy, after the end of strings in 798 SV's, so it is likely that using this macro will prevent inadvertent reading 799 beyond the end of the input buffer, even if it is malformed UTF-8. 800 801 This macro is intended to be used by XS modules where the inputs could be 802 malformed, and it isn't feasible to restructure to use the safer 803 C<L</UTF8_SAFE_SKIP>>, for example when interfacing with a C library. 804 805 =cut 806 */ 807 808 #define UTF8_CHK_SKIP(s) \ 809 (UNLIKELY(s[0] == '\0') ? 1 : MIN(UTF8SKIP(s), \ 810 my_strnlen((char *) (s), UTF8SKIP(s)))) 811 /* 812 813 =for apidoc Am|STRLEN|UTF8_SAFE_SKIP|char* s|char* e 814 returns 0 if S<C<s E<gt>= e>>; otherwise returns the number of bytes in the 815 UTF-8 encoded character whose first byte is pointed to by C<s>. But it never 816 returns beyond C<e>. On DEBUGGING builds, it asserts that S<C<s E<lt>= e>>. 817 818 =cut 819 */ 820 #define UTF8_SAFE_SKIP(s, e) (__ASSERT_((e) >= (s)) \ 821 UNLIKELY(((e) - (s)) <= 0) \ 822 ? 0 \ 823 : MIN(((e) - (s)), UTF8_SKIP(s))) 824 825 /* Most code that says 'UNI_' really means the native value for code points up 826 * through 255 */ 827 #define UNI_IS_INVARIANT(cp) UVCHR_IS_INVARIANT(cp) 828 829 /* 830 =for apidoc Am|bool|UTF8_IS_INVARIANT|char c 831 832 Evaluates to 1 if the byte C<c> represents the same character when encoded in 833 UTF-8 as when not; otherwise evaluates to 0. UTF-8 invariant characters can be 834 copied as-is when converting to/from UTF-8, saving time. 835 836 In spite of the name, this macro gives the correct result if the input string 837 from which C<c> comes is not encoded in UTF-8. 838 839 See C<L</UVCHR_IS_INVARIANT>> for checking if a UV is invariant. 840 841 =cut 842 843 The reason it works on both UTF-8 encoded strings and non-UTF-8 encoded, is 844 that it returns TRUE in each for the exact same set of bit patterns. It is 845 valid on a subset of what UVCHR_IS_INVARIANT is valid on, so can just use that; 846 and the compiler should optimize out anything extraneous given the 847 implementation of the latter. */ 848 #define UTF8_IS_INVARIANT(c) UVCHR_IS_INVARIANT(ASSERT_NOT_PTR(c)) 849 850 /* Like the above, but its name implies a non-UTF8 input, which as the comments 851 * above show, doesn't matter as to its implementation */ 852 #define NATIVE_BYTE_IS_INVARIANT(c) UVCHR_IS_INVARIANT(c) 853 854 /* Misleadingly named: is the UTF8-encoded byte 'c' part of a variant sequence 855 * in UTF-8? This is the inverse of UTF8_IS_INVARIANT. */ 856 #define UTF8_IS_CONTINUED(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 857 (! UTF8_IS_INVARIANT(c))) 858 859 /* The macros in the next 4 sets are used to generate the two utf8 or utfebcdic 860 * bytes from an ordinal that is known to fit into exactly two (not one) bytes; 861 * it must be less than 0x3FF to work across both encodings. */ 862 863 /* These two are helper macros for the other three sets, and should not be used 864 * directly anywhere else. 'translate_function' is either NATIVE_TO_LATIN1 865 * (which works for code points up through 0xFF) or NATIVE_TO_UNI which works 866 * for any code point */ 867 #define __BASE_TWO_BYTE_HI(c, translate_function) \ 868 (__ASSERT_(! UVCHR_IS_INVARIANT(c)) \ 869 I8_TO_NATIVE_UTF8((translate_function(c) >> UTF_ACCUMULATION_SHIFT) \ 870 | UTF_START_MARK(2))) 871 #define __BASE_TWO_BYTE_LO(c, translate_function) \ 872 (__ASSERT_(! UVCHR_IS_INVARIANT(c)) \ 873 I8_TO_NATIVE_UTF8((translate_function(c) & UTF_CONTINUATION_MASK) \ 874 | UTF_CONTINUATION_MARK)) 875 876 /* The next two macros should not be used. They were designed to be usable as 877 * the case label of a switch statement, but this doesn't work for EBCDIC. Use 878 * regen/unicode_constants.pl instead */ 879 #define UTF8_TWO_BYTE_HI_nocast(c) __BASE_TWO_BYTE_HI(c, NATIVE_TO_UNI) 880 #define UTF8_TWO_BYTE_LO_nocast(c) __BASE_TWO_BYTE_LO(c, NATIVE_TO_UNI) 881 882 /* The next two macros are used when the source should be a single byte 883 * character; checked for under DEBUGGING */ 884 #define UTF8_EIGHT_BIT_HI(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 885 ( __BASE_TWO_BYTE_HI(c, NATIVE_TO_LATIN1))) 886 #define UTF8_EIGHT_BIT_LO(c) (__ASSERT_(FITS_IN_8_BITS(c)) \ 887 (__BASE_TWO_BYTE_LO(c, NATIVE_TO_LATIN1))) 888 889 /* These final two macros in the series are used when the source can be any 890 * code point whose UTF-8 is known to occupy 2 bytes; they are less efficient 891 * than the EIGHT_BIT versions on EBCDIC platforms. We use the logical '~' 892 * operator instead of "<=" to avoid getting compiler warnings. 893 * MAX_UTF8_TWO_BYTE should be exactly all one bits in the lower few 894 * places, so the ~ works */ 895 #define UTF8_TWO_BYTE_HI(c) \ 896 (__ASSERT_((sizeof(c) == 1) \ 897 || !(((WIDEST_UTYPE)(c)) & ~MAX_UTF8_TWO_BYTE)) \ 898 (__BASE_TWO_BYTE_HI(c, NATIVE_TO_UNI))) 899 #define UTF8_TWO_BYTE_LO(c) \ 900 (__ASSERT_((sizeof(c) == 1) \ 901 || !(((WIDEST_UTYPE)(c)) & ~MAX_UTF8_TWO_BYTE)) \ 902 (__BASE_TWO_BYTE_LO(c, NATIVE_TO_UNI))) 903 904 /* This is illegal in any well-formed UTF-8 in both EBCDIC and ASCII 905 * as it is only in overlongs. */ 906 #define ILLEGAL_UTF8_BYTE I8_TO_NATIVE_UTF8(0xC1) 907 908 /* 909 * 'UTF' is whether or not p is encoded in UTF8. The names 'foo_lazy_if' stem 910 * from an earlier version of these macros in which they didn't call the 911 * foo_utf8() macros (i.e. were 'lazy') unless they decided that *p is the 912 * beginning of a utf8 character. Now that foo_utf8() determines that itself, 913 * no need to do it again here 914 */ 915 #define isIDFIRST_lazy_if_safe(p, e, UTF) \ 916 ((IN_BYTES || !UTF) \ 917 ? isIDFIRST(*(p)) \ 918 : isIDFIRST_utf8_safe(p, e)) 919 #define isWORDCHAR_lazy_if_safe(p, e, UTF) \ 920 ((IN_BYTES || !UTF) \ 921 ? isWORDCHAR(*(p)) \ 922 : isWORDCHAR_utf8_safe((U8 *) p, (U8 *) e)) 923 #define isALNUM_lazy_if_safe(p, e, UTF) isWORDCHAR_lazy_if_safe(p, e, UTF) 924 925 #define UTF8_MAXLEN UTF8_MAXBYTES 926 927 /* A Unicode character can fold to up to 3 characters */ 928 #define UTF8_MAX_FOLD_CHAR_EXPAND 3 929 930 #define IN_BYTES UNLIKELY(CopHINTS_get(PL_curcop) & HINT_BYTES) 931 932 /* 933 934 =for apidoc Am|bool|DO_UTF8|SV* sv 935 Returns a bool giving whether or not the PV in C<sv> is to be treated as being 936 encoded in UTF-8. 937 938 You should use this I<after> a call to C<SvPV()> or one of its variants, in 939 case any call to string overloading updates the internal UTF-8 encoding flag. 940 941 =cut 942 */ 943 #define DO_UTF8(sv) (SvUTF8(sv) && !IN_BYTES) 944 945 /* Should all strings be treated as Unicode, and not just UTF-8 encoded ones? 946 * Is so within 'feature unicode_strings' or 'locale :not_characters', and not 947 * within 'use bytes'. UTF-8 locales are not tested for here, because it gets 948 * complicated by the probability of having categories in different locales. */ 949 #define IN_UNI_8_BIT \ 950 (( ( (CopHINTS_get(PL_curcop) & HINT_UNI_8_BIT)) \ 951 || ( CopHINTS_get(PL_curcop) & HINT_LOCALE_PARTIAL \ 952 /* -1 below is for :not_characters */ \ 953 && _is_in_locale_category(FALSE, -1))) \ 954 && (! IN_BYTES)) 955 956 #define UNICODE_SURROGATE_FIRST 0xD800 957 #define UNICODE_SURROGATE_LAST 0xDFFF 958 959 /* 960 =for apidoc Am|bool|UNICODE_IS_SURROGATE|const UV uv 961 962 Returns a boolean as to whether or not C<uv> is one of the Unicode surrogate 963 code points 964 965 =for apidoc Am|bool|UTF8_IS_SURROGATE|const U8 *s|const U8 *e 966 967 Evaluates to non-zero if the first few bytes of the string starting at C<s> and 968 looking no further than S<C<e - 1>> are well-formed UTF-8 that represents one 969 of the Unicode surrogate code points; otherwise it evaluates to 0. If 970 non-zero, the value gives how many bytes starting at C<s> comprise the code 971 point's representation. 972 973 =cut 974 */ 975 976 #define UNICODE_IS_SURROGATE(uv) UNLIKELY(inRANGE(uv, UNICODE_SURROGATE_FIRST, \ 977 UNICODE_SURROGATE_LAST)) 978 #define UTF8_IS_SURROGATE(s, e) is_SURROGATE_utf8_safe(s, e) 979 980 /* 981 982 =for apidoc AmnU|UV|UNICODE_REPLACEMENT 983 984 Evaluates to 0xFFFD, the code point of the Unicode REPLACEMENT CHARACTER 985 986 =for apidoc Am|bool|UNICODE_IS_REPLACEMENT|const UV uv 987 988 Returns a boolean as to whether or not C<uv> is the Unicode REPLACEMENT 989 CHARACTER 990 991 =for apidoc Am|bool|UTF8_IS_REPLACEMENT|const U8 *s|const U8 *e 992 993 Evaluates to non-zero if the first few bytes of the string starting at C<s> and 994 looking no further than S<C<e - 1>> are well-formed UTF-8 that represents the 995 Unicode REPLACEMENT CHARACTER; otherwise it evaluates to 0. If non-zero, the 996 value gives how many bytes starting at C<s> comprise the code point's 997 representation. 998 999 =cut 1000 */ 1001 #define UNICODE_REPLACEMENT 0xFFFD 1002 #define UNICODE_IS_REPLACEMENT(uv) UNLIKELY((UV) (uv) == UNICODE_REPLACEMENT) 1003 #define UTF8_IS_REPLACEMENT(s, send) \ 1004 UNLIKELY( \ 1005 ((send) - (s)) >= ((SSize_t)(sizeof(REPLACEMENT_CHARACTER_UTF8) - 1))\ 1006 && memEQ((s), REPLACEMENT_CHARACTER_UTF8, \ 1007 sizeof(REPLACEMENT_CHARACTER_UTF8) - 1)) 1008 1009 /* Max legal code point according to Unicode */ 1010 #define PERL_UNICODE_MAX 0x10FFFF 1011 1012 /* 1013 1014 =for apidoc Am|bool|UNICODE_IS_SUPER|const UV uv 1015 1016 Returns a boolean as to whether or not C<uv> is above the maximum legal Unicode 1017 code point of U+10FFFF. 1018 1019 =cut 1020 */ 1021 1022 #define UNICODE_IS_SUPER(uv) UNLIKELY((UV) (uv) > PERL_UNICODE_MAX) 1023 1024 /* 1025 =for apidoc Am|bool|UTF8_IS_SUPER|const U8 *s|const U8 *e 1026 1027 Recall that Perl recognizes an extension to UTF-8 that can encode code 1028 points larger than the ones defined by Unicode, which are 0..0x10FFFF. 1029 1030 This macro evaluates to non-zero if the first few bytes of the string starting 1031 at C<s> and looking no further than S<C<e - 1>> are from this UTF-8 extension; 1032 otherwise it evaluates to 0. If non-zero, the return is how many bytes 1033 starting at C<s> comprise the code point's representation. 1034 1035 0 is returned if the bytes are not well-formed extended UTF-8, or if they 1036 represent a code point that cannot fit in a UV on the current platform. Hence 1037 this macro can give different results when run on a 64-bit word machine than on 1038 one with a 32-bit word size. 1039 1040 Note that it is illegal in Perl to have code points that are larger than what can 1041 fit in an IV on the current machine; and illegal in Unicode to have any that 1042 this macro matches 1043 1044 =cut 1045 1046 * ASCII EBCDIC I8 1047 * U+10FFFF: \xF4\x8F\xBF\xBF \xF9\xA1\xBF\xBF\xBF max legal Unicode 1048 * U+110000: \xF4\x90\x80\x80 \xF9\xA2\xA0\xA0\xA0 1049 * U+110001: \xF4\x90\x80\x81 \xF9\xA2\xA0\xA0\xA1 1050 */ 1051 #define UTF_START_BYTE_110000_ UTF_START_BYTE(PERL_UNICODE_MAX + 1, 21) 1052 #define UTF_FIRST_CONT_BYTE_110000_ \ 1053 UTF_FIRST_CONT_BYTE(PERL_UNICODE_MAX + 1, 21) 1054 #define UTF8_IS_SUPER(s, e) \ 1055 ( ((e) - (s)) >= UNISKIP_BY_MSB_(20) \ 1056 && ( NATIVE_UTF8_TO_I8(s[0]) >= UTF_START_BYTE_110000_ \ 1057 && ( NATIVE_UTF8_TO_I8(s[0]) > UTF_START_BYTE_110000_ \ 1058 || NATIVE_UTF8_TO_I8(s[1]) >= UTF_FIRST_CONT_BYTE_110000_))) \ 1059 ? isUTF8_CHAR(s, e) \ 1060 : 0 1061 1062 /* 1063 =for apidoc Am|bool|UNICODE_IS_NONCHAR|const UV uv 1064 1065 Returns a boolean as to whether or not C<uv> is one of the Unicode 1066 non-character code points 1067 1068 =cut 1069 */ 1070 1071 /* Is 'uv' one of the 32 contiguous-range noncharacters? */ 1072 #define UNICODE_IS_32_CONTIGUOUS_NONCHARS(uv) \ 1073 UNLIKELY(inRANGE(uv, 0xFDD0, 0xFDEF)) 1074 1075 /* Is 'uv' one of the 34 plane-ending noncharacters 0xFFFE, 0xFFFF, 0x1FFFE, 1076 * 0x1FFFF, ... 0x10FFFE, 0x10FFFF, given that we know that 'uv' is not above 1077 * the Unicode legal max */ 1078 #define UNICODE_IS_END_PLANE_NONCHAR_GIVEN_NOT_SUPER(uv) \ 1079 UNLIKELY(((UV) (uv) & 0xFFFE) == 0xFFFE) 1080 1081 #define UNICODE_IS_NONCHAR(uv) \ 1082 ( UNLIKELY(UNICODE_IS_32_CONTIGUOUS_NONCHARS(uv)) \ 1083 || ( UNLIKELY(UNICODE_IS_END_PLANE_NONCHAR_GIVEN_NOT_SUPER(uv)) \ 1084 && LIKELY(! UNICODE_IS_SUPER(uv)))) 1085 1086 /* 1087 =for apidoc Am|bool|UTF8_IS_NONCHAR|const U8 *s|const U8 *e 1088 1089 Evaluates to non-zero if the first few bytes of the string starting at C<s> and 1090 looking no further than S<C<e - 1>> are well-formed UTF-8 that represents one 1091 of the Unicode non-character code points; otherwise it evaluates to 0. If 1092 non-zero, the value gives how many bytes starting at C<s> comprise the code 1093 point's representation. 1094 1095 =cut 1096 */ 1097 #define UTF8_IS_NONCHAR(s, e) is_NONCHAR_utf8_safe(s,e) 1098 1099 /* This is now machine generated, and the 'given' clause is no longer 1100 * applicable */ 1101 #define UTF8_IS_NONCHAR_GIVEN_THAT_NON_SUPER_AND_GE_PROBLEMATIC(s, e) \ 1102 UTF8_IS_NONCHAR(s, e) 1103 1104 /* Surrogates, non-character code points and above-Unicode code points are 1105 * problematic in some contexts. These macros allow code that needs to check 1106 * for those to quickly exclude the vast majority of code points it will 1107 * encounter. 1108 * 1109 * The lowest such code point is the smallest surrogate, U+D800. We calculate 1110 * the start byte of that. 0xD800 occupies 16 bits. */ 1111 #define isUNICODE_POSSIBLY_PROBLEMATIC(uv) ((uv) >= UNICODE_SURROGATE_FIRST) 1112 #define isUTF8_POSSIBLY_PROBLEMATIC(c) \ 1113 (NATIVE_UTF8_TO_I8(c) >= UTF_START_BYTE(UNICODE_SURROGATE_FIRST, 16)) 1114 1115 /* Perl extends Unicode so that it is possible to encode (as extended UTF-8 or 1116 * UTF-EBCDIC) any 64-bit value. No standard known to khw ever encoded higher 1117 * than a 31 bit value. On ASCII platforms this just meant arbitrarily saying 1118 * nothing could be higher than this. On these the start byte FD gets you to 1119 * 31 bits, and FE and FF are forbidden as start bytes. On EBCDIC platforms, 1120 * FD gets you only to 26 bits; adding FE to mean 7 total bytes gets you to 30 1121 * bits. To get to 31 bits, they treated an initial FF byte idiosyncratically. 1122 * It was considered to be the start byte FE meaning it had 7 total bytes, and 1123 * the final 1 was treated as an information bit, getting you to 31 bits. 1124 * 1125 * Perl used to accept this idiosyncratic interpretation of FF, but now rejects 1126 * it in order to get to being able to encode 64 bits. The bottom line is that 1127 * it is a Perl extension to use the start bytes FE and FF on ASCII platforms, 1128 * and the start byte FF on EBCDIC ones. That translates into that it is a 1129 * Perl extension to represent anything occupying more than 31 bits on ASCII 1130 * platforms; 30 bits on EBCDIC. */ 1131 #define UNICODE_IS_PERL_EXTENDED(uv) \ 1132 UNLIKELY((UV) (uv) > nBIT_UMAX(31 - ONE_IF_EBCDIC_ZERO_IF_NOT)) 1133 #define UTF8_IS_PERL_EXTENDED(s) \ 1134 (UTF8SKIP(s) > 6 + ONE_IF_EBCDIC_ZERO_IF_NOT) 1135 1136 /* Largest code point we accept from external sources */ 1137 #define MAX_LEGAL_CP ((UV)IV_MAX) 1138 1139 #define UTF8_ALLOW_EMPTY 0x0001 /* Allow a zero length string */ 1140 #define UTF8_GOT_EMPTY UTF8_ALLOW_EMPTY 1141 1142 /* Allow first byte to be a continuation byte */ 1143 #define UTF8_ALLOW_CONTINUATION 0x0002 1144 #define UTF8_GOT_CONTINUATION UTF8_ALLOW_CONTINUATION 1145 1146 /* Unexpected non-continuation byte */ 1147 #define UTF8_ALLOW_NON_CONTINUATION 0x0004 1148 #define UTF8_GOT_NON_CONTINUATION UTF8_ALLOW_NON_CONTINUATION 1149 1150 /* expecting more bytes than were available in the string */ 1151 #define UTF8_ALLOW_SHORT 0x0008 1152 #define UTF8_GOT_SHORT UTF8_ALLOW_SHORT 1153 1154 /* Overlong sequence; i.e., the code point can be specified in fewer bytes. 1155 * First one will convert the overlong to the REPLACEMENT CHARACTER; second 1156 * will return what the overlong evaluates to */ 1157 #define UTF8_ALLOW_LONG 0x0010 1158 #define UTF8_ALLOW_LONG_AND_ITS_VALUE (UTF8_ALLOW_LONG|0x0020) 1159 #define UTF8_GOT_LONG UTF8_ALLOW_LONG 1160 1161 #define UTF8_ALLOW_OVERFLOW 0x0080 1162 #define UTF8_GOT_OVERFLOW UTF8_ALLOW_OVERFLOW 1163 1164 #define UTF8_DISALLOW_SURROGATE 0x0100 /* Unicode surrogates */ 1165 #define UTF8_GOT_SURROGATE UTF8_DISALLOW_SURROGATE 1166 #define UTF8_WARN_SURROGATE 0x0200 1167 1168 /* Unicode non-character code points */ 1169 #define UTF8_DISALLOW_NONCHAR 0x0400 1170 #define UTF8_GOT_NONCHAR UTF8_DISALLOW_NONCHAR 1171 #define UTF8_WARN_NONCHAR 0x0800 1172 1173 /* Super-set of Unicode: code points above the legal max */ 1174 #define UTF8_DISALLOW_SUPER 0x1000 1175 #define UTF8_GOT_SUPER UTF8_DISALLOW_SUPER 1176 #define UTF8_WARN_SUPER 0x2000 1177 1178 /* The original UTF-8 standard did not define UTF-8 with start bytes of 0xFE or 1179 * 0xFF, though UTF-EBCDIC did. This allowed both versions to represent code 1180 * points up to 2 ** 31 - 1. Perl extends UTF-8 so that 0xFE and 0xFF are 1181 * usable on ASCII platforms, and 0xFF means something different than 1182 * UTF-EBCDIC defines. These changes allow code points of 64 bits (actually 1183 * somewhat more) to be represented on both platforms. But these are Perl 1184 * extensions, and not likely to be interchangeable with other languages. Note 1185 * that on ASCII platforms, FE overflows a signed 32-bit word, and FF an 1186 * unsigned one. */ 1187 #define UTF8_DISALLOW_PERL_EXTENDED 0x4000 1188 #define UTF8_GOT_PERL_EXTENDED UTF8_DISALLOW_PERL_EXTENDED 1189 #define UTF8_WARN_PERL_EXTENDED 0x8000 1190 1191 /* For back compat, these old names are misleading for overlongs and 1192 * UTF_EBCDIC. */ 1193 #define UTF8_DISALLOW_ABOVE_31_BIT UTF8_DISALLOW_PERL_EXTENDED 1194 #define UTF8_GOT_ABOVE_31_BIT UTF8_GOT_PERL_EXTENDED 1195 #define UTF8_WARN_ABOVE_31_BIT UTF8_WARN_PERL_EXTENDED 1196 #define UTF8_DISALLOW_FE_FF UTF8_DISALLOW_PERL_EXTENDED 1197 #define UTF8_WARN_FE_FF UTF8_WARN_PERL_EXTENDED 1198 1199 #define UTF8_CHECK_ONLY 0x10000 1200 #define _UTF8_NO_CONFIDENCE_IN_CURLEN 0x20000 /* Internal core use only */ 1201 1202 /* For backwards source compatibility. They do nothing, as the default now 1203 * includes what they used to mean. The first one's meaning was to allow the 1204 * just the single non-character 0xFFFF */ 1205 #define UTF8_ALLOW_FFFF 0 1206 #define UTF8_ALLOW_FE_FF 0 1207 #define UTF8_ALLOW_SURROGATE 0 1208 1209 /* C9 refers to Unicode Corrigendum #9: allows but discourages non-chars */ 1210 #define UTF8_DISALLOW_ILLEGAL_C9_INTERCHANGE \ 1211 (UTF8_DISALLOW_SUPER|UTF8_DISALLOW_SURROGATE) 1212 #define UTF8_WARN_ILLEGAL_C9_INTERCHANGE (UTF8_WARN_SUPER|UTF8_WARN_SURROGATE) 1213 1214 #define UTF8_DISALLOW_ILLEGAL_INTERCHANGE \ 1215 (UTF8_DISALLOW_ILLEGAL_C9_INTERCHANGE|UTF8_DISALLOW_NONCHAR) 1216 #define UTF8_WARN_ILLEGAL_INTERCHANGE \ 1217 (UTF8_WARN_ILLEGAL_C9_INTERCHANGE|UTF8_WARN_NONCHAR) 1218 1219 /* This is typically used for code that processes UTF-8 input and doesn't want 1220 * to have to deal with any malformations that might be present. All such will 1221 * be safely replaced by the REPLACEMENT CHARACTER, unless other flags 1222 * overriding this are also present. */ 1223 #define UTF8_ALLOW_ANY ( UTF8_ALLOW_CONTINUATION \ 1224 |UTF8_ALLOW_NON_CONTINUATION \ 1225 |UTF8_ALLOW_SHORT \ 1226 |UTF8_ALLOW_LONG \ 1227 |UTF8_ALLOW_OVERFLOW) 1228 1229 /* Accept any Perl-extended UTF-8 that evaluates to any UV on the platform, but 1230 * not any malformed. This is the default. */ 1231 #define UTF8_ALLOW_ANYUV 0 1232 #define UTF8_ALLOW_DEFAULT UTF8_ALLOW_ANYUV 1233 1234 #define UNICODE_WARN_SURROGATE 0x0001 /* UTF-16 surrogates */ 1235 #define UNICODE_WARN_NONCHAR 0x0002 /* Non-char code points */ 1236 #define UNICODE_WARN_SUPER 0x0004 /* Above 0x10FFFF */ 1237 #define UNICODE_WARN_PERL_EXTENDED 0x0008 /* Above 0x7FFF_FFFF */ 1238 #define UNICODE_WARN_ABOVE_31_BIT UNICODE_WARN_PERL_EXTENDED 1239 #define UNICODE_DISALLOW_SURROGATE 0x0010 1240 #define UNICODE_DISALLOW_NONCHAR 0x0020 1241 #define UNICODE_DISALLOW_SUPER 0x0040 1242 #define UNICODE_DISALLOW_PERL_EXTENDED 0x0080 1243 1244 #ifdef PERL_CORE 1245 # define UNICODE_ALLOW_ABOVE_IV_MAX 0x0100 1246 #endif 1247 #define UNICODE_DISALLOW_ABOVE_31_BIT UNICODE_DISALLOW_PERL_EXTENDED 1248 1249 #define UNICODE_GOT_SURROGATE UNICODE_DISALLOW_SURROGATE 1250 #define UNICODE_GOT_NONCHAR UNICODE_DISALLOW_NONCHAR 1251 #define UNICODE_GOT_SUPER UNICODE_DISALLOW_SUPER 1252 #define UNICODE_GOT_PERL_EXTENDED UNICODE_DISALLOW_PERL_EXTENDED 1253 1254 #define UNICODE_WARN_ILLEGAL_C9_INTERCHANGE \ 1255 (UNICODE_WARN_SURROGATE|UNICODE_WARN_SUPER) 1256 #define UNICODE_WARN_ILLEGAL_INTERCHANGE \ 1257 (UNICODE_WARN_ILLEGAL_C9_INTERCHANGE|UNICODE_WARN_NONCHAR) 1258 #define UNICODE_DISALLOW_ILLEGAL_C9_INTERCHANGE \ 1259 (UNICODE_DISALLOW_SURROGATE|UNICODE_DISALLOW_SUPER) 1260 #define UNICODE_DISALLOW_ILLEGAL_INTERCHANGE \ 1261 (UNICODE_DISALLOW_ILLEGAL_C9_INTERCHANGE|UNICODE_DISALLOW_NONCHAR) 1262 1263 /* For backward source compatibility, as are now the default */ 1264 #define UNICODE_ALLOW_SURROGATE 0 1265 #define UNICODE_ALLOW_SUPER 0 1266 #define UNICODE_ALLOW_ANY 0 1267 1268 #define UNICODE_BYTE_ORDER_MARK 0xFEFF 1269 #define UNICODE_IS_BYTE_ORDER_MARK(uv) UNLIKELY((UV) (uv) \ 1270 == UNICODE_BYTE_ORDER_MARK) 1271 1272 #define LATIN_SMALL_LETTER_SHARP_S LATIN_SMALL_LETTER_SHARP_S_NATIVE 1273 #define LATIN_SMALL_LETTER_Y_WITH_DIAERESIS \ 1274 LATIN_SMALL_LETTER_Y_WITH_DIAERESIS_NATIVE 1275 #define MICRO_SIGN MICRO_SIGN_NATIVE 1276 #define LATIN_CAPITAL_LETTER_A_WITH_RING_ABOVE \ 1277 LATIN_CAPITAL_LETTER_A_WITH_RING_ABOVE_NATIVE 1278 #define LATIN_SMALL_LETTER_A_WITH_RING_ABOVE \ 1279 LATIN_SMALL_LETTER_A_WITH_RING_ABOVE_NATIVE 1280 #define UNICODE_GREEK_CAPITAL_LETTER_SIGMA 0x03A3 1281 #define UNICODE_GREEK_SMALL_LETTER_FINAL_SIGMA 0x03C2 1282 #define UNICODE_GREEK_SMALL_LETTER_SIGMA 0x03C3 1283 #define GREEK_SMALL_LETTER_MU 0x03BC 1284 #define GREEK_CAPITAL_LETTER_MU 0x039C /* Upper and title case 1285 of MICRON */ 1286 #define LATIN_CAPITAL_LETTER_Y_WITH_DIAERESIS 0x0178 /* Also is title case */ 1287 #ifdef LATIN_CAPITAL_LETTER_SHARP_S_UTF8 1288 # define LATIN_CAPITAL_LETTER_SHARP_S 0x1E9E 1289 #endif 1290 #define LATIN_CAPITAL_LETTER_I_WITH_DOT_ABOVE 0x130 1291 #define LATIN_SMALL_LETTER_DOTLESS_I 0x131 1292 #define LATIN_SMALL_LETTER_LONG_S 0x017F 1293 #define LATIN_SMALL_LIGATURE_LONG_S_T 0xFB05 1294 #define LATIN_SMALL_LIGATURE_ST 0xFB06 1295 #define KELVIN_SIGN 0x212A 1296 #define ANGSTROM_SIGN 0x212B 1297 1298 #define UNI_DISPLAY_ISPRINT 0x0001 1299 #define UNI_DISPLAY_BACKSLASH 0x0002 1300 #define UNI_DISPLAY_BACKSPACE 0x0004 /* Allow \b when also 1301 UNI_DISPLAY_BACKSLASH */ 1302 #define UNI_DISPLAY_QQ (UNI_DISPLAY_ISPRINT \ 1303 |UNI_DISPLAY_BACKSLASH \ 1304 |UNI_DISPLAY_BACKSPACE) 1305 1306 /* Character classes could also allow \b, but not patterns in general */ 1307 #define UNI_DISPLAY_REGEX (UNI_DISPLAY_ISPRINT|UNI_DISPLAY_BACKSLASH) 1308 1309 /* Should be removed; maybe deprecated, but not used in CPAN */ 1310 #define SHARP_S_SKIP 2 1311 1312 #define is_utf8_char_buf(buf, buf_end) isUTF8_CHAR(buf, buf_end) 1313 #define bytes_from_utf8(s, lenp, is_utf8p) \ 1314 bytes_from_utf8_loc(s, lenp, is_utf8p, 0) 1315 1316 /* Do not use; should be deprecated. Use isUTF8_CHAR() instead; this is 1317 * retained solely for backwards compatibility */ 1318 #define IS_UTF8_CHAR(p, n) (isUTF8_CHAR(p, (p) + (n)) == n) 1319 1320 #endif /* PERL_UTF8_H_ */ 1321 1322 /* 1323 * ex: set ts=8 sts=4 sw=4 et: 1324 */ 1325