#!/usr/bin/env python3 # SPDX-License-Identifier: GPL-2.0 # # Leverage Python's unidecode module to generate ucs_fallback_table.h # # The generated table maps complex characters to their simpler fallback forms # for a terminal display when corresponding glyphs are unavailable. # # Usage: # python3 gen_ucs_fallback_table.py # Generate fallback tables # python3 gen_ucs_fallback_table.py -o FILE # Specify output file import unicodedata from unidecode import unidecode import sys import argparse from collections import defaultdict # Try to get unidecode version try: from importlib.metadata import version unidecode_version = version('unidecode') except: unidecode_version = 'unknown' # This script's file name from pathlib import Path this_file = Path(__file__).name # Default output file name DEFAULT_OUT_FILE = "ucs_fallback_table.h" # Define the range marker value RANGE_MARKER = 0x00 def generate_fallback_map(): """Generate a fallback map using unidecode for all relevant Unicode points.""" fallback_map = {} # Process BMP characters (0x0000 - 0xFFFF) to keep table size manageable for cp in range(0x0080, 0x10000): # Skip ASCII range (0x00-0x7F) char = chr(cp) # Skip unassigned/control characters try: if not unicodedata.name(char, ''): continue except ValueError: continue # Get the unidecode transliteration ascii_version = unidecode(char) # Only store if it results in a single character mapping if len(ascii_version) == 1: fallback_map[cp] = ord(ascii_version) # Apply manual overrides for special cases fallback_map.update(get_special_overrides()) return fallback_map def get_special_overrides(): """Get special case overrides that need different handling than unidecode provides... or doesn't provide at all.""" overrides = {} # Multi-character unidecode output # These map to single chars instead of unidecode's multiple-char mappings # In a terminal fallback context, we need a single character rather than multiple overrides[0x00C6] = ord('E') # Æ LATIN CAPITAL LETTER AE -> E (unidecode: "AE") overrides[0x00E6] = ord('e') # æ LATIN SMALL LETTER AE -> e (unidecode: "ae") overrides[0x0152] = ord('E') # Œ LATIN CAPITAL LIGATURE OE -> E (unidecode: "OE") overrides[0x0153] = ord('e') # œ LATIN SMALL LETTER LIGATURE OE -> e (unidecode: "oe") overrides[0x00DF] = ord('s') # ß LATIN SMALL LETTER SHARP S -> s (unidecode: "ss") # Comparison operators that unidecode renders as multiple characters overrides[0x2264] = ord('<') # ≤ LESS-THAN OR EQUAL TO -> < (unidecode: "<=") overrides[0x2265] = ord('>') # ≥ GREATER-THAN OR EQUAL TO -> > (unidecode: ">=") # Unidecode returns an empty string for these overrides[0x2260] = ord('#') # ≠ NOT EQUAL TO -> # (unidecode: empty string) # Quadrant block characters that unidecode doesn't map for cp in range(0x2596, 0x259F+1): overrides[cp] = ord('#') # ▖ ▗ ▘ ▙ etc. - map to # (unidecode: empty string) # Directional arrows # These provide better semantic meaning than unidecode's mappings overrides[0x2192] = ord('>') # → RIGHTWARDS ARROW -> > (unidecode: "-") overrides[0x2190] = ord('<') # ← LEFTWARDS ARROW -> < (unidecode: "-") overrides[0x2191] = ord('^') # ↑ UPWARDS ARROW -> ^ (unidecode: "|") overrides[0x2193] = ord('v') # ↓ DOWNWARDS ARROW -> v (unidecode: "|") # Double arrows with their directional semantic mappings overrides[0x21D0] = ord('<') # ⇐ LEFTWARDS DOUBLE ARROW -> < overrides[0x21D1] = ord('^') # ⇑ UPWARDS DOUBLE ARROW -> ^ overrides[0x21D2] = ord('>') # ⇒ RIGHTWARDS DOUBLE ARROW -> > overrides[0x21D3] = ord('v') # ⇓ DOWNWARDS DOUBLE ARROW -> v # Halfwidth arrows # These need the same treatment as their normal-width counterparts overrides[0xFFE9] = ord('<') # ← HALFWIDTH LEFTWARDS ARROW -> < (unidecode: "-") overrides[0xFFEA] = ord('^') # ↑ HALFWIDTH UPWARDS ARROW -> ^ (unidecode: "|") overrides[0xFFEB] = ord('>') # → HALFWIDTH RIGHTWARDS ARROW -> > (unidecode: "-") overrides[0xFFEC] = ord('v') # ↓ HALFWIDTH DOWNWARDS ARROW -> v (unidecode: "|") # Currency symbols - each mapped to a representative letter overrides[0x00A2] = ord('c') # ¢ CENT SIGN -> c overrides[0x00A3] = ord('L') # £ POUND SIGN -> L overrides[0x00A5] = ord('Y') # ¥ YEN SIGN -> Y overrides[0x20AC] = ord('E') # € EURO SIGN -> E # Symbols mapped to letters overrides[0x00A7] = ord('S') # § SECTION SIGN -> S overrides[0x00A9] = ord('C') # © COPYRIGHT SIGN -> C overrides[0x00AE] = ord('R') # ® REGISTERED SIGN -> R overrides[0x2122] = ord('T') # ™ TRADE MARK SIGN -> T # Degree-related symbols overrides[0x00B0] = ord('o') # ° DEGREE SIGN -> o overrides[0x2103] = ord('C') # ℃ DEGREE CELSIUS -> C overrides[0x2109] = ord('F') # ℉ DEGREE FAHRENHEIT -> F # Angle quotation marks overrides[0x00AB] = ord('<') # « LEFT-POINTING DOUBLE ANGLE QUOTATION MARK -> < overrides[0x00BB] = ord('>') # » RIGHT-POINTING DOUBLE ANGLE QUOTATION MARK -> > # Operators with circular shape overrides[0x2218] = ord('o') # ∘ RING OPERATOR -> o overrides[0x2219] = ord('.') # ∙ BULLET OPERATOR -> . # Negated mathematical symbols (preserving the negation semantics) # Negated symbols mapped to exclamation mark (semantically "not") for cp in (0x2204, 0x2209, 0x220C, 0x2224, 0x2226, 0x226E, 0x226F, 0x2280, 0x2281, 0x2284, 0x2285): overrides[cp] = ord('!') # Negated math symbols -> ! (not) # Negated symbols mapped to hash sign (semantically "not equal") for cp in (0x2241, 0x2244, 0x2249, 0x2262, 0x2268, 0x2269, 0x226D, 0x228A, 0x228B): overrides[cp] = ord('#') # Negated equality symbols -> # (not equal) # Negated arrows - all mapped to exclamation mark for cp in (0x219A, 0x219B, 0x21AE, 0x21CD, 0x21CE, 0x21CF): overrides[cp] = ord('!') # Negated arrows -> ! (not) # Dashes and hyphens for cp in (0x2010, 0x2011, 0x2012, 0x2013, 0x2014, 0x2015, 0x2043, 0x2052): overrides[cp] = ord('-') # Dashes and hyphens -> - # Question mark punctuation for cp in (0x203D, 0x2047, 0x2048): overrides[cp] = ord('?') # Question marks -> ? # Exclamation mark punctuation for cp in (0x203C, 0x2049): overrides[cp] = ord('!') # Exclamation marks -> ! # Asterisk-like symbols for cp in (0x2042, 0x2051, 0x2055): overrides[cp] = ord('*') # Other specific punctuation with unique mappings overrides[0x201E] = ord('"') # „ DOUBLE LOW-9 QUOTATION MARK overrides[0x2023] = ord('>') # ‣ TRIANGULAR BULLET overrides[0x2026] = ord('.') # … HORIZONTAL ELLIPSIS overrides[0x2033] = ord('"') # ″ DOUBLE PRIME overrides[0x204B] = ord('P') # ⁋ REVERSED PILCROW SIGN overrides[0x204C] = ord('<') # ⁌ BLACK LEFTWARDS BULLET overrides[0x204D] = ord('>') # ⁍ BLACK RIGHTWARDS BULLET overrides[0x204F] = ord(';') # ⁏ REVERSED SEMICOLON overrides[0x205B] = ord(':') # ⁛ FOUR DOT MARK # Check marks overrides[0x2713] = ord('v') # ✓ CHECK MARK overrides[0x2714] = ord('V') # ✔ HEAVY CHECK MARK # X marks - lowercase for regular, uppercase for heavy for cp in (0x2715, 0x2717): overrides[cp] = ord('x') # Regular X marks -> x for cp in (0x2716, 0x2718): overrides[cp] = ord('X') # Heavy X marks -> X # Stars and asterisk-like symbols mapped to '*' for cp in (0x2605, 0x2606, 0x262A, 0x269D, 0x2698): overrides[cp] = ord('*') # All star and asterisk symbols -> * for cp in range(0x2721, 0x2746+1): overrides[cp] = ord('*') # All star and asterisk symbols -> * for cp in range(0x2749, 0x274B+1): overrides[cp] = ord('*') # Last set of asterisk symbols -> * for cp in (0x229B, 0x22C6, 0x235F, 0x2363): overrides[cp] = ord('*') # Star operators -> * # Special exclusions with fallback value of 0 # These will be filtered out in organize_by_pages() # Exclude U+2028 (LINE SEPARATOR) overrides[0x2028] = 0 # LINE SEPARATOR (unidecode: '\n') # Full-width to ASCII mapping (covering all printable ASCII 33-126) # 0xFF01 (!) to 0xFF5E (~) -> ASCII 33 (!) to 126 (~) # Those are excluded here to reduce the table size. # It is more efficient to process them programmatically in # ucs.c:ucs_get_fallback(). for cp in range(0xFF01, 0xFF5E + 1): overrides[cp] = 0 # Double-width ASCII characters return overrides def organize_by_pages(fallback_map): """Organize the fallback mappings by their high byte (page).""" # Group by high byte (page) page_groups = defaultdict(list) for code, fallback in fallback_map.items(): # Skip characters with fallback value of 0 (excluded characters) if fallback == 0: continue page = code >> 8 # Get the high byte (page) offset = code & 0xFF # Get the low byte (offset within page) page_groups[page].append((offset, fallback)) # Sort each page's entries by offset for page in page_groups: page_groups[page].sort() return page_groups def compress_ranges(page_groups): """Compress consecutive entries with the same fallback character into ranges. A range is only compressed if it contains 3 or more consecutive entries.""" compressed_pages = {} for page, entries in page_groups.items(): compressed_entries = [] i = 0 while i < len(entries): start_offset, fallback = entries[i] # Look ahead to find consecutive entries with the same fallback j = i + 1 while (j < len(entries) and entries[j][0] == entries[j-1][0] + 1 and # consecutive offsets entries[j][1] == fallback): # same fallback j += 1 # Calculate the range end end_offset = entries[j-1][0] # If we found a range with 3 or more entries (worth compressing) if j - i >= 3: # Add a range entry compressed_entries.append((start_offset, RANGE_MARKER)) compressed_entries.append((end_offset, fallback)) else: # Add the individual entries as is for k in range(i, j): compressed_entries.append(entries[k]) i = j compressed_pages[page] = compressed_entries return compressed_pages def cp_name(cp): """Get the Unicode character name for a code point.""" try: return unicodedata.name(chr(cp)) except: return f"U+{cp:04X}" def generate_fallback_tables(out_file=DEFAULT_OUT_FILE): """Generate the fallback character tables.""" # Generate fallback map using unidecode fallback_map = generate_fallback_map() print(f"Generated {len(fallback_map)} total fallback mappings") # Organize by pages page_groups = organize_by_pages(fallback_map) print(f"Organized into {len(page_groups)} pages") # Compress ranges compressed_pages = compress_ranges(page_groups) total_compressed_entries = sum(len(entries) for entries in compressed_pages.values()) print(f"Total compressed entries: {total_compressed_entries}") # Create output file with open(out_file, 'w') as f: f.write(f"""\ /* SPDX-License-Identifier: GPL-2.0 */ /* * {out_file} - Unicode character fallback table * * Auto-generated by {this_file} * * Unicode Version: {unicodedata.unidata_version} * Unidecode Version: {unidecode_version} * * This file contains optimized tables that map complex Unicode characters * to simpler fallback characters for terminal display when corresponding * glyphs are unavailable. */ static const struct ucs_page_desc ucs_fallback_pages[] = {{ """) # Convert compressed_pages to a sorted list of (page, entries) tuples sorted_pages = sorted(compressed_pages.items()) # Track the start index for each page start_index = 0 # Write page descriptors for page, entries in sorted_pages: count = len(entries) f.write(f"\t{{ 0x{page:02X}, {count}, {start_index} }},\n") start_index += count # Write entries array f.write("""\ }; /* Page entries array (referenced by page descriptors) */ static const struct ucs_page_entry ucs_fallback_entries[] = { """) # Write all entries for page, entries in sorted_pages: page_hex = f"0x{page:02X}" f.write(f"\t/* Entries for page {page_hex} */\n") for i, (offset, fallback) in enumerate(entries): # Convert to hex for better readability offset_hex = f"0x{offset:02X}" fallback_hex = f"0x{fallback:02X}" # Handle comments codepoint = (page << 8) | offset if fallback == RANGE_MARKER: comment = f"{cp_name(codepoint)} -> ..." else: comment = f"{cp_name(codepoint)} -> '{chr(fallback)}'" f.write(f"\t{{ 0x{offset:02X}, 0x{fallback:02X} }}, /* {comment} */\n") f.write(f"""\ }}; #define UCS_PAGE_ENTRY_RANGE_MARKER {RANGE_MARKER} """) if __name__ == "__main__": parser = argparse.ArgumentParser(description="Generate Unicode fallback character tables") parser.add_argument("-o", "--output", dest="output_file", default=DEFAULT_OUT_FILE, help=f"Output file name (default: {DEFAULT_OUT_FILE})") args = parser.parse_args() generate_fallback_tables(out_file=args.output_file)