swh:1:snp:7ce5f1105410d5ee1ad6abfdc873986c25b579e5
Raw File
Tip revision: 4b3e10665499f7ebd8ec2bf073dfbca000eaa9ba authored by Dirk Roorda on 10 July 2018, 08:50:53 UTC
fixes
Tip revision: 4b3e106
transcription.py
import re


class Transcription(object):
  decomp = {
      '\u05E9\u05C1': "\uFB2A",
      '\u05E9\u05C2': "\uFB2B",
  }
  hebrew_mapping = {
      '_': ' ',  # space inside word
      '92': "\u0591",  # etnahta = atnach
      '01': "\u0592",  # segolta
      '65': "\u0593",  # shalshelet
      '80': "\u0594",  # zaqef_qatan
      '85': "\u0595",  # zaqef_gadol
      '73': "\u0596",  # tipeha = tifcha
      '81': "\u0597",  # revia = rebia
      '82': "\u0598",  # zarqa = tsinorit = zinorit = sinnorit
      '03': "\u0599",  # pashta
      '10': "\u059A",  # yetiv = yetib
      '91': "\u059B",  # tevir = tebir
      '61': "\u059C",  # geresh
      '11': "\u059D",  # geresh muqdam = mugrash
      '62': "\u059E",  # gershayim = garshayim
      '84': "\u059F",  # qarney para = pazer_gadol
      '14': "\u05A0",  # telisha_gedola
      '44': "\u05A0",  # telisha_gedola = telisha_gedola_med
      '83': "\u05A1",  # pazer
      '74': "\u05A3",  # munah = munach
      '70': "\u05A4",  # mahapakh = mehuppach
      '71': "\u05A5",  # merkha = merecha
      '72': "\u05A6",  # merkha kefula = merecha_kepula
      '94': "\u05A7",  # darga
      '63': "\u05A8",  # qadma = azla
      '33': "\u05A8",  # pashta_med < qadma
      '04': "\u05A9",  # telisha_qetana
      '24': "\u05A9",  # telisha_qetana = telisha_qetana_med
      '93': "\u05AA",  # yera ben yomo = yerach
      '60': "\u05AB",  # ole = ole_weyored
      '64': "\u05AC",  # iluy = illuy
      '13': "\u05AD",  # dehi = dechi
      '02': "\u05AE",  # zinor = sinnor
      '*': "\u05AF",  # masora circle
      ':': "\u05B0",  # sheva = shewa
      ':E': "\u05B1",  # hataf segol = chataph_segol
      ':A': "\u05B2",  # hataf patah = chataph_patach
      ':@': "\u05B3",  # hataf qamats = chataph_qamats
      'I': "\u05B4",  # hiriq = chiriq
      ';': "\u05B5",  # tsere
      'E': "\u05B6",  # segol
      'A': "\u05B7",  # patach
      '@': "\u05B8",  # qamats
      'O': "\u05B9",  # holam = cholam
      'U': "\u05BB",  # qubuts = qubbuts
      '.': "\u05BC",  # dagesh
      '25': "\u05BD",  # silluq yamin
      '45': "\u05BD",  # meteg
      '35': "\u05BD",  # meteg (tikon)
      '75': "\u05BD",  # siluq = silluq
      '95': "\u05BD",  # meteg = meteg_yamin
      '&': "\u05BE",  # maqaf
      ',': "\u05BF",  # rafe = raphe
      '05': "\u05C0",  # paseq
      '.c': "\u05C1",  # shin dot
      '.f': "\u05C2",  # sin dot
      '00': "\u05C3",  # sof_pasuq
      '52': "\u05C4",  # upper dot = puncta_above
      '53': "\u05C5",  # lower dot = puncta_below
      'ñ': "\u05C6\u0307",  # nun hafukha
      'Ñ': "\u05C6\u0307",  # nun hafukha
      '>': "\u05D0",  # alef
      'B': "\u05D1",  # bet
      'G': "\u05D2",  # gimel
      'D': "\u05D3",  # dalet
      'H': "\u05D4",  # he
      'W': "\u05D5",  # vav
      'Z': "\u05D6",  # zayin
      'X': "\u05D7",  # het
      'V': "\u05D8",  # tet
      'J': "\u05D9",  # yod
      'k': "\u05DA",  # kaf final
      'K': "\u05DB",  # kaf
      'L': "\u05DC",  # lamed
      'm': "\u05DD",  # mem final
      'M': "\u05DE",  # mem
      'n': "\u05DF",  # nun final
      'N': "\u05E0",  # nun
      'S': "\u05E1",  # samekh
      '<': "\u05E2",  # ayin
      'p': "\u05E3",  # pe final
      'P': "\u05E4",  # pe
      'y': "\u05E5",  # tsadi final
      'Y': "\u05E6",  # tsadi
      'Q': "\u05E7",  # qof
      'R': "\u05E8",  # resh
      '#': "\u05E9",  # sin unpointed
      'T': "\u05EA",  # tav
      'C': "\uFB2A",  # shin pointed
      'F': "\uFB2B",  # sin pointed
      '55': "<UNMAPPED 55=large letter>",  # large_letter
      '56': "<UNMAPPED 56=small letter>",  # small_letter
      '57': "<UNMAPPED 57=suspended letter>",  # suspended_letter
      '-': "",  # suppress space afterward
  }
  hebrew_cons = '>BGDHWZXVJKLMNS<PYQRFCT'
  trans_final_pat = re.compile(
      r'([' + hebrew_cons + r'][^_&]*)([KMNPY])([^' + hebrew_cons + r'_&]*(:?[_&]|\Z))'
  )
  trans_hebrew_pat = re.compile(r'(:[AE@]|.[cf]|:|[0-9][0-9]|.)')
  swap_accent_pat = re.compile(r'(\A|[_&])([0-9][0-9])([' + hebrew_cons + r'])([:;,.EAIOU@]*)')
  remove_accent_pat = re.compile(r'((?:[0-9][0-9])|[,*])')
  remove_point_pat = re.compile(r'((?:[0-9][0-9])|(?:\.[cf])|(?::[@AE])|[,.:;@AEIOU*])')
  remove_psn_pat = re.compile(r'00[ _SPNÑñ]*')
  remove_psq_pat = re.compile(r'(?:[ _]+05[ _]*)|(?:05[ _]+)')
  shin_pat = re.compile(r'[CF]')
  ph_simple_pat = re.compile(r'([ˈˌᵊᵃᵒᵉāo*])')
  noorigspace = re.compile(
      '''
          (?: [&-]\Z)           # space, maqef or nospace
        | (?:
               0[05]            # sof pasuq or paseq
               (?:_[SNP])*      # nun hafukha, setumah, petuhah at end of verse
               \Z
          )
        | (?:_[SPN])+           #  nun hafukha, setumah, petuhah between words
    ''', re.X
  )

  syriac_mapping = {
      '>': "\u0710",  # alaph
      'B': "\u0712",  # beth
      'G': "\u0713",  # gamal
      'D': "\u0715",  # dalat
      'H': "\u0717",  # he
      'W': "\u0718",  # waw
      'Z': "\u0719",  # zain
      'X': "\u071A",  # heth
      'V': "\u071B",  # teth
      'J': "\u071D",  # yudh
      'K': "\u071F",  # kaph
      'L': "\u0720",  # lamadh
      'M': "\u0721",  # mim
      'N': "\u0722",  # nun
      'S': "\u0723",  # semkath
      '<': "\u0725",  # e
      'P': "\u0726",  # pe
      'Y': "\u0728",  # sadhe
      'Q': "\u0729",  # qaph
      'R': "\u072A",  # rish
      'C': "\u072B",  # shin
      'T': "\u072C",  # taw
      's': "\u0724",  # semkath final
      'p': "\u0727",  # pe reversed
  }

  def __init__(self):
    self.hebrew_consonants = {Transcription.hebrew_mapping[x] for x in Transcription.hebrew_cons}
    self.hebrew_consonants.add('\u05E9')
    self.hebrew_mappingi = dict((v, k)
                                for (k, v) in Transcription.hebrew_mapping.items()
                                if k != '')
    # special treatment needed for nun hafukha,
    # since it is consists of two characters
    self.hebrew_mappingi['\u05C6'] = 'ñ'
    self.hebrew_mappingi['\u0307'] = ''

  def _comp(s):
    for (d, c) in Transcription.decomp.items():
      s = s.replace(d, c)
    return s

  def _decomp(s):
    for (d, c) in Transcription.decomp.items():
      s = s.replace(c, d)
    return s

  def suffix_and_finales(word):
    # first split the word proper from the suffix,
    # and add a space if there is no other suffix
    add_space = ''
    suffix = ''
    new_word = word
    if not word:
      return (new_word, suffix + add_space)
    lastch = new_word[-1]
    if lastch == '-' or lastch == '&':
      new_word = new_word[0:-1]
      suffix = lastch
    else:
      if len(new_word) >= 2:
        lastch = new_word[-1]
        llastch = new_word[-2]
        if llastch == '_' and (lastch == 'P' or lastch == 'S'):
          new_word = new_word[0:-2]
          suffix = ' ' + lastch + suffix + ' '
      if len(new_word) >= 2:
        lastch = new_word[-1]
        llastch = new_word[-2]
        if llastch == '_' and (lastch == 'N'):
          new_word = new_word[0:-2]
          suffix = ' ñ' + suffix + ' '
      if len(new_word) >= 2:
        lastch = new_word[-1]
        llastch = new_word[-2]
        if llastch == '0' and (lastch == '0' or lastch == '5'):
          new_word = new_word[0:-2]
          suffix = (' ' if lastch == '5' else '') + llastch + lastch + suffix
          add_space = '\n' if lastch == '0' else ' '
    if suffix == '':
      add_space = ' '
    elif suffix == '-':
      add_space = ''
      suffix = ''
    # second: replace consonants by their final forms when needed
    new_word = Transcription.trans_final_pat.sub(Transcription._map_final, new_word)
    return (new_word, suffix + add_space)

  def _map_final(m):
    return m.group(1) + m.group(2).lower() + m.group(3)

  def _map_hebrew(m):
    return Transcription.hebrew_mapping.get(m.group(1), m.group(1))

  def _swap_accent(m):
    return m.group(1) + m.group(3) + m.group(4) + m.group(2)

  def _remove_accent(m):
    return '00' if m.group(1) == '00' else '05' if m.group(1) == '05' else ''

  def _remove_point(m):
    return '00' if m.group(1) == '00' else '05' if m.group(1) == '05' else ''

  def _ph_simple(m):
    return 'å' if m.group(1) in 'āo' else ''

  # unicode normalization is harmful
  # if there is a combination of dagesh, vowel and accent.

  def suppress_space(word):
    return Transcription.noorigspace.search(word)

  def to_etcbc_v(word):
    return Transcription.remove_accent_pat.sub(Transcription._remove_accent, word)

  def to_etcbc_c(word):
    word = Transcription.remove_point_pat.sub(Transcription._remove_point, word)
    word = Transcription.remove_psn_pat.sub(
        '00', word
    )  # remove nun hafukha, setumah, petuhah at the end of a verse
    word = Transcription.remove_psq_pat.sub(
        ' ', word
    )  # replace paseq with attached spaces by single space
    word = word.upper()  # no final forms of consonants
    return Transcription.shin_pat.sub('#', word)

  def to_hebrew(word):
    word = Transcription.swap_accent_pat.sub(Transcription._swap_accent, word)
    return Transcription.trans_hebrew_pat.sub(Transcription._map_hebrew, word)

  def to_hebrew_v(word):
    return Transcription.trans_hebrew_pat.sub(
        Transcription._map_hebrew, Transcription.to_etcbc_v(word)
    )

  def to_hebrew_c(word):
    return Transcription.trans_hebrew_pat.sub(
        Transcription._map_hebrew, Transcription.to_etcbc_c(word)
    )

  def to_hebrew_x(word):
    return Transcription.trans_hebrew_pat.sub(Transcription._map_hebrew, word)

  def ph_simplify(pword):
    return Transcription.ph_simple_pat.sub(Transcription._ph_simple, pword)

  def from_hebrew(self, word):
    return ''.join(self.hebrew_mappingi.get(x, x) for x in Transcription._comp(word))

  def to_syriac(self, word):
    return Transcription._decomp(
        ''.join(self.syriac_mapping.get(x, x) for x in Transcription._comp(word))
    )

  def from_syriac(self, word):
    return ''.join(self.syriac_mappingi.get(x, x) for x in Transcription._comp(word))
back to top