You can create a simple parser by first tokenizing your input with re
:
import re
class Token:
def __init__(self, _t, val):
self._type, self.val = _t, val
def __repr__(self):
return f'{self.__class__.__name__}({self._type}, {self.val})'
class Tokenize:
gram, _t = r'neg|or|imp|and|iff|\(|\)|\w+', [(r'neg|or|imp|and|iff', 'func'), (r'\(', 'oparen'), (r'\)', 'cparen'), (r'\w+', 'value')]
@classmethod
def tokenize(cls, _input):
return [Token([b for a, b in cls._t if re.findall(a, i)][0], i) for i in re.findall(cls.gram, _input)]
def parse(d, stop=None):
s = next(d, None)
if s is None or s._type == stop:
return ''
if s._type == 'func':
return f'{s.val}({parse(d, stop=stop)})'
if s._type == 'oparen':
s = parse(d, stop='cparen')
_n = next(d, None)
if _n and _n._type == stop:
return getattr(s, 'val', s)
return getattr(s, 'val', s) if _n is None else f'{_n.val}({getattr(s, "val", s)}, {parse(d, stop=stop)})'
n = ['[neg (p or q)]', '[p imp q, (neg r) imp (neg q)]', '[(p and q) and r]', '[neg (p iff (neg q))]']
result = [[parse(iter(Tokenize.tokenize(i))) for i in a[1:-1].split(',')] for a in n]
Output:
[['neg(or(p, q))'], ['imp(p, q)', 'imp(neg(r), neg(q))'], ['and(and(p, q), r)'], ['neg(iff(p, neg(q)))']]