| import copy |
|
|
|
|
| class ChatTemplate: |
| cache = {} |
| roles = set() |
|
|
| def __init__(self, model, im_start=r'<|im_start|>', im_end=r'<|im_end|>', nl='\n'): |
| self.model = model |
| self.nl = nl |
| self.im_start = im_start |
| self.im_start_token = model.tokenize(self.im_start.encode('utf-8'), add_bos=False, special=True) |
| self.im_end = im_end |
| self.im_end_nl = model.tokenize((self.im_end + self.nl).encode('utf-8'), add_bos=False, special=True) |
| self.eos = [model._token_eos, self.im_end_nl[0]] |
| self.onenl = [self.im_end_nl[-1]] |
| tmp = model.tokenize(('\r' + self.nl).encode('utf-8'), add_bos=False, special=True) |
| if len(tmp) == 1: |
| self.onenl.append(tmp[0]) |
| self.onerl = model.tokenize(b'\r', add_bos=False, special=True) |
| self.nlnl = None |
| tmp = model.tokenize((self.nl + self.nl).encode('utf-8'), add_bos=False, special=True) |
| if len(tmp) == 1: |
| self.nlnl = tmp[0] |
| print('ChatTemplate', self.eos, self.im_end_nl, self.onerl, self.onenl, self.nlnl) |
|
|
| def _get(self, key: str): |
| if key in self.cache: |
| return copy.deepcopy(self.cache[key]) |
| else: |
| value = self.model.tokenize((self.im_start + key + self.nl).encode('utf-8'), add_bos=False, special=True) |
| self.cache[key] = copy.deepcopy(value) |
| return value |
|
|
| def _add_role(self, _role): |
| if _role: |
| self.roles.add('\n' + _role) |
|
|
| def eos_in_role(self, history: str, t_bot): |
| if not (history.endswith('\n') or history.endswith('\r')): |
| return 0 |
| tmp = history.rstrip() |
| for _role in self.roles: |
| if tmp.endswith(_role): |
| n = len(t_bot) |
| for i in range(1, n): |
| tmp = self.model.str_detokenize(t_bot[n - i:]) |
| if tmp.rstrip().endswith(_role): |
| print('eos_in_role', t_bot[n - i:], repr(tmp)) |
| return i |
| print('eos_in_role missing') |
| break |
| return 0 |
|
|
| def eos_in_nlnl(self, history: str, t_bot): |
| if not (history.endswith('\n\n') or history.endswith('\n\r\n')): |
| return 0 |
| n = len(t_bot) |
| for i in range(1, n): |
| tmp = self.model.str_detokenize(t_bot[n - i:]) |
| if tmp.endswith('\n\n') or tmp.endswith('\n\r\n'): |
| if tmp.startswith(']'): |
| return 0 |
| print('eos_in_nlnl', t_bot[n - i:], repr(tmp)) |
| return i |
| print('eos_in_nlnl missing') |
| return 0 |
|
|
| def __call__(self, _role, prompt=None): |
| self._add_role(_role) |
| if prompt is None: |
| return self._get(_role) |
| |
| prompt = self.im_start + _role + self.nl + prompt |
| prompt = self.model.tokenize(prompt.encode('utf-8'), add_bos=False, special=True) + self.im_end_nl |
| |
| return prompt |
|
|