| | from turtle import st |
| | from zipfile import ZipFile, ZIP_DEFLATED |
| | from shutil import rmtree |
| | import json |
| | import os |
| | from tqdm import tqdm |
| | from collections import Counter |
| | from pprint import pprint |
| | import re |
| | import requests |
| | from dateutil import parser as date_parser |
| | from string import punctuation |
| | from copy import deepcopy |
| |
|
| |
|
| | def value_in_utt(value, utt): |
| | """return character level (start, end) if value in utt""" |
| | value = value.strip(punctuation).lower() |
| | utt = utt |
| | p = '(^|[\s,\.:\?!-])(?P<v>{})([\s,\.:\?!-\']|$)'.format(re.escape(value)) |
| | p = re.compile(p, re.I) |
| | m = re.search(p, utt) |
| | if m: |
| | |
| | return True, m.span('v') |
| | else: |
| | try: |
| | |
| | date_parser.parse(value) |
| | if (value.endswith('pm') or value.endswith('am')) and ''.join(value.split(' ')) in ''.join(utt.split(' ')): |
| | return True, None |
| | |
| | except: |
| | if value in utt: |
| | |
| | return True, None |
| |
|
| | return False, None |
| |
|
| |
|
| | def preprocess(): |
| | data_file = "kvret_dataset_public.zip" |
| | if not os.path.exists(data_file): |
| | response = requests.get("http://nlp.stanford.edu/projects/kvret/kvret_dataset_public.zip") |
| | open(data_file, "wb").write(response.content) |
| |
|
| | archive = ZipFile(data_file) |
| |
|
| | new_data_dir = 'data' |
| |
|
| | os.makedirs(new_data_dir, exist_ok=True) |
| |
|
| | dataset = 'kvret' |
| | splits = ['train', 'validation', 'test'] |
| | dialogues_by_split = {split:[] for split in splits} |
| |
|
| | ontology = {'domains': {}, |
| | 'intents': { |
| | 'inform': {'description': ''}, |
| | 'request': {'description': ''} |
| | }, |
| | 'state': {}, |
| | 'dialogue_acts': { |
| | "categorical": {}, |
| | "non-categorical": {}, |
| | "binary": {} |
| | }} |
| |
|
| | domain2slot = { |
| | 'schedule': ['event', 'time', 'date', 'party', 'room', 'agenda'], |
| | 'weather': ['location', 'weekly_time', 'temperature', 'weather_attribute'], |
| | 'navigate': ['poi', 'traffic_info', 'poi_type', 'address', 'distance'] |
| | } |
| | slot2domain = {slot: domain for domain in domain2slot for slot in domain2slot[domain]} |
| |
|
| | db = [] |
| | with archive.open(f'kvret_entities.json') as f: |
| | entities = json.load(f) |
| | for slot, values in entities.items(): |
| | domain = slot2domain[slot] |
| | ontology['domains'].setdefault(domain, {'description': '', 'slots': {}}) |
| | if slot == 'poi': |
| | for s in ['poi', 'address', 'poi_type']: |
| | ontology['domains'][domain]['slots'][s] = {'description': '', 'is_categorical': False, 'possible_values': []} |
| | for item in values: |
| | poi, address, poi_type = item['poi'], item['address'], item['type'] |
| | db.append({'poi': poi, 'address': address, 'poi_type': poi_type}) |
| | for s in ['poi', 'address', 'poi_type']: |
| | ontology['domains'][domain]['slots'][s]['possible_values'].append(db[-1][s]) |
| | continue |
| | elif slot == 'weekly_time': |
| | slot = 'date' |
| | elif slot == 'temperature': |
| | values = [f"{x}F" for x in values] |
| | elif slot == 'distance': |
| | values = [f"{x} miles" for x in values] |
| | |
| | ontology['domains'][domain]['slots'][slot] = {'description': '', 'is_categorical': False, 'possible_values': values} |
| |
|
| | for domain in ontology['domains']: |
| | for slot in ontology['domains'][domain]['slots']: |
| | ontology['domains'][domain]['slots'][slot]['possible_values'] = sorted(list(set(ontology['domains'][domain]['slots'][slot]['possible_values']))) |
| |
|
| | for data_split in splits: |
| | filename = data_split if data_split != 'validation' else 'dev' |
| | with archive.open(f'kvret_{filename}_public.json') as f: |
| | data = json.load(f) |
| | for item in tqdm(data): |
| | if len(item['dialogue']) == 0: |
| | continue |
| | scenario = item['scenario'] |
| | domain = scenario['task']['intent'] |
| |
|
| | slots = scenario['kb']['column_names'] |
| | db_results = {domain: []} |
| | if scenario['kb']['items']: |
| | for entry in scenario['kb']['items']: |
| | db_results[domain].append({s: entry[s] for s in slots}) |
| | |
| | dialogue_id = f'{dataset}-{data_split}-{len(dialogues_by_split[data_split])}' |
| | dialogue = { |
| | 'dataset': dataset, |
| | 'data_split': data_split, |
| | 'dialogue_id': dialogue_id, |
| | 'original_id': f'{data_split}-{len(dialogues_by_split[data_split])}', |
| | 'domains': [domain], |
| | 'turns': [] |
| | } |
| | init_state = {domain: {}} |
| |
|
| | for turn in item['dialogue']: |
| | speaker = 'user' if turn['turn'] == 'driver' else 'system' |
| | utt = turn['data']['utterance'].strip() |
| | if len(dialogue['turns']) > 0 and speaker == dialogue['turns'][-1]['speaker']: |
| | |
| | if utt == dialogue['turns'][-1]['utterance']: |
| | continue |
| | else: |
| | dialogue['turns'].pop(-1) |
| |
|
| | dialogue['turns'].append({ |
| | 'speaker': speaker, |
| | 'utterance': utt, |
| | 'utt_idx': len(dialogue['turns']), |
| | 'dialogue_acts': { |
| | 'binary': [], |
| | 'categorical': [], |
| | 'non-categorical': [], |
| | }, |
| | }) |
| | |
| | if speaker == 'user': |
| | dialogue['turns'][-1]['state'] = deepcopy(init_state) |
| | else: |
| | user_da = {'binary': [], 'categorical': [], 'non-categorical': []} |
| | user_utt = dialogue['turns'][-2]['utterance'] |
| |
|
| | for slot, value in turn['data']['slots'].items(): |
| | value = value.strip() |
| | is_appear, span = value_in_utt(value, user_utt) |
| | if is_appear: |
| | if span: |
| | start, end = span |
| | user_da['non-categorical'].append({ |
| | 'intent': 'inform', 'domain': domain, 'slot': slot, 'value': user_utt[start:end], |
| | 'start': start, 'end': end |
| | }) |
| | else: |
| | user_da['non-categorical'].append({ |
| | 'intent': 'inform', 'domain': domain, 'slot': slot, 'value': value, |
| | }) |
| | init_state[domain][slot] = value |
| | ontology['state'].setdefault(domain, {}) |
| | ontology['state'][domain].setdefault(slot, '') |
| | dialogue['turns'][-2]['state'] = deepcopy(init_state) |
| | |
| | for slot, present in turn['data']['requested'].items(): |
| | if slot not in turn['data']['slots'] and present: |
| | user_da['binary'].append({'intent': 'request', 'domain': domain, 'slot': slot}) |
| | |
| | dialogue['turns'][-2]['dialogue_acts'] = user_da |
| | dialogue['turns'][-1]['db_results'] = db_results |
| |
|
| | for da_type in user_da: |
| | das = user_da[da_type] |
| | for da in das: |
| | ontology["dialogue_acts"][da_type].setdefault((da['intent'], da['domain'], da['slot']), {}) |
| | ontology["dialogue_acts"][da_type][(da['intent'], da['domain'], da['slot'])]['user'] = True |
| |
|
| | assert all([s in ontology['domains'][domain]['slots'] for s in turn['data']['requested']]), print(turn['data']['requested'], ontology['domains'][domain]['slots'].keys()) |
| | assert all([s in ontology['domains'][domain]['slots'] for s in turn['data']['slots']]), print(turn['data']['slots'], ontology['domains'][domain]['slots'].keys()) |
| |
|
| | dialogues_by_split[data_split].append(dialogue) |
| |
|
| | for da_type in ontology['dialogue_acts']: |
| | ontology["dialogue_acts"][da_type] = sorted([str({'user': speakers.get('user', False), 'system': speakers.get('system', False), 'intent':da[0],'domain':da[1], 'slot':da[2]}) for da, speakers in ontology["dialogue_acts"][da_type].items()]) |
| | dialogues = dialogues_by_split['train']+dialogues_by_split['validation']+dialogues_by_split['test'] |
| | json.dump(dialogues[:10], open(f'dummy_data.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| | json.dump(ontology, open(f'{new_data_dir}/ontology.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| | json.dump(dialogues, open(f'{new_data_dir}/dialogues.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| | json.dump(db, open(f'{new_data_dir}/db.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| | with ZipFile('data.zip', 'w', ZIP_DEFLATED) as zf: |
| | for filename in os.listdir(new_data_dir): |
| | zf.write(f'{new_data_dir}/{filename}') |
| | rmtree(new_data_dir) |
| | return dialogues, ontology |
| |
|
| |
|
| | if __name__ == '__main__': |
| | preprocess() |
| |
|