DSTC9 Track 1 - Beyond Domain APIs: Task-oriented Conversational Modeling with Unstructured Knowledge Access
Most prior work on task-oriented dialogue systems are restricted to a limited coverage of domain APIs, while users oftentimes have domain related requests that are not covered by the APIs. In this paper, we propose to expand coverage of task-oriented dialogue systems by incorporating external unstructured knowledge sources. We define three sub-tasks: knowledge-seeking turn detection, knowledge selection, and knowledge-grounded response generation, which can be modeled individually or jointly. We introduce an augmented version of MultiWOZ 2.1, which includes new out-of-API-coverage turns and responses grounded on external knowledge sources. We present baselines for each sub-task using both conventional and neural approaches. Our experimental results demonstrate the need for further research in this direction to enable more informative conversational systems.READ FULL TEXT VIEW PDF
DSTC9 Track 1 - Beyond Domain APIs: Task-oriented Conversational Modeling with Unstructured Knowledge Access
Traditionally, task-oriented dialogue systems have focused on providing information and performing actions that can be handled only by given databases or APIs. However, in addition to task-focused requests, users also have needs that go beyond what is provided by the backend resources. For example, while most virtual assistants can help users book a hotel, a restaurant or movie tickets, they fall short of answering potential follow-up questions users may have, such as: where to park vehicles; whether they are allowed to bring pets or children to the reserved place; or what the cancellation policy is. No API/DB entry is usually available to handle such requests. On the other hand, relevant domain knowledge is already available on web pages in the form of descriptions, FAQs and customer reviews for many of these out-of-coverage scenarios. Since current dialogue systems don’t incorporate these external knowledge sources into task-oriented conversational modeling, users need to visit the websites by themselves to find out any additional information beyond API/DB coverage, making conversational interactions inefficient.
In this work, we propose a new conversational modeling task towards frictionless task-oriented scenarios, where the flow of the conversation does not break when users have requests that are out of the coverage of APIs/DB but potentially are already available in external knowledge sources. Inspired by recent studies on knowledge-grounded conversational modeling Zhou et al. (2018); Dinan et al. (2018); Galley et al. (2019); Gopalakrishnan et al. (2019), our proposed task aims to develop end-to-end dialogue systems to understand relevant domain knowledge, and generate system responses with the selected knowledge. Different from previous work on social conversations Ritter et al. (2011); Vinyals and Le (2015); Serban et al. (2017), this task addresses task-oriented conversations grounded on fine-grained domain-level or entity-level knowledge sources related to given dialogue contexts.
Figure 1 shows an example conversation with unstructured knowledge access. The user utterances at turns and request the policy details about bringing pets and making payments, respectively, which are out of the coverage of the structured domain APIs. On the other hand, the relevant knowledge contents can be found from the external sources as in the rightmost column which includes the QA snippets from the FAQ lists for each corresponding entity within domains such as train, hotel, or restaurant. With access to these unstructured external knowledge sources, the agent managed to continue the conversation with no friction by responding adequately at the turns .
|Speaker||Utterance ()||Relevant Knowledge Snippets from FAQs|
|1||User||I need a train from King’s Lynn to Cambridge arriving by 17:45 on Sunday.|
|2||Agent||I have train TR6003 leaving cambridge at 5:11 and arriving at 5:58 on sunday|
|[.4pt/1pt] 3||User||I also need to bring my dog. Do they allow pets?||Q: Can I bring my pets to trains?
A: We happily welcome dogs and cats up to 20 pounds.
|4||Agent||Yes, you can travel with your dog up to 20 pounds.|
|[.4pt/1pt] 5||User||That sounds great. I also need help finding a place to stay in the South.|
|6||Agent||I have 2 options available. The Lensfield Hotel and Gonville Hotel.|
|[.4pt/1pt] 7||User||Do either of them allow to stay with my dog?||Q1: Can I bring my dog to Gonville Hotel?
A1: Pets are permitted. But charges may be applicable.
|8||Agent||You could stay with your dog at Gonville with a fee.|
|4-4[.4pt/1pt]||Would you like me to book it?||Q2: Can I bring my dog?
A2: Pets are not allowed at the Lensfield Hotel.
|[.4pt/1pt] 9||User||Maybe later. How about a place to eat nearby? Chinese food would be great.|
|10||Agent||The peking restaurant is a nice place. Do you need reservations?|
|[.4pt/1pt] 11||User||Before that, could you confirm that this restaurant accepts AMEX?||Q: What type of payments are accepted?
A: Peking Restaurant accepts cash only.
|12||Agent||Unfortunately, the Peking restaurant accepts cash only. Would it work?|
|[.4pt/1pt] 13||User||Okay. Can you book a table for 4 at 18:30 on Monday, please?|
|14||Agent||Booking was successful. Do you have any other question?|
|[.4pt/1pt] 15||User||What about the hotel? Can I use my credit card there?||Q: What credit cards are accepted by Gonville Hotel?
A: You can use AMEX, Visa and Mastercard.
|16||Agent||Yes, Gonville Hotel accepts all major credit cards including AMEX.|
Task-oriented dialogue systems aim to enable users to complete tasks by interacting with an automated agent in natural language Young et al. (2013). These systems typically convert user utterances to a semantic representation (such as domain, intent, and slots Tur and De Mori (2011)
) based on what is used by the backend resources (such as APIs) that accomplish the tasks. At each turn, the dialogue system decides the next action to take based on the estimated dialogue state as well as any results or responses from the backend resourcesLevin et al. (2000); Singh et al. (2002); Williams and Young (2007)
. The next action, which is typically in the form of a semantic frame formed of dialogue acts, arguments and values, is converted to a natural language response to the user by natural language generationPerera and Nand (2017).
On the other hand, social conversational systems typically follow an end-to-end approach, and aim to generate target responses based on the previous conversation context Ritter et al. (2011); Vinyals and Le (2015); Serban et al. (2017). Ghazvininejad et al. Ghazvininejad et al. (2018) proposed an extension to these models that grounds the responses on unstructured, textual knowledge, by using end-to-end memory networks where an attention over the knowledge relevant to the conversation context is estimated. Along similar lines, Liu et al. Liu et al. (2018)et al. (2018)
proposed both static and dynamic graph attention mechanisms for knowledge selection and response generation, respectively, using knowledge graphs. More recently, Dinan et al.Dinan et al. (2018) and Gopalakrishnan et al. Gopalakrishnan et al. (2019) both have publicly released large conversational data sets, where knowledge sentences related to each conversation turn are annotated. Our proposed task, data, and baseline models in this work differ from these studies in the following aspects: we target task-oriented conversations with more clear goals and explicit dialogue states than social conversations; and we aim to incorporate task-specific domain knowledge instead of commonsense knowledge.
The other line of related work is machine reading comprehension which aims to answer questions given unstructured text Richardson et al. (2013); Hermann et al. (2015); Rajpurkar et al. (2016) and has later been extended to conversational question answering Choi et al. (2018); Reddy et al. (2019). In our work, the document required to generate a response needs to be identified according to the conversation context. The responses are also different in that, rather than plain answers to factual questions, we aim to form factually accurate responses that seamlessly blend into the conversation.
We define an unstructured knowledge-grounded task-oriented conversational modeling task based on a simple baseline architecture (Figure 2) which decouples turns that could be handled by existing task-oriented conversational models with no extra knowledge and turns that require external knowledge resources. In this work, we assume that a conventional API-based system already exists and focus on the new knowledge access branch which takes a dialogue context and knowledge snippets , where is the -th utterance in a given dialogue, is the time-step of the current user utterance to be processed, is the dialogue context window size.
Our proposed task aims to generate a context-appropriate system response grounded on a set of relevant knowledge snippets . The remainder of this section presents the detailed formulations of the following three sub-tasks: ‘Knowledge-seeking Turn Detection’, ‘Knowledge Selection’, and ‘Knowledge-grounded Response Generation’.
For each given turn at , a system first needs to decide whether to continue an existing API-based scenario or trigger the knowledge access branch. We call this task Knowledge-seeking Turn Detection. This problem is defined as a binary classification task formulated as follows:
which we assume that every turn can be handled by either branch in this work. For the examples in Figure 1, for the knowledge-seeking turns at , while for the other user turns at .
Once a given user turn at is determined as a knowledge-seeking turn by , it moves forward with Knowledge Selection to sort out the relevant knowledge snippets. This task takes each pair of and and predicts whether they are relevant or not as follows:
Different from other information retrieval problems taking only a short single query, this knowledge selection task must be highly aware of the dialogue context. For example, and themselves in Figure 1 share the same question type with similar surface form, but the relevant knowledge snippets would vary depending on their dialogue states across different domains. Even within a single domain, fine-grained dialogue context needs to be taken into account to select proper knowledge snippets corresponding to a specific entity, for example, ‘Peking Restaurant’ and ‘Gonville Hotel’ for and against any other restaurants and hotels, respectively.
Since more than one knowledge snippet can be relevant to a single turn, as for in Figure 1, we form a task output including all the positive knowledge snippets from , as follows:
Finally, a system response is generated based on both dialogue context and the selected knowledge snippets , as follows:
Each generated response is supposed to provide the user with the requested information grounded on the properly selected knowledge sources. In addition, the response should be naturally connected to the previous turns. The knowledge-grounded responses in Figure 1 focus not only on delivery of the information by knowledge access, but also maintain natural conversation. For example, the responses at paraphrase written sentences into a colloquial style, the responses at acknowledge before giving a statements, the responses at ask a follow-up question to the user.
To address the proposed research problems, we collected an augmented version of MultiWOZ 2.1 Budzianowski et al. (2018); Eric et al. (2019) with out-of-API-coverage turns grounded on external knowledge sources beyond the original database entries. This was incrementally done by the following three crowdsourcing tasks.
First, crowd workers were given a dialogue sampled from the original MultiWOZ 2.1 conversations and asked to indicate an appropriate position to insert a new turn about a selected subject from external knowledge categories (Figure 2(a)). This task aims to collect user behaviors about when to ask a knowledge-seeking question for a given subject. It corresponds to the knowledge-seeking turn detection sub-task in Section 3.1.
Then, they were asked to write down a new user utterance at each selected position in the first task to discuss about a given corresponding subject (Figure 2(b)), which is for both knowledge-seeking turn detection (Section 3.1) and knowledge selection (Section 3.2) sub-tasks. In order to collect various expressions, a single task with the same dialogue context and knowledge category was assigned to multiple crowd workers in parallel.
Finally, we collected the agent’s response to each question collected in the previous step. In this task (Figure 2(c)), crowd workers were given external knowledge sources for each category and asked to convert them into a system response which is more colloquial and coherent to both the question and dialogue context. This task aims at knowledge-grounded response generation (Section 3.3).
Our proposed pipeline for data collection has the following advantages over Wizard-of-Oz (WoZ) approaches. First, it is more efficient and scalable, since every task can be done by a single crowd worker independently from others, while WoZ requires to pair up two crowd workers in real time. This aspect enables us to have more control in the whole process compared to the end-to-end data collection entirely by crowd workers from scratch. Furthermore, the intermediate outcomes from each phase can be utilized to build conversational models with no additional annotation.
|Split||# dialogues||# augmented turns||# utterances|
|Domain||# snippets||# entities||# snippets|
Table 1 shows the statistics of the collected data sets. A total of 9,072 utterance pairs are newly collected in addition to the original MultiWOZ dialogues, each of which is linked to corresponding knowledge snippets among 938 question-answer pairs (Table 2) collected from the FAQ webpages about the domains and the entities in MultiWOZ databases. Figure 4 shows the length distribution of the augmented utterances. Similar to the original MultiWOZ Budzianowski et al. (2018), the agent responses are longer than the user utterances, which have 12.45 and 9.85 tokens on average spoken by agents and users, respectively. Figure 5 presents the distribution of trigram prefixes of the augmented user utterances with various types of follow-up questions that go beyond the coverage of domain APIs.
In this section, we present baseline methods for the problems defined in Section 3
. Specifically, we introduce both a non-machine learning approach and a neural baseline model for each sub-task.
For the knowledge-seeking turn detection, we compare two baselines with unsupervised anomaly detection and supervised classification methods.
In the first baseline, we consider the task as an anomaly detection problem that aims to identify the turns that are out of the coverage of conventional API-based requests. Given the assumption that there is no knowledge-seeking turn available in most task-oriented dialogue data, we applied an unsupervised anomaly detection algorithm, Local Outlier Factor (LOF)Breunig et al. (2000). The algorithm compares the local densities between a given input instance and its nearest neighbors. If the input has a significantly lower density than the neighbors, it is considered an anomaly.
We built a knowledge-seeking turn detector with the LOF implementation in PyOD Zhao et al. (2019) with its default configurations. The system includes all the user utterances in the original MultiWOZ 2.1 training set. Every utterance in both training and test sets was encoded by the uncased pre-trained BERT Devlin et al. (2019) model.
If training data is available for the knowledge-seeking turn detection, the most straightforward solution will be training a binary classifier in a supervised manner. In this experiment, we fine-tuned the uncased pre-trained BERTDevlin et al. (2019) model on the training data in Section 4. The model takes each single user utterance as an input and generates the utterance representation as the final layer output for
which is a special token in the beginning of the input sequence. We added a single layer feedforward network on top of the utterance embeddings, which was trained with binary cross-entropy loss for three epochs. We used a mini-batch size of 128 with truncated utterances up to 256 tokens.
In our experiments, we consider two variants of the knowledge selector: unsupervised knowledge-retrieval baselines and supervised neural Transformer architectures.
First, we propose the unsupervised knowledge selection baselines using information retrieval (IR) algorithms (Figure 6). Let us denote an encoder function
mapping the concatenation of all the sentences in a query or a document to a fixed-dimensional weight vector. In this work, we take the dialogue contextas a query and each knowledge snippet as a candidate document. When scoring entity-level knowledge, we also add the name of the entity to each document being scored as this helps differentiate among potentially ambiguous knowledge contents that may be applicable to multiple entities.
Our IR model then computes the following cosine similarity score per knowledge snippet:
where we finally take the most relevant document as a selected knowledge in the following fashion:
We use two types of standard IR baselines: a TF-IDF Manning et al. (2008) and a BM25 Robertson and Zaragoza (2009) system. We also consider another IR baseline that employs an uncased pretrained BERT model as a static utterance encoder. In this baseline, we encode and each separately and then compute the cosine similarity between the pooled utterance outputs.
We also employ a BERT-based Devlin et al. (2019) neural model as a baseline knowledge selection system. In particular, we train a binary classification model (Figure 7) over a pair of encoded texts as is done in prior Transformer sentence relationship models Nogueira and Cho (2019). The model takes the concatenation of the utterances in and the sentences in as an input instance. We use the final layer output at the same position to the
token as input to a single layer feedforward network to obtain a probabilitythat the is relevant to the given dialogue context .
We finetune a pretrained BERT model using a binary cross-entropy loss as follows:
where refers to the set of knowledges that are relevant for the given dialogue context and refers to those that are not.
During training of the knowledge classifier, we experimented with sampling methods of negative knowledge candidates to be paired with a given dialogue context. For dialogues annotated with domain-level knowledge, we chose negative candidates by sampling other documents in the same domain as the annotation. For entity-level knowledge dialogues, we chose negative candidates by sampling other documents from the same entity as the provided annotation. We built models in which the number of negative candidates for each positive example was varied from 1 to 13 in increments of 4 and found the best-performing model used negative candidates for each positive candidate.
In this section, we propose both extractive and generative approaches for the knowledge-grounded response generation task.
The simplest method for knowledge-grounded response generation is to output a part of the selected knowledge snippets. In this experiment, we developed an answer extraction baseline with the following heuristics:
If multiple knowledge snippets are related to a given turn, randomly pick one of them. Otherwise, a sole snippet is taken as the source for answer extraction.
If the target snippet includes multiple paragraphs, extract only the first paragraph as a system response. Otherwise, the whole paragraph is considered as the output.
Given the tremendous interest and success in leveraging large pre-trained language models for downstream NLP tasks in the community, our neural baseline leverages the Generative Pre-trained Transformer (GPT-2) model Radford et al. (2019). We fine-tuned the GPT-2 small model with a standard language modeling objective on our dataset, using both the knowledge-augmented and regular system turns as target sequences. To show the influence of knowledge, we compared two variants of models with different inputs, as follows:
GPT-2 w/o knowledge: no knowledge was used during fine-tuning.
GPT-2 w/ knowledge: the ground-truth knowledge snippets were concatenated to each input dialog context (Figure 8) for fine-tuning.
We used the transformers library Wolf et al. (2019a) 111https://huggingface.co/transformers/ to fine-tune the models for a fixed number of 3 epochs with a truncation window of 256 tokens for both dialog context and knowledge snippet
. We used a train batch size of 2, performed gradient accumulation every 8 steps and gradient clipping with a max norm of, used the Adam optimizer and linearly decayed the learning rate from 6.25e-5 to 0 during fine-tuning.
We added special tokens for both speakers user and agent to our vocabulary, initialized their parameters randomly and learned them during fine-tuning. We enriched the corresponding turns in the input with speaker embeddings at a token-level by identifying their token types, exactly as described in Wolf et al. (2019b). We used top-, top- nucleus sampling with temperature Holtzman et al. (2019) for decoding, where , and . We also set a maximum decode length of 40 tokens.
First, we evaluated the knowledge-seeking turn detection performances of unsupervised anomaly detection (Section 5.1.1) and supervised neural classification (Section 5.2.2) methods. Both models were built on all the user utterances in the training set and evaluated on the test set user turns in accuracy, precision, recall, and F-measure.
Table 3 shows that the unsupervised baseline has a limitation in distinguishing between API-based and knowledge-seeking turns, especially with many false positives. On the other hand, the neural classifier achieved almost perfect performance in all the metrics. Nevertheless, this utterance classifier may work well when restricted only to this data set or similar, due to lack of knowledge or API details incorporated into the model. There is much room for improvement in making the model more generalizable to unseen domains or knowledge sources.
|[.4pt/1pt] Classification (BERT)||0.891||0.834||0.976|
Knowledge selection was evaluated using a number of standard IR metrics including recall (R@1 and R@5), and mean reciprocal rank (MRR@5). For domain-knowledge dialogues, our total candidate set included all domain knowledges for the annotated domain, and for entity-knowledge dialogues our total candidate set included all entity knowledges for the annotated entity.
Table 4 shows that our bag-of-words IR baselines (Section 5.2.1) outperformed the static BERT encoder across all three metrics. However, the neural classifier model (Section 5.2.2) significantly outperformed the IR baselines, demonstrating the efficacy of downstream fine-tuning of large pre-trained neural representations. That being said, there is still a substantial performance gap in the R@1 and MRR@5 metrics, leaving room for further research into knowledge selection on this data.
|Method||PPL||Unigram F1||Div. ( = 1)||Div. ( = 2)||BLEU-4||METEOR||ROUGE-L|
|[.4pt/1pt] GPT-2 w/o knowledge||5.0906||0.2620||0.0509||0.1589||0.0559||0.2202||0.1979|
|GPT-2 with knowledge||4.1723||0.3175||0.0509||0.1559||0.0840||0.2796||0.2403|
|GPT-2 w/o knowledge||Human||-||-||-||4.59||27.76||67.65|
|GPT-2 with knowledge||Human||36.02||59.49||4.49||5.31||22.96||71.74|
|[.4pt/1pt] GPT-2 with knowledge||Answer Extraction||56.33||31.02||12.65||-||-||-|
|GPT-2 with knowledge||GPT-2 w/o knowledge||-||-||-||22.55||17.04||60.41|
Responses by answer extraction (Section 5.3.1) and neural generation models (Section 5.3.2) were first evaluated using the following automated metrics: perplexity, unigram F1, n-gram diversity, BLEU-4, METEOR, and ROUGE-L. The evaluation was done only on the augmented turns with the ground-truth knowledge, in order to characterize the models’ ability to handle the external knowledge scenario. Table 5 shows that our generation models achieved better scores than the extractive baseline on most metrics. Especially, the GPT-2 model with knowledge outperformed both the answer extraction baseline and the other GPT-2 variant with no knowledge in BLEU-4, METEOR, and ROUGE-L, which indicates that our proposed neural model generates more human-like responses than the extractive baseline.
In addition, we also performed human evaluations of the generated responses with the following two crowdsourcing tasks:
Appropriateness: given a dialogue context and a pair of responses generated by two methods, crowdworkers were asked to select a more appropriate response to the context.
Accuracy: given a knowledge snippet and a pair of responses generated by two methods, crowdworkers were asked to select a more accurate response to the knowledge.
In both tasks, we presented each instance to three crowdworkers; asked them to choose either response or ‘not sure’ for the cases that are equally good or bad; and took the majority as the final label for the instance. Table 6 shows that our GPT-2 models generated more appropriate responses than the answer extraction baseline. Comparing between two GPT-2 variants, the model with knowledge provided more accurate information based on explicitly given knowledge than the one without knowledge. However, this accuracy gap between two models is not very big, which depicts the need to add more diversity in knowledge content which cannot be handled just by memorizing facts from the training data.
This paper proposed a new task-oriented conversational modeling problem grounded on unstructured domain knowledge, which aims to handle out-of-API coverage user requests. To support research on our proposed tasks, we introduced an augmented version of MultiWOZ 2.1 dialogues with additional knowledge-seeking turns collected given external knowledge sources. We presented baseline methods based both on non-machine learning approaches and neural model architectures.
Furthering this work, we plan to collect more dialogues including different domains, entities, and locales from the original ones for MultiWOZ 2.1. Moreover, this new data set will include not only written conversations, but also spoken dialogues to evaluate the system performances for more realistic scenarios. Then, all the data sets and the baselines will be released for establishing a new public benchmark in dialogue research.
In addition, we will continue to iterate on the models with the following potential enhancements: end-to-end learning instead of the pipelined processing, joint modeling of both knowledge-seeking and API-driven branches, and few shot transfer learning for unseen domains or knowledge sources.
Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 5016–5026. Cited by: §4, §4.
Wizard of wikipedia: knowledge-powered conversational agents. arXiv preprint arXiv:1811.01241. Cited by: §1, §2.
Thirty-Second AAAI Conference on Artificial Intelligence, Cited by: §2.
Optimizing dialogue management with reinforcement learning: experiments with the njfun system. Journal of Artificial Intelligence Research 16, pp. 105–133. Cited by: §2.
Partially observable markov decision processes for spoken dialog systems. Computer Speech & Language 21 (2), pp. 393–422. Cited by: §2.
Transfertransfo: a transfer learning approach for neural network based conversational agents. arXiv preprint arXiv:1901.08149. Cited by: §5.3.2.
PyOD: a python toolbox for scalable outlier detection. Journal of Machine Learning Research 20 (96), pp. 1–7. External Links: Cited by: §5.1.1.
Figure 9 and Figure 10 show examples of knowledge snippets used in our data collection for domain- and entity-specific augmented turns, respectively. While domain-level snippets include generic information that could be applicable over all the domain entities, entity-level knowledge varies depending on a given entity even for the same question.
|[.4pt/1pt] Title||How can I get an invoice?|
|Body||The property can provide you with an invoice for your stay, so please contact them directly.|
|Body||You can cancel a reservation online or call the restaurant directly. Please note that some restaurants have implemented a 24-48 hour cancellation policy.|
|[.4pt/1pt] Title||Discount Information for Children|
|Body||One child ages 2-12 is eligible to receive a 50% discount on the lowest available adult rail fare on most trains with each fare-paying adult (age 18+).|
|[.4pt/1pt] Title||What is the parking charge?|
|Body||Parking costs GBP 14 per day.|
|[.4pt/1pt] Title||Is there free parking at your lodge?|
|Body||Hamilton Lodge offers free parking.|
|[.4pt/1pt] Title||What credit cards are accepted?|
|Body||La Margherita takes Visa, Mastercard and AMEX.|
|[.4pt/1pt] Title||What type of payments are accepted at this location?|
|Body||Peking Restaurant accepts cash only.|
Figure 13 compares the example responses generated by the proposed methods. GPT-2 models generated more natural and appropriate responses to be spoken by the agent than the answer extraction baseline. Comparing between two GPT-2 variants, the model with knowledge provided more accurate information than the one without knowledge. However, there is still room for improvement in both appropriateness and accuracy compared to the human responses.
|User||I’m also looking for info on a hotel called allenbell.|
|Agent||Would you like me to book it for you?|
|User||Am I able to smoke while staying at this hotel?|
|Q: Can I smoke there?|
|A: Allenbell is a non-smoking establishment.|
|Human||The Allenbell does not allow smoking in their hotel. Can I help you answer any other questions?|
|[.4pt/1pt] Extracted||Allenbell is a non-smoking establishment.|
|[.4pt/1pt] GPT-2 w/o k.||Please contact us for more information|
|[.4pt/1pt] GPT-2 with k.||No, they don’t allow smoking in the hotel. Would you like to try a different hotel?|
|User||Great! I’m also looking for a place to eat in the same area. Something not too expensive, but not cheap.|
|Agent||There are 21 places in the centre of town. What type of cuisine would you like?|
|User||I am interested in moderately priced international food.|
|Agent||I’d recommend the Varsity Restaurant. Would you like to book a table?|
|Q: Payment options|
|A: We prefer cash payment. We accept MasterCard and Visa with a $20 minimum purchase.|
|Human||Yes. Cash is actually the preferred payment method.|
|[.4pt/1pt] Extracted||We prefer cash payment. We accept MasterCard and Visa with a $20 minimum purchase.|
|[.4pt/1pt] GPT-2 w/o k.||Yes, all major credit cards are accepted. Would you like me to make a reservation for you?|
|[.4pt/1pt] GPT-2 with K.||Yes, the restaurant accepts cash. Do you need anything else?|
|User||Can you help me find a restaurant that serves African food with a moderate price range please|
|Agent||I am sorry, there aren’t any options available. May I ask if there is another type of restaurant you would be interested in?|
|User||Yes how about Asian food in the same price range?|
|Agent||Yes I have the Yippee Noodle Bar in the center of town on King street in the moderate price range. They serve Asian cuisine. Is there anything else I can do for you?|
|User||Will you please book a table for 6 at 13:45 on Friday?|
|Agent||Booking was successful. The table will be reserved for 15 minutes. Reference number is : 3GLLHSNN.|
|Q: What parking options are available at your location?|
|A: Only street parking is available.|
|Human||There is only parking on the street. Anything else I can do for you?|
|[.4pt/1pt] Extracted||Only street parking is available.|
|[.4pt/1pt] GPT-2 w/o k.||No, the restaurant does not offer parking for parking. Would you like me to book a reservation for you?|
|[.4pt/1pt] GPT-2 with k.||Yes it does. Would you like me to make reservations for you?|