0:00:16okay thanks and but i'll and the buttons and then this a joint work with
0:00:20the next five rate and small integration
0:00:24so
0:00:26some present detection on viral are any detection has recently become a popular and of
0:00:31the problem
0:00:32and the method of the these detection research is looking into the utterances of what
0:00:37the point and of utterance and in isolation
0:00:40however
0:00:41as we know that sarcasm is a complex phenomena in natural language and speaker intent
0:00:46is often unclear a list to provide additional context so this conversation context background knowledge
0:00:53from of doctors
0:00:54the topic of discussion in section
0:00:56and that's why some researchers have argued that these are the sort of additional context
0:01:02has to be provided to understand sarcasm bitter
0:01:05and that actually triggers to our research here we look into basically to research questions
0:01:11that far does conversation context held in predicting some present but conference in context is
0:01:17that one type of context
0:01:19and can identify what part of the conversation context actually triggering a sarcastic reply
0:01:26so before going to the modeling let's look into some of the examples that we're
0:01:31looking you know researcher
0:01:33so this that we'd
0:01:34and whether a user is saying that one more reason to feel really great of
0:01:39what's line and then to put the has text are present which would be used
0:01:42as a label
0:01:43now is looking to that with this is
0:01:46a hard to understand that why this to it is a sarcastic to eat right
0:01:50but of the same time we see that the street was a reply to one
0:01:54other tweet with the symbolic can see that at the rate
0:01:57and
0:01:58that was the reply to another to it from another user
0:02:01where it says that plane window shades or open so that people can see
0:02:05so we can understand that the user is sarcastic about the experience of line here
0:02:10and that's why to put i would that has text are present
0:02:13no let's look at a another domain
0:02:16and this
0:02:16these the discussion thread or stun it's taken from the internet argument corpus
0:02:21but this user again
0:02:24replying to some other user and reading about sarcastic about their reading have it
0:02:30and think that all your reading too much into by
0:02:33is to look into the context forcing we see that the point x this way
0:02:36a lot longer than we because it's taken from a discussion forum posts and i'm
0:02:41going to the detail of the post but what's the person is saying that you
0:02:44think that well the what is not sixty million years old it's only two thousand
0:02:48years old and they bring
0:02:50definitely different arguments
0:02:52and the user to is just saying that go you're eating too much into by
0:02:55the and that's what they're
0:02:57sarcastic about
0:02:59so the outline of my job here two days like with fast discuss about though
0:03:04i don't results are some of the state-of-the-art in sarcasm detection i'll also talk about
0:03:08but data
0:03:09and then i'll go over the first research question that can conversation context l in
0:03:14detection some thousand and then they're seconds question that what the context actually triggering the
0:03:19sarcastic reply and can identify that
0:03:22and finally alicante would and a point to some future work
0:03:28so as explained before there's a sergeant sarcasm environment detection in the recent years
0:03:33and that of the research looking to see this as a typical binary classification problem
0:03:38of not be which we we're doing like for an entity tagging or
0:03:42relation extraction that sector
0:03:44but there are some of the directions of the research also has come up in
0:03:47started and the second recently and out just if you talk about that are for
0:03:50instance right of it all the look into the context in can really characteristic of
0:03:54sarcasm a then in our previous work in yemen we looked sarcasm detection as a
0:04:00what sense disambiguation problem
0:04:02i miss rate all they have looked into cognitive feature set as i tracking feature
0:04:06for sarcasm data sent to find a lot they look into timbre posts where images
0:04:11and the text can be combined for a second predictions and yesterday we saw one
0:04:16poster from a horribly at all that the looked into rhetorical questions where
0:04:20a lot of sarcastic utterances are actually rhetorical questions
0:04:24from the respected of role of context
0:04:27that's all the users have looked into orders context so that they actually model orders
0:04:31previous x in a previous
0:04:34post like into it
0:04:35to understand that if the authors of sarcastic on that
0:04:38are there is some work in the conversation context also especially balance but they look
0:04:42into like n-gram characteristics and their performance we don't the context was not much different
0:04:48and finally we think that combining wall knowledge how would be like really crucial to
0:04:54understand sarcasm but
0:04:55and the same time we think it's
0:04:56much harder problem so for instance if we see this to it
0:05:01that users thing that driveway that we are having a long time so that means
0:05:04that
0:05:05alignment has something to do with green and this sort of all knowledge is very
0:05:09difficult
0:05:10to include
0:05:11in a system
0:05:13your research we
0:05:14particularly look into the conversational context here
0:05:18so we use two sets of data the four cities like discussion four
0:05:22and that was that it partially released in the lasagna conference eigenvalue rugby at all
0:05:28and the structure is that you have a sarcastic response and which replies to a
0:05:33context and vocalic discussion forum post
0:05:37some of the characteristic of the data that the data was like allocated that comment
0:05:40level using crowdsourcing and it's a balance set of training data close to five thousand
0:05:46a post
0:05:47from the context and response
0:05:50we also collected data from twitter
0:05:52and we look specifically into sarcastic with which are actually like previous tweet and when
0:05:58we
0:05:59we collected the previous fight using this parameter like at user
0:06:03and then whenever possible we just collect the full set of the dialogue so sometimes
0:06:08like one user is sarcastic about out of their tweet and that we use they
0:06:12can reply to another with so we have collected the full trade
0:06:15and labels are provided by the authors so that's one different from the discussion forum
0:06:20where it's not just one delegating here the labels are provided and we use that
0:06:25have sex sarcasm and has tech sarcastic to understand what to do sarcastic to each
0:06:30other
0:06:31again
0:06:32it sort of on about in balance that but it's still close like two thousand
0:06:36and thirteen thousand it for sarcastic and on circuits degrees respectively
0:06:41and more than thirty bucks of the data we had a couple of sentences as
0:06:45a context
0:06:45also it should be given as like into it as max for the contextual characteristics
0:06:50here
0:06:52looking to the model can conversation context help in sarcasm detection
0:06:57our baseline
0:06:59features for svm linear kernel with discrete features
0:07:02and do that the features with actually a part one very well in previous the
0:07:06sarcasm detection research and that's where you're using
0:07:09n-gram features we use like bigram and trigram and unigram we also used two sets
0:07:15of lexicons like they are lexical for sentiment and
0:07:18liwc louis for pragmatic features
0:07:21absolutely like alteration of the sentiments between the point text and sorry a sarcastic post
0:07:27and then we used a set of sarcasm markers which are basically indicators of sarcasm
0:07:32a and that's been used in many research and linguistic and communication signs are for
0:07:37instance like morpho syntactic features so use of various interjections questions examine some signs
0:07:44and then when you're doing like speech recognition i'm sarcasm the voice into an snr
0:07:49modulations is like they have like strong features
0:07:52but one opinion to natural language and people like typing the put like this kind
0:07:55of typographic feature set this quotation mark a different type of you multi phones
0:07:59you move these are used also capitalisation such as this what the time showing like
0:08:04never
0:08:05and finally a list of intensify record is also shown to be a strong features
0:08:10in sarcasm detection like the worst degrade based and beta next section
0:08:14we also experience with the lstm network of which are able to a lower long
0:08:19distance dependencies
0:08:20and in our architecture we are using two lstm one read stuff context and one
0:08:25reads the response
0:08:27your six print with that consent base variations where first we used a lot and
0:08:31sentence level attention and that's like a hierarchical model
0:08:35and in the second we give the word embeddings static and we didn't do any
0:08:39a bit attention on the words but we only put that consent on the sentence
0:08:43of the context because we also one two
0:08:46answer our second visits question that
0:08:47what part of the context also help in identifying the sarcasm so that's why you're
0:08:52doing the sentence level only
0:08:55this is a schematic diagram of our utterancelevel for the sentence a sentence but seconds
0:09:01and so you can see that the left hand side we have one lstm to
0:09:04learn the content
0:09:05and in right hand side we have one lstm to learn the response the real
0:09:09there is a sentence embedding then you have the hidden layer then we have attention
0:09:13and then you have the context vector
0:09:15and what we mean here that we are concatenating the final vector representation from the
0:09:19context and the response concatenating and then we're passing it to the softmax
0:09:25a finally we also experimented with other variation the fella stimulus the conditional density and
0:09:30that was introduced from the deep mine group for the a natural language inferencing tasks
0:09:34like textual entailment
0:09:36what we're doing here is that again we are using two lstm
0:09:39but the response lstm is conditioned on they represent on the context of spam
0:09:44so which means that
0:09:45the sale state is for the response is initialized by the final state of the
0:09:50context and
0:09:51that's
0:09:52been shown that it's like really a strong character structure architecture for
0:09:57nestling in inferencing task
0:10:01more some more details that we split the data into a dt in general like
0:10:04training is like eighty percent data ten percent data we used for the parameter tuning
0:10:09and we use two sets of word embeddings for discussion forums so we use that
0:10:13standard go would be ten few hundred dimensional what effect model that for twitter we
0:10:18used as basic model that was been trained on tweets and we have used you
0:10:21know this is for
0:10:24no so the results are this talk about some of the macro-average the that allowed
0:10:28us point out something important how values and
0:10:31the more i thought the number seven in their in more details in the paper
0:10:36here first we look into the comparison of the is em results with the response
0:10:41and context bus response we said that into a terry there is a some of
0:10:46that some part of the improvement is they're using the context
0:10:49but not in the discussion forum data and we suspect that
0:10:52it's might be possible because a lot of and the context of soul all in
0:10:56the discussion forums and
0:10:57is v m which are mostly based on n-gram features for not every to learn
0:11:01that
0:11:03if you look into the lstm variations we only the we don't that instant model
0:11:08we see that
0:11:09the first observation is using context help
0:11:12for both lstm models here
0:11:15but at the same time they list in conditional model which was kind is the
0:11:19weather response was conditioned on the context
0:11:21it performs reasonably well so we see there is like
0:11:25five to six percent improvement of if one for the discussion forum and for the
0:11:29preacher results more
0:11:31and finally that the results from using attention over the sentences with the response and
0:11:37with the context and we see that sort the treated it to the context and
0:11:41response performs past and so the discussion forum it just like comparable to the conditional
0:11:47now we i'm not showing the results we when we do the attentional what in
0:11:52sentence because that are phones a little more than that
0:11:55and what we think there isn't could be that
0:11:57are we don't have enough training data and when you're putting attention what on the
0:12:00warden sentence level that are too many parameters to model so in the future we
0:12:04hope that
0:12:05well when you're we can put more data under discussion forum at which are especially
0:12:09to
0:12:11experiment in that direction and that might prove better
0:12:15no
0:12:16we go to the second research question that and the identify
0:12:20what of the context figure out the sarcastic reply
0:12:23on what we're doing here that's basically we're looking to that happens and wait and
0:12:27if the indicate what part of the context
0:12:30triggers the sarcastic reply here
0:12:33so we evaluated i mention weight with a crowdsourcing experiment on amazon mechanical turk where
0:12:38we asked i don't parse that we
0:12:41poster sarcastic reply on the context from the discussion forum and we have that are
0:12:46those that are can you identify one or more sentences from the context
0:12:52but you think that we got the sarcastic reply
0:12:54and we selected the
0:12:56dark colours where like really
0:12:59much higher qualifications so we thought that due to be able to do that starts
0:13:04we're and with our own the eighty five heats here and we selected only the
0:13:07post with the context length are between three to seven sentence because we think that
0:13:11if we put context much longer which will be a hard task for that are
0:13:14percent they might be interested to do that
0:13:17and we use the major regenerating and interestingly when you compared to that can wait
0:13:22we found that
0:13:22ask for more than forty possible but i'm the sentence that got the highest
0:13:27attention weight is also
0:13:29comparable to what that are course the thinking and the put the majority voting for
0:13:33that
0:13:34so if you see into the same example that i should be for are regarding
0:13:39the reading have it up because i will
0:13:42left hand side of the sentences and the right hand side with show that heat
0:13:46map
0:13:46so for the heat map also in the left hand side that engine the weight
0:13:49and the right hand side it just how we represent the majority voting
0:13:53we see that the for sentence was selected from that can generate and also but
0:13:57are cars
0:13:58but at the same time that can send don't put much weight into the other
0:14:02sentences for the target still think that
0:14:05there is some wait for the second
0:14:06and part sentence
0:14:12we also looked into to analyze that consent weight in the light of like how
0:14:17it's actually telling us about sarcasm
0:14:20and we in the paper we actually discussed what various characteristic of sarcasm but in
0:14:24the interest of time i'll us basically look into one issue which no less the
0:14:29context in country
0:14:30so you can think of the conquered context incongruity is characteristics of sarcasm and
0:14:36with this very simple example i can say that okay say
0:14:40i'll going to the
0:14:42emergency room one every week so there is a sentiment like positive sentiment loss
0:14:47but there is a negative situation like going to the abundance a room on every
0:14:51weekend and so you can think of this early in the country
0:14:54or inconsistency between the sentiment of the situation and they're a lot of research in
0:14:59a linguistics and communication and also in philadelphia where people talk about this context in
0:15:04one greedy and the think that
0:15:06this one of the very strong characteristic of sarcasm
0:15:09a here in our research we actually found out that a lot of time that
0:15:13i've been sent especially for that which you're the one example am showing
0:15:17has actually puts more weight into that includes features here so for instance in this
0:15:23example of the context which was that
0:15:26it talking about like advertisement saying right now it just totally do you press
0:15:30about c of mediocre s
0:15:32and the response was like all really got see what you're doing
0:15:35and we see that depressed and mediocre actually got more weight
0:15:40with the
0:15:41lots one responses got see so you can see that there is something some sort
0:15:44of incongruity and there are some more examples in our paper is that we
0:15:49i showed that how this context incongruity is happening here
0:15:54so at the same time we also looked into sarcasm markers are which is basically
0:15:58indicators like explicit indicators of sarcasm
0:16:02and we seems like especially here i can send a concluding much weights on the
0:16:06markers such as various multi columns anymore jeez
0:16:09you can think of and those actually actions and put more weight and
0:16:12that means the model of thinking that these are like some features
0:16:16a one interesting thing was we found out that sometime in the context you have
0:16:20a like cost like anymore tickle like a set face
0:16:23and then in the response maybe you have a like smiley face
0:16:27and that attention models actually learning that these opposite characters of emotive content may be
0:16:33small feature and that would be
0:16:34a small weight over there so that so interesting observation
0:16:38at the same data like interjections like are and those words actually have received a
0:16:44lot of attention which so
0:16:47but
0:16:48one of constant here that
0:16:51however
0:16:52plus the classification task was not forced only to run on the attention the weight
0:16:56and it's again
0:16:58not does it say tall in two thousand fifteen to have discussed
0:17:02on this topic that
0:17:03the integration based on any attention rate has to be taken
0:17:07care taken with here and so you have to be little courses here because
0:17:10the classification was not only working on this with
0:17:13so
0:17:15we're interpreting here and we are mostly like observing the races
0:17:19so to confuse
0:17:21we found out that for sarcasm detection when you're using contextual information from the dialogs
0:17:25it shows like better accuracy
0:17:27also we i think to identify what portion of the context may trigger the sarcasm
0:17:32and we ran some experience on the interpretation of the context to analyze a different
0:17:37attitude of such as a like context incorporate the n
0:17:39other characteristics or read discussed in the paper
0:17:43a sort of future work we are interested in to like for large scale experiment
0:17:47and we hope that there will be like more data released from the discussion forum
0:17:51that we can use for more parameter tuning especially like into the award and second
0:17:56sentence level able to attention models
0:18:00also we saw yesterday on posted at how the comments which are reply to the
0:18:05sarcastic was where you stand
0:18:07in this kind of analysis so we also interesting
0:18:11another thing here to observe you have that the three the data is a self-labeling
0:18:15sarcastic data so people are
0:18:17posting to its and they are posting the highest tax themselves
0:18:21but in the discussion forum the sarcasm was perceived to because you have other annotators
0:18:25coming and then they are not doing so
0:18:28we are a very interesting to do this sort of analysis that what's the difference
0:18:31between cells level and possible sarcasm
0:18:35and finally we follow the in our experiment after doing a lot of error analysis
0:18:39that
0:18:40there are specific aspect of conversation like you more than on that users have used
0:18:46many times and when we also downloaded or like the mechanical turk or they were
0:18:50actually able to identify those here morrison ones which could actually trigger sarcasm
0:18:55but i want model actually couldn't so that's one interesting direction that
0:19:00we hope that will continue or research into
0:19:03thank you
0:19:11you questions
0:19:16you like this slide on an attention experiment i guess for you had the one
0:19:20example
0:19:23a series of tweets and i think you where you compared against a crowdsource workers
0:19:27this can you in that i think i
0:19:30this is like for attention or no so this is like a we're showing the
0:19:34discussion forums one so we experiment with that are course only on the discussion forum
0:19:38and we
0:19:39proposed like the sarcastic pose
0:19:41and also the context for the context for like multiple sentences so we have that
0:19:45doctors that
0:19:46okay like
0:19:47we already telling that this is a sarcastic or stand this the context that actually
0:19:51has triggered the sub doesn't
0:19:52so how about that can identify the tweet sentence is more important
0:19:56and we just like five doctors and they put their voting on the sentences that
0:20:00they think that okay it's one is more important than is to an effect your
0:20:04and here the right hand side it's sort of the heat map but it's showing
0:20:08the method of importing
0:20:10and then we compare the majority voting with that mention that was and here for
0:20:14this to be example yes one
0:20:16actually has got a maximum attention the weight from or experiment at the same time
0:20:20that are first thing that actually also like doctors think that this of the thing
0:20:24is that actually
0:20:25i might be triggering the sarcasm
0:20:48in this
0:20:49sponded then
0:20:52again
0:20:56so this the macro-average f
0:21:00sorry the macro-average if once we are showing here between the sarcastic and on circuits
0:21:04to capture
0:21:10so we basically to the macro-average if one from the
0:21:14both binary classes so in the paper we actually shortly
0:21:18separately all categories but just for the interest of time taken that macro-average report
0:21:30so c plus are means that we're doing the c is like the context
0:21:34and the response and the all like all the response model
0:21:39is that
0:21:48right this normal questions then that's thanks to john again