0:00:15alright first let me thank you for the invitation and the opportunity to
0:00:20to come to all the modes
0:00:22it's so funny because a friend of mine saying all you going to the middle
0:00:25of nowhere i said no i'm going to the middle more idea
0:00:30and i really enjoy coming to new places that i've never been to
0:00:35so i talk about thirty is
0:00:38and new trend sort of technology trend that is really stripping merging and taking off
0:00:43and that is this notion of anticipatory search and how much a speech can contribute
0:00:49to that
0:00:52here sort of our region imagine you having a conversation with a friend and she
0:00:56says only to atone in spitting five minutes and as and putting down the phone
0:01:00and i'm and i look at the screen this is what i wanna see right
0:01:05i wanna
0:01:06basically have the directions of word to any to go and what do we need
0:01:10to be in five minutes
0:01:12and if you think about it we can have all the pieces already right would
0:01:16have user location we have good maps we have good directions we have speech recognition
0:01:22we have some reasonable understanding and so it's kind of a matter of putting it
0:01:27all together into one compelling application
0:01:32so that's kind of the premise we realize that the way that you find information
0:01:37is changing
0:01:39and we're moving towards kind of a query free search in the sense that instead
0:01:44of having to you proactively when you have a something going to find out having
0:01:48to fire up a browser final
0:01:50finally a search box and type in your query getting results it can be much
0:01:55more proactive you when you're context and what you've said and what where you are
0:02:00the information can come to you as opposed to are you having to find information
0:02:06but of course we're not alone in this in this idea
0:02:09recourse well the technology isn't future is that recently joined google had is a has
0:02:15a pretty similar vision so as search engines may be seen also search engine is
0:02:20that they one weight to be ask questions
0:02:22so releasing in our conversations
0:02:25what we say what we
0:02:26right but we would we here and they want to submit are needs
0:02:30and that's
0:02:31remotes that the same premise that
0:02:34expect maps was built on
0:02:36so let's look at some of the enabling trends
0:02:40for and to separate research
0:02:42there's mobile devices
0:02:44there's a i that is making progress
0:02:47and then and so if you put it together there's applications that can take contextual
0:02:52information and start making good predictions about what the user what informational needs of the
0:02:57user might be
0:02:58so like let's look at these you know in more detail
0:03:02it's obviously not surprise that
0:03:04about the whites sre you could as you can probably go anywhere
0:03:08to and you know a few minutes later there's a couple of
0:03:12you
0:03:13videos on youtube already about that event and you know hundreds of pictures the in
0:03:18fact there's technologies now that are trying to recreate some sort of a three D
0:03:22map just based on the fact that you have images from different point of view
0:03:27so
0:03:29then there's the amazing sort of growth of mobile devices so this is a statistic
0:03:35for are smart phones and tablets both running
0:03:38i O S and an and right and of course the absolute count there's us
0:03:43in china because of the
0:03:44population that have the highest up numbers but if you look at the growing market
0:03:49is basically southeast asia and on and stuff so the merrick and some other a
0:03:54growing market
0:03:56so
0:03:57we're ending up in a position where pretty much any adult is gonna have
0:04:03the smart phone in their pockets
0:04:05and so that really changes to the possibilities of what you can do with that
0:04:11because this martin this is mobile devices have a lot of sensors and you can
0:04:16think of well of course we have cameras we have microphones this why there is
0:04:22a gps
0:04:23but also if you look closely for example in this so
0:04:26let's see is for there's gestures sensors proximity sensors covers politics or amateurs
0:04:33there's even a humility sensor so that you could drop your phone in the water
0:04:38they can what the warranty
0:04:41and
0:04:42barometer
0:04:43so basically it turns out that this device is that we are not pockets in
0:04:48so to some extent no more about where we are then we ourselves might be
0:04:52aware
0:04:56and there's more right
0:04:58we all know about sort of logos of that also has
0:05:02you know bone-conduction transducer in addition to well other stuff and then more futuristic things
0:05:08right like there's research actually by
0:05:12and you hear unusual that is able to do recognition just based on the other
0:05:18facial must look activity right you have these sensors so i could be talking and
0:05:25i said without formation a you'd be able to still recognise so in fact i
0:05:30was talking to well to marry that may be an interesting challenge
0:05:34for some
0:05:36feature and used evaluation
0:05:39then there's this more you know three stick a electro and the follow gram headsets
0:05:45that it still kind of you know not very clear what you can do with
0:05:50them but they're becoming more stylish so people might start wearing them
0:05:55and then there's interesting things like this happen application from what roller
0:06:00where
0:06:01basically they have this idea that we all the nowhere an electric a tattoo here
0:06:07nor next
0:06:07that is gonna have the microphone and you can help also with speech recognition
0:06:13there's all kinds of ideas about how to
0:06:17collect more data about what we do in where we are
0:06:21and then there's sort of progressive in the back and right once we get this
0:06:25information what can we do with it
0:06:28and there's been some talk here about how much progress we're making we're all familiar
0:06:33with this
0:06:34with this chart of the famous a word error rates for different tasks
0:06:39no are we reaching some sort of a plateau but we know that that's not
0:06:43the case because there's working dynamic speaker adaptation there's all these work in the in
0:06:49the deep neural networks that we've been talking about also work in extremely large language
0:06:53models that are making the recognition be better
0:06:58there's also some working and all you not language understanding around conversation and topic modeling
0:07:03there's a knowledge grabbed all talking a second and so if you put all these
0:07:07together with some machine learning algorithms we're getting to a point where can be
0:07:13start to be reasonably good at understanding
0:07:16a human conversation
0:07:19so
0:07:20this is in this audience this is this is obviously very well known but it
0:07:24is gonna remarkable that we now have
0:07:27these a fairly substantial improvements in down to convert accuracy things to these
0:07:33do you will networks and there's work here from microsoft ibm google and there's
0:07:37others in the room that are working on this
0:07:40something that you might not be as familiar which is the fact that deep learning
0:07:45is also being applied to not a language understanding
0:07:49and i would
0:07:51when you to
0:07:52but to make sure that you're aware of the so called down for sentiment treebank
0:07:56was recently released by at stanford university
0:08:00and there's is a nice paper recursive give models for semantic compositional at over sentiment
0:08:05treebank by other soccer and was also i mean the same group as andrew on
0:08:09and on chris manning
0:08:12and what they do is
0:08:14the
0:08:16they published made available this corpus all over eleven thousand annotated utterances where they've been
0:08:25parsed in this binary parse tree and then every node is mean annotated with the
0:08:30sentiment about whether it's from very negative neutral prosody very positive
0:08:37and so the and then to the interesting part is
0:08:41how
0:08:42they man so the make they make use of theme multiple
0:08:48layers you know deep neural network to actually model the saying that the levels in
0:08:54a parse tree
0:08:56so that bottom-up can composition really fine the sentiment about a value at any you
0:09:03know by doing these steps
0:09:05so for example if you look at the sentence this film doesn't care about cleverness
0:09:09weirder and you know that kind of intelligent humour
0:09:12there's words like humour the case of plus a very positive one intelligent also so
0:09:17this whole parse-tree
0:09:19we sparsity
0:09:20except when you reach the negation just doesn't
0:09:24care about these and so the overall sentiment is negative
0:09:28and this is very powerful because after now the traditional model has been back of
0:09:33words
0:09:34a vector space and it's
0:09:38heart to model these relationships and
0:09:41we all know that up
0:09:42language has a deep structure it's kind of a recursive structure and
0:09:47there's is long distance relationships with
0:09:49certain modules within the sentence
0:09:51that are harder to capture enough
0:09:54in unless you
0:09:56really get a false sense of the parse tree
0:09:58so applying this
0:10:01they gate getting gains of well
0:10:05you know what twenty five percent
0:10:07improvement in the
0:10:08accuracy of the recognition of the sentiment over these this corpus which by the ways
0:10:13about movies this is from
0:10:15movie reviews
0:10:17so that so encouraging that
0:10:19that this technique that is not popular enough asr can also be transferred to natural
0:10:25language understanding
0:10:27then there's another a very important train
0:10:30the way i seed in how we can improve that which understanding
0:10:35and
0:10:36just of all these earlier today with saying well the kind of the you in
0:10:40asr use gone of missing a bit
0:10:42i think knowledge graphs a really the answer to that
0:10:46and wise that well because
0:10:48we can go from this kind of disembodied strings
0:10:52two and kurt entities in the real world right there is a nice but possible
0:10:57that says from strings to thinks
0:11:00so what that what is what is that
0:11:03and knowledge graph really you can think of it as these giant network what the
0:11:08nodes are concepts and then they're slings that really one entity to another for example
0:11:13you know george cloning appears in ocean's twelve and you know this is movies and
0:11:19an actors
0:11:20and how they're really to each other
0:11:23and the interesting part is if you know some history
0:11:27you might remember psych
0:11:30which was an attempt was still open sec still exist
0:11:34it's an attempt to kind of create these very complex representation of
0:11:39all known human
0:11:41knowledge especially strip common sense
0:11:44but the problem is that one is be able by hand
0:11:47and they spend a lot of time deciding whether a property of an object is
0:11:51intrinsic or extrinsic
0:11:54kind of splitting hairs a something that is not there so it quite relevant the
0:11:58way that this knowledge graphs are being built now is different
0:12:01you will start with
0:12:05start with wikipedia
0:12:08and there you know
0:12:09there's a at the data sets of machine readable version would you pdf that you
0:12:13can ingest and then you can start extracting these entities and the relationships and there's
0:12:18some certain degree of money alteration we can get pretty far with an automatic process
0:12:22and so companies are doing this
0:12:25and
0:12:26for example has knowledge graph that has ten million entities and thirty million properties in
0:12:32time you know connections microsoft have their own the court's authority and they have three
0:12:36hundred billion entities
0:12:38well five have a five hundred twenty million and it is an eighteen good properties
0:12:43and then there's also more specialised ones
0:12:46like factual for example which is a database of places point of interest local businesses
0:12:52and they're also getting to sixty six million entries
0:12:56in fifty different kind
0:12:58and then of course you can take social media
0:13:01and see their origin of entities and relations use which is people as a as
0:13:07the version of a knowledge graph and so linked units now what twenty five million
0:13:11users and facebook is over a billion
0:13:15so
0:13:16if you think carefully about these it means that
0:13:20anytime the do you relate to what concept
0:13:23or named entity like a place robotically organisation or person
0:13:27you could actually you're able to grab that and map it onto one of these
0:13:32entities
0:13:34so that the traditional idea more in the linguistic side of
0:13:39we do part-of-speech and we find this subject and the object
0:13:43we can is they'll be some relationship
0:13:45but this is still not really it's groups
0:13:48i a bit easier material with the knowledge graph you kind of and for these
0:13:53and say you're referring to this movie you're bring to that person and then there's
0:13:58all kinds of inferences and disambiguation that you can do all
0:14:02without knowledge right
0:14:04so
0:14:04i think to the fact that we can start to represent pretty much more human
0:14:09knowledge
0:14:10at least in the terms of sir
0:14:12concepts and entities
0:14:14in a way that it's read fit you know you know you know a commercial
0:14:18representation is very important and that's very big step towards real natural language understanding because
0:14:23it's more grounded
0:14:27one of the usages
0:14:29for
0:14:31for a knowledge graphics for disambiguation and there's is classic sentence from linguistics rate i
0:14:38saw the men on the keel
0:14:39with the telescope
0:14:41that can be interpreted in a variety of ways similar which are depicted in this
0:14:45funny graph right so it's what the linguists call a prepositional phrase attachment
0:14:51problem is it
0:14:52with a telescope is it attached to the hill or to the man
0:14:56or to me and on the hill again does it types of the manner to
0:14:59me so
0:15:02traditionally there's been really no way to solve this except for context but if you
0:15:07think about imagine that you have access to my amazon purchase history
0:15:14how do you and you saw
0:15:15but i just bought a telescope you know two weeks ago pen you would have
0:15:19a kind of a this idea of the priors right you could have a very
0:15:22strong prior that it is me who is using the telescope to see the man
0:15:26on the hill
0:15:27so
0:15:28it's obvious that the more context and the different sources of this context that we
0:15:33can have access to
0:15:34gonna help disambiguate natural language
0:15:37that's context in one aspect and then gonna with different idea is that we also
0:15:42know that you're intent and what you're looking for also depends on where you are
0:15:47so that's another
0:15:48place where
0:15:50location now is important contextual location
0:15:54this is this is not new there's a bunch of companies that are using for
0:15:58example exploring the yours as location local search obviously by sort for japanese restaurants depending
0:16:05on where i am gonna get different results
0:16:07one yell for example
0:16:09then there's also company select employee i that focus on
0:16:12sort of predicting what you might need based on your calendar entries there's Q at
0:16:17startup that was recently part by apple also in this space and then there's also
0:16:21obviously google now
0:16:22that
0:16:23sort of
0:16:24use able to ingest things like your email and makes sense at and understand that
0:16:29you wanna have a flight or a hotel reservation and then take it makes use
0:16:32of that information to bring a relevant alerts when the time is right
0:16:38and finally the last piece is the recommend or systems right we're all familiar with
0:16:44things that they like and amazon you get recommendations for books depending on the stuff
0:16:48that you've but for
0:16:49and the way the systems work is kind of semantic like a lot of spell
0:16:53data but the users and then they class of the users and see all your
0:16:57similar to these users so you might also like this on the book and this
0:17:01is expanding for your net flicks from movies and or an spotty five for music
0:17:05a link in facebook for people that you might know et cetera so
0:17:10all these
0:17:11systems are using context to kind of make predictions or anticipate things that you might
0:17:17mean
0:17:18so
0:17:19it is within this general context of the emergence of anticipatory sort that we start
0:17:26this company and expect laps is the technology company based in san francisco
0:17:31that we start about
0:17:32twenty five years ago
0:17:34with this idea of creating a technology platform that especially designed
0:17:41for
0:17:42this real-time applications that are gonna be able to ingest a lot of states
0:17:47give you relevant contextual information
0:17:50so
0:17:51in sort of run step
0:17:52the way works as we
0:17:55are able to receive
0:17:57it's real time and dates about what you are
0:18:00what you might be saying
0:18:02what you reading like on a new email
0:18:05and you can
0:18:05assign different weights to some of these modalities right so something but i say or
0:18:10something that i treat is gonna have a higher
0:18:14wait and something that
0:18:15i'm an email that i receive which i may just sort of scheme or read
0:18:19as opposed to
0:18:21deep
0:18:22read deeply
0:18:23so but we take all these inputs in real time and this allows and we
0:18:28process then we extract important pieces of information from all the sources and that creates
0:18:33dynamic model our best representation of what the user is doing and their intent and
0:18:40therefore were able to
0:18:42all sorts cap for information across many different data sources to try to provide information
0:18:47there's gonna be useful to that user at that point i
0:18:52and as a forty example of this platform
0:18:55which created mine mel
0:18:58mind meld it's right now and i put our
0:19:00that understands or conversation
0:19:02and fines content as you speak
0:19:05you can think a little bit of the sky where you can invite people and
0:19:09start talking
0:19:10and then we'll get
0:19:12interesting content based on that
0:19:16and all gonna give a demo in a second
0:19:19important aspect of the design of my mlps that we wanted it to make it
0:19:22very easy to share information because if it ever tried to have a kind of
0:19:27a collaboration session a using sky people quickly find especially the i
0:19:32on the ipod that it's difficult to say you wanna share a an article you
0:19:36have to leave the sky at have to find a browser or to some searches
0:19:41and then you find sort of the url and then to try to send the
0:19:45url thrust of the sky i am which may or may not be active and
0:19:49so it's a bit cumbersome so we wanted to
0:19:52make it very easy for users to be able to discover
0:19:55to
0:19:57to navigate and then to share information
0:20:02in the stuff that you share becomes a permanent archive of the conversation then you
0:20:07can look back to use
0:20:10right so with that things that
0:20:13when a give it a little demo
0:20:18my email
0:20:20see how that
0:20:21works
0:20:24so this is my ml and you can see that i have access to
0:20:27some of the sessions or conversations that have taken place in the past we can
0:20:33think of you may have a recording meetings like every tuesday you have your update
0:20:38with your colleagues and so you would joint that section because everybody's already
0:20:43invited
0:20:44and plus you can have all the context
0:20:47all the things to
0:20:48the shared items and the and the conversation that when that was previously happening that
0:20:54session
0:20:55but for now i'm gonna start a new session
0:20:59and i can give a name
0:21:03learn what's
0:21:04i can make it friends only
0:21:07what can make it public rights invite only
0:21:11and
0:21:16it's if the connection works
0:21:23this is now making a call to facebook
0:21:25the face at i
0:21:27that
0:21:28okay here we go so
0:21:31let's say that i will invite alex
0:21:35likes my able
0:21:37okay
0:21:41so
0:21:42now what i'm the only one in the conversation and so otherwise if as soon
0:21:47as alex joins you would also see
0:21:49information about the speaker right
0:21:51you know the thing that we found when you talk to people like no
0:21:54web text run to
0:21:56on the
0:21:57on some sort of a conference call
0:21:59people tend to kind of google each other and find the lincoln profile well here
0:22:03is in which is you that to you right
0:22:05and this is a discovery screen so i'm the only one seeing this information
0:22:11but if i decide to share then everybody else in the conversation would see that
0:22:15which is why for example
0:22:17you know they find the current location
0:22:21of the user right here in the
0:22:23in this whole what's congress hotel
0:22:28so
0:22:29the most interesting parties
0:22:31when you have multiple speakers but for now i'm just gonna give
0:22:35so we will real demo of how this looks like
0:22:40okay mine mel
0:22:46in mind meld
0:22:49so was
0:22:50wondering a whether you by some part about present no batman's brain mapping initiative
0:22:56i so this new technical clarity that makes brains transparent
0:23:00that might be a help for L
0:23:02for these mapping initiative
0:23:12so
0:23:12you can see that you know the we show you that about a ticker items
0:23:18here of
0:23:19what we own what we recognise we try to extract some of the of the
0:23:23key phrases
0:23:26and
0:23:27and then we know we do some post processing and bring irrelevant results
0:23:33see what else
0:23:36okay mine mel
0:23:38so we're gonna have some friends over maybe we should cook some italian food
0:23:43it we can do a mean a strong to so
0:23:46fitted you know for it
0:23:48maybe that would be nice
0:24:01so you can see the mean wait works
0:24:06if i like this for example i can drag
0:24:10and share it
0:24:11and this is what becomes part of the of the archive
0:24:15which then everybody in a conversation C and also becomes experiment archive but i can
0:24:20also access through a browser
0:24:28anybody has a topic or something that might be interested in
0:24:42i is okay my mel so paper more anyways interested in deep belief neural networks
0:24:48that's something that we've been talking about
0:24:51at this L ieee asru
0:24:54conference in other modes
0:25:12so
0:25:14one of the issues is i think pattern i are not connected in facebook
0:25:19because otherwise we would have found
0:25:22the right "'cause" are model
0:25:25i
0:25:36however if we are
0:25:42not even this one okay
0:25:44this is but you can see right so something
0:25:49let's stick to ieee okay i
0:25:54so one of the things that we do is we do look at the intersection
0:25:58of the social graph of the different participant you know call
0:26:01so that we can then
0:26:03be better at
0:26:06disambiguating
0:26:07no named entities right so
0:26:09so if we had been connected and
0:26:12pay a pit on brno would have been the real they don't know what in
0:26:15right here
0:26:23alright so
0:26:25but
0:26:27let me go back to the
0:26:29presentation real quick here
0:26:31so
0:26:32this is the platform that we've than the we build and
0:26:36if you wanna sort of
0:26:39dig a little bit deeper
0:26:41one of the novelties i think is that were combining the traditional and all P
0:26:45with a more we call and of search style approach
0:26:50because the interesting part is that were able to model
0:26:53semantic relevance
0:26:55based on the context
0:26:57the what we're speaker least be easily set and the user model and also from
0:27:01the different data sources that you can you have access to
0:27:05so basis something like work we go for dinner and then the other person says
0:27:09i don't know you like japanese sure any good base around union square
0:27:13we're building these incremental context
0:27:16about the overall intent of the conversation
0:27:19and so
0:27:21were able to then you know
0:27:23do natural language processing the usual stuff part-of-speech tagging noun phrase chunking named entity extraction
0:27:28anaphora resolution semantic parsing topic modeling and some degree of discourse modelling and pragmatics
0:27:35but then the or the piece is that depending on the signal
0:27:39that we get from each of these different data sources and you can think of
0:27:42my social graph that was mentioning
0:27:45the local businesses that factual or el can give you
0:27:48personal files right you give us access to drop box or to europe will drive
0:27:54we can make take that of the data source
0:27:57and then there's more the more general web with needles and general content and videos
0:28:04but what's interesting is that even this the response that we get when we do
0:28:08all these searches
0:28:09that also informed as about what is relevant and what is not
0:28:13about that particular
0:28:14you know conversation
0:28:17put in other words if for example you work to build an application that only
0:28:20deals with movies and T V shows an actor stand any reference to something else
0:28:25that would not find a match
0:28:27would basically not give you
0:28:28results
0:28:29but that also means that would be much more precise right in terms of the
0:28:34answers the that you give the relevancy of the content
0:28:38in so this is something that
0:28:40because we have well
0:28:42kind of very scalable and fast backend
0:28:45allows us to do multiple searches
0:28:48and we have some cash as well but basically these
0:28:50makes as
0:28:52be able to compute the semantic relevance of an utterance never a dynamic way
0:28:56based on context and also based on the type of results that we obtain
0:29:02so this is a you know technology conference so what tech technical conference some of
0:29:07the ongoing R and D as you can imagine is quite substantial
0:29:11in the on the speech side
0:29:13there's
0:29:14we have two engines we have an embedded engine that runs on the ad
0:29:18and also we have passed club a speech processing so an interesting
0:29:22research is you know how to balance that and how to
0:29:27how to be able to on the one hand listen continuously put on the other
0:29:31also be robust to network issues
0:29:34and then there's in terms of practical usage there's things that you can imagine detecting
0:29:38sub optimal audio conditions like when the speakers so far on the mic noise environments
0:29:44as we all know heavy accents are an issue
0:29:47and then
0:29:48one of things we found is because is an ipod at it's very natural for
0:29:51people to kind of leave it on the table and two things happened they speak
0:29:55to each from far away and also the can be multiple people
0:29:58speaking on you know to the same device and our models try to do some
0:30:02speaker adaptation
0:30:04and sometimes that doesn't work that well
0:30:08and then sort of the issue with this kind of the holy grail of could
0:30:11we detect you know a sequence of long
0:30:14and grammatical works and
0:30:18when he's gone of you bridge
0:30:19and of course there's techniques to do that but
0:30:21we're trying to get
0:30:23improve the accuracy of that
0:30:24and then in terms of natural language processing in information retrieval also kind of a
0:30:28design question are things like the class i cannot P problems like word sense disambiguation
0:30:33although obviously the knowledge graph helps a lot
0:30:36and then enough resolution and some of these things we do with the social graph
0:30:42an important aspect is
0:30:43these knowledge graph is useful but
0:30:45how do you dynamically updated how do you keep it fresh
0:30:49and we have some
0:30:50some techniques for that but it's
0:30:53it so
0:30:54ongoing research
0:30:56then every important aspect is
0:30:59deciding that the sorts working this right
0:31:02as we all know if we if you leave a speech engine on
0:31:05but i remember an anecdote from are alex waibel that you told me once it
0:31:09as an engine running in his house and then when he was doing the dishes
0:31:13with a look cling incline that you know the search engine was spouting all kinds
0:31:17of the interesting
0:31:19hypotheses
0:31:21this is been alluded to of course you can have a fairly robust voice activity
0:31:24detection
0:31:25but there's
0:31:27there's always room for improvement
0:31:30the search more than is as i mention is not just
0:31:33understanding that something is speech but also detecting of how relevant something is within this
0:31:38within the context and this comes of these other point of the interrupt ability and
0:31:45mind meld is a bit too verbose right this is just a showcase of what
0:31:49you can do also because the ipod has a lot of real state sequence shoulders
0:31:53different articles in practice and through the a i'll talk about in a second
0:31:57you have a lot of control about how like twenty one to be interrupted when
0:32:02you wanna
0:32:03a search result for an article to be
0:32:07to be shown and this is
0:32:09a function of at least two factors one is
0:32:13have
0:32:13in place in the request is how much the user ones to have certain information
0:32:18and the other one is what i was mentioning about the nature of the information
0:32:23found how strong is the signal from the data sources about the relevancy of what
0:32:27i'm gonna show
0:32:29and what i mean by that is
0:32:31you can think of
0:32:33but you by set
0:32:36the difference between
0:32:38what is the latest movie by woody allen
0:32:42versus i've been talking about woody allen in
0:32:44and i mentioned the that
0:32:46the keys latest movie et cetera
0:32:48right so one is a direct question where am the intent is clear more like
0:32:53a serial like application where and trying to find the specific information the other one
0:32:58is a reference sort of in passing about
0:33:00something
0:33:01i'm and so
0:33:02that
0:33:03would be the these understanding of
0:33:06how eager i am to receive that bit of information
0:33:09so that's work that is ongoing being able to model that
0:33:14and then finally
0:33:16we have a fair amount of feedback from this right especially when the user shares
0:33:21an article that's a pretty strong signal that was relevant
0:33:25on the negative side haven't shown you this but you consider flick one of the
0:33:31entries on the on the right on the left hand side that eager items as
0:33:34we call them you can delete them so that would be good of negative feedback
0:33:38about
0:33:39certain entity or a key phrase that was not
0:33:41deemed relevant by the user
0:33:44how to
0:33:45optimize the learning that we can obtain from taking that user feedback
0:33:49is also something that
0:33:50that we working on
0:33:52especially because
0:33:54the decision to show certain article based is complex enough that
0:33:58sometimes it's harder to assign the right sort of credit or blame for how we
0:34:03got there
0:34:06so just do well
0:34:09sort of
0:34:10twenty five what we're doing there's two products that we're offering
0:34:14one is that might melt
0:34:16my not obvious what you see here
0:34:18and as a matter of fact
0:34:19the mind meld out
0:34:21is gonna be alive on the apple store tonight
0:34:25so
0:34:27we've been working need for awhile and it's finally happening
0:34:30so if a if you're welcome to tried out
0:34:33i guess will be tonight well
0:34:36for whatever time zone you're up store a is set to so i think
0:34:41new zealand users might already be able to download it
0:34:44and then for the us will be
0:34:46in a few hours
0:34:50so that's a mimo but then
0:34:52the other thing is
0:34:54were also offering these the same functionality when api about a rest based api
0:35:00that
0:35:01you're able to well
0:35:03give this creates sessions and you and users and give this real time updates so
0:35:08that and then you can query for what is the most relevant you can also
0:35:13select the different data sources and so it any given point you can ask for
0:35:17what are modeled thing system most relevant set of articles
0:35:22with a certain parameters for ranking et cetera so we're having
0:35:27already a system
0:35:29degree of well of scoring
0:35:31how lots
0:35:32with comes
0:35:33for example some or all of our backers which include by the way google ventures
0:35:38and also sums and
0:35:40intel top twenty car
0:35:42liberty mutual
0:35:44they're all in the backers that we're trying to do some prototypes with
0:35:49so
0:35:50i'm character to try it out and
0:35:53i was thinking that
0:35:55because i'm actually gonna be missing the launch party that is happening in san francisco
0:35:59i'm gonna take our banquet that the bishop's palace as the ones party for might
0:36:04know
0:36:10that's what i want to say and we have some time for questions
0:36:32was at all
0:36:34the i was wondering i'll the track the users they the example the key we
0:36:41want to eat something and then
0:36:44is it is still sticking to the restaurant domain and me and no
0:36:49what the example you show that's all you're adding information and how about you change
0:36:55information that you previously used switch to another domain
0:37:00how you jack use
0:37:03there's to right of information that we use for that one is simply time right
0:37:08that sort of as time passes you can of so you decay certain previous and
0:37:12trees
0:37:13the other one is some
0:37:15kind of topic detection clustering the we're doing so that
0:37:19sentences that still seem to relate to the same topic kind of you know how
0:37:24help a
0:37:25sort of ground round that topic
0:37:28and then there's also
0:37:31some user modeling about you know you're previous sessions so that we have certain
0:37:37prior weights
0:37:48what
0:37:53well so you know there there's
0:37:58i'm not gonna sitting some specific algorithm that we use but you can imagine there's
0:38:03some you know statistical techniques to
0:38:06to do that modeling
0:38:09where small startup we can not like reveal everything
0:38:15so like very much so it's great another question so
0:38:21i one point you happened mentioned
0:38:26asr you and all the modes probably enough came out as a that's are you
0:38:31a slu and columbus
0:38:34no it's
0:38:36it would same
0:38:38the really what
0:38:39that what you've shown us are ways of organising information at the output and the
0:38:44process
0:38:45but also same particularly not example when the beanie the
0:38:50not only that it's actually well it does know exactly where you work it's without
0:38:55map
0:38:56and it might even figure out that you're at this thank all layers are you
0:39:00but this things we're not being reflected in the lower level transcription process so i
0:39:06was wondering how the mites you don't have to tell us anything it's buster's father
0:39:12figured and to train nice things
0:39:15well it's obviously a that the research issue of how you
0:39:20make the most of the contextual information and unfortunately
0:39:24asr specially the well the these cloud based asr
0:39:30de at this point doesn't
0:39:32fully support the kind of adaptation and comp and dynamic modification that would like to
0:39:38do
0:39:39but that's kind of a and an obvious thing to do in the same way
0:39:43that you constructs larger contexts and fine you know the all the people that you're
0:39:47related to and at that you're specific lexicon having something like the location and the
0:39:52towns nearby or would be something
0:39:55very no sort of natural to do
0:39:58but we're not being this
0:40:02i have to say your search for better more innocent implement because
0:40:06the previous one used to be a and the step and it has so that
0:40:10you made
0:40:11so when you search for permanent you go okay so this is better
0:40:16well that the asr was no hundred percent accuracy
0:40:20which one to use
0:40:23actually we use the writing including a new ones and cools
0:40:35sex for a talk also pitch wondering about privacy occurrence i was on those impression
0:40:41that's the more
0:40:43i want to
0:40:45interact with this mind meld at some live in or a need to be transparent
0:40:50for the for that and my personal data
0:40:57well i have a
0:40:59actually a philosophical reflection
0:41:02that
0:41:02as a society with this technology we are going to words what i'm calling with
0:41:07transparent bring
0:41:08a and
0:41:10if you think closely about it up
0:41:12the better we are collecting data but users and modeling their thing intentions
0:41:18we can get to a point where
0:41:21you can almost all of complete your thought
0:41:24right assume that you start typing the query and gonna be knows what you might
0:41:27one
0:41:28and of course is just a little bit of science fiction but
0:41:31we're kind of getting there and so i think the way to address that is
0:41:35by doing very transparent about this process
0:41:40and giving you full control that what is it that you wanna share for how
0:41:43long
0:41:44because
0:41:45that's really the only way to modulated it's not just say one gonna opt out
0:41:49and not just gonna use
0:41:50any of these
0:41:52anticipate research because basically will be unavoidable right but so i think it's
0:41:58it's
0:41:59what we need to do is how well some clear
0:42:03settings about
0:42:04what you wanna share with this out for how long
0:42:07and then insuring the back and that that's really
0:42:09the only way the only usage
0:42:11of that information
0:42:15but as an example
0:42:16we're not recording
0:42:18this
0:42:19the voice rate
0:42:20and is the only thing that is permanent in this particular mind all application
0:42:24are this the articles that you've specifically share
0:42:28that's the only think that
0:42:34so i'm happy that maybe if you're looking at pedro if you are task pedro
0:42:40in police record would you see something
0:42:44it may be you wouldn't wanna see so is there anyway i like when you're
0:42:47looking at your space
0:42:48do you have certain
0:42:52contexts that you're searching for things when you bring information back like let's say you
0:42:57know this descending order social setting or some other context
0:43:02yes so what one of the shortcomings of the little demo idea is first of
0:43:06all you was only one speaker it's always more interesting when israel conversation
0:43:10and the second is it wasn't really a long ranging conversation about certain topic where
0:43:16mine mel excels at least in say you wanna
0:43:20planned application with you know some of the frames are somewhere else and you sail
0:43:24well gonna go here then you explore different places you can stay things you can
0:43:27do and you share that when you have
0:43:30a long range in conversation with this with the kind of overarching goal
0:43:34that's where it works the best if you keep sort of switching around then it
0:43:38becomes more like a serial like search that doesn't have much
0:43:42in just a quick question so how do you build your pronunciation so if you
0:43:46look at asr you would spell line out that if you look at icassp you
0:43:50actually see it doesn't work
0:43:52that's it's mostly in the lexicon there's certain
0:43:56abbreviations there are more typically
0:44:00separated like you know i guess are you or some other ones like need to
0:44:02alright guys that would be a spoken is a war
0:44:05it's so it's becomes in the pronunciation lexicon pretty much
0:44:12you more questions