| 0:00:26 | oh |
|---|
| 0:00:27 | hi |
|---|
| 0:00:28 | this is uh a joint work with uh four |
|---|
| 0:00:30 | back from my in yeah this shot team in a selective but |
|---|
| 0:00:33 | and a |
|---|
| 0:00:34 | from a you know as an |
|---|
| 0:00:36 | just sticks team from a to to comply that can we uh |
|---|
| 0:00:39 | i'm not and going |
|---|
| 0:00:40 | to talk about |
|---|
| 0:00:42 | tech was say to a non-negative matrix factorization with group sparsity so uh there have been a several talks about |
|---|
| 0:00:47 | the the quite site two |
|---|
| 0:00:50 | a negative matrix factorization so |
|---|
| 0:00:51 | uh |
|---|
| 0:00:52 | we have been working on |
|---|
| 0:00:53 | adding uh uh priors |
|---|
| 0:00:56 | uh |
|---|
| 0:00:57 | with |
|---|
| 0:00:58 | this frame |
|---|
| 0:00:59 | so i go over a quickly and non-negative matrix factorization uh |
|---|
| 0:01:04 | the the next your slides so uh |
|---|
| 0:01:07 | yeah |
|---|
| 0:01:08 | yeah you can see a a a a steep it's simple example of uh |
|---|
| 0:01:11 | a just signal |
|---|
| 0:01:13 | it's uh it's composed of a can or not |
|---|
| 0:01:15 | it's uh |
|---|
| 0:01:16 | it's a |
|---|
| 0:01:18 | uh and uh at to to |
|---|
| 0:01:20 | at |
|---|
| 0:01:20 | each each or you can see that uh first |
|---|
| 0:01:23 | for notes of can are light and then combinations of two which are |
|---|
| 0:01:27 | how money |
|---|
| 0:01:28 | which are |
|---|
| 0:01:29 | a a one up one dave uh |
|---|
| 0:01:31 | to the other |
|---|
| 0:01:32 | so uh this is a |
|---|
| 0:01:34 | this is an example of a very a very difficult to and source separation a compact |
|---|
| 0:01:39 | and uh what we can see here is that the so you have the data and uh a money to |
|---|
| 0:01:43 | matrix factorisation in learning |
|---|
| 0:01:45 | a basis dictionaries so with the basis spectra |
|---|
| 0:01:48 | and they time activations yeah we can see the dictionary and the time activations and you can see that |
|---|
| 0:01:53 | uh a very clearly you can see a the the notes |
|---|
| 0:01:56 | are uh a separate it and you can see that the make actually since a very |
|---|
| 0:02:00 | or easily so you can see that for notes are played together and then combination |
|---|
| 0:02:04 | of to notes |
|---|
| 0:02:06 | and uh uh there are still two components that are left one that explains the one that explains the noise |
|---|
| 0:02:12 | and one that you can see here |
|---|
| 0:02:14 | but uh uh so |
|---|
| 0:02:17 | if we could listen to it uh sounds like the hammer |
|---|
| 0:02:19 | of the of the channel |
|---|
| 0:02:21 | so this is an example of a or where uh the cross a to nonnegative matrix factorisation works are really |
|---|
| 0:02:26 | where i i'm seven example of a |
|---|
| 0:02:29 | a a nonnegative matrix factorization |
|---|
| 0:02:31 | using another a a a a uh another all us which is the euclidean us |
|---|
| 0:02:35 | so here's the the same |
|---|
| 0:02:37 | uh it's the same type plots |
|---|
| 0:02:39 | uh except |
|---|
| 0:02:41 | you can see that uh for um |
|---|
| 0:02:44 | or the first not reece is the the thought components here |
|---|
| 0:02:47 | uh you can see that the the top component gets split up with other components |
|---|
| 0:02:52 | so the separation is not so good as a as before |
|---|
| 0:02:56 | and that is explained by the fact that uh a take a to uh a measure of divergence |
|---|
| 0:03:00 | is uh more sensitive uh most since if to to uh |
|---|
| 0:03:05 | to high frequency uh and to choose |
|---|
| 0:03:07 | so it seems a |
|---|
| 0:03:10 | but a suppression for |
|---|
| 0:03:14 | so uh now if we want to more complicated uh |
|---|
| 0:03:18 | well just signals |
|---|
| 0:03:20 | uh |
|---|
| 0:03:21 | a problem uh a appears that uh |
|---|
| 0:03:24 | if uh if you have only two sources uh each source can "'em" meets several uh several different spectra for |
|---|
| 0:03:30 | example when i speak |
|---|
| 0:03:31 | uh there are so spectra but you can associate my course i'm not only |
|---|
| 0:03:36 | uh always saying that be the same thing |
|---|
| 0:03:38 | and uh so there is the problem of grouping the components |
|---|
| 0:03:43 | uh into two sources assigning several components to sources so |
|---|
| 0:03:47 | uh uh for instance you can you can simply a run and M F and look at uh the activation |
|---|
| 0:03:51 | coefficient |
|---|
| 0:03:52 | okay can see matrix H and you can see that the |
|---|
| 0:03:55 | in this uh in this very simple or uh example where are you have a base |
|---|
| 0:04:00 | and uh yeah |
|---|
| 0:04:01 | and there are overlapping |
|---|
| 0:04:03 | this the region |
|---|
| 0:04:04 | you can see that the for some components there is already |
|---|
| 0:04:08 | uh very clear means that that these components second be assigned to the base |
|---|
| 0:04:12 | and uh other components to uh |
|---|
| 0:04:15 | data |
|---|
| 0:04:16 | so uh one approach is to look at the dictionary and is are guided by a stick or just uh |
|---|
| 0:04:22 | with the yeah |
|---|
| 0:04:23 | uh the an engineer can uh |
|---|
| 0:04:27 | the design the best uh the best grouping of components and |
|---|
| 0:04:30 | two sources |
|---|
| 0:04:32 | but uh the problem is that |
|---|
| 0:04:34 | as the tracks get longer as you get a a a more tracks |
|---|
| 0:04:38 | and uh also as the dictionary get larger |
|---|
| 0:04:41 | uh is because more complicated so for the engine now because there is a a lot a more work to |
|---|
| 0:04:45 | do |
|---|
| 0:04:46 | uh uh and if you are used uh a get it by heuristic |
|---|
| 0:04:49 | uh this series stick will involve a considering or permutations |
|---|
| 0:04:53 | of of the of your generate so uh if you have five permutation mutation you have a factor your five |
|---|
| 0:04:58 | a limitations |
|---|
| 0:04:59 | to see too small |
|---|
| 0:05:01 | but if you want to ten |
|---|
| 0:05:02 | twenty uh components and of the jury this becomes |
|---|
| 0:05:05 | uh |
|---|
| 0:05:06 | wait |
|---|
| 0:05:06 | wait too long it uh so you would run uh and an F of thoughts to seconds and would spend |
|---|
| 0:05:11 | one day |
|---|
| 0:05:12 | considering all the permutations |
|---|
| 0:05:14 | a source of so |
|---|
| 0:05:15 | uh |
|---|
| 0:05:16 | with the |
|---|
| 0:05:17 | want to do is to include the grouping in the learning of the of the dictionary |
|---|
| 0:05:22 | so um |
|---|
| 0:05:24 | when way of uh when we have thinking uh a how to group for the components is to uh is |
|---|
| 0:05:28 | i think about the the the some levels of uh each source |
|---|
| 0:05:32 | at uh at a given time |
|---|
| 0:05:33 | so uh uh here uh for a given track a a uh i i are did the volume for |
|---|
| 0:05:38 | each so the base get down the voice |
|---|
| 0:05:41 | and uh |
|---|
| 0:05:42 | you can see that there are some uh he that you can use for instance uh uh at this time |
|---|
| 0:05:46 | you can see that the the basis a very low level |
|---|
| 0:05:49 | uh compared to the other sources so you could say that that's some points |
|---|
| 0:05:53 | one source is inactive or as the other as a are active |
|---|
| 0:05:56 | and also uh |
|---|
| 0:05:58 | another idea yeah is to exploit the fact that their shapes of uh is volume activations are are are very |
|---|
| 0:06:03 | different |
|---|
| 0:06:04 | so |
|---|
| 0:06:05 | uh |
|---|
| 0:06:07 | so uh |
|---|
| 0:06:09 | not coming back to the the from of was set the the notations a a little bit |
|---|
| 0:06:13 | so uh what we have been looking at uh so that that there is a you |
|---|
| 0:06:17 | of the power |
|---|
| 0:06:18 | spectrum |
|---|
| 0:06:19 | uh and at time you can consider out that uh in a model that if you of a |
|---|
| 0:06:24 | and it's eve uh you know model |
|---|
| 0:06:26 | uh there was a spectrogram |
|---|
| 0:06:28 | is |
|---|
| 0:06:28 | uh |
|---|
| 0:06:29 | is gonna but that that's the sum of uh |
|---|
| 0:06:32 | for several components |
|---|
| 0:06:34 | and |
|---|
| 0:06:34 | each component |
|---|
| 0:06:36 | uh |
|---|
| 0:06:36 | each components of the complex spectrum |
|---|
| 0:06:38 | uh it's not the gaussian |
|---|
| 0:06:40 | with a diagonal covariance and uh |
|---|
| 0:06:44 | nonnegative matrix factorization consists |
|---|
| 0:06:46 | in |
|---|
| 0:06:46 | uh computing using uh a factorization of uh the parameters of a matter |
|---|
| 0:06:51 | so uh |
|---|
| 0:06:53 | in this case in in the case of uh D uh tech why said to uh your a chance |
|---|
| 0:06:58 | this corresponds to uh i mean you gaussian model |
|---|
| 0:07:01 | uh which means uh that's we have a truly the additive model for the power spectrum runs so even if |
|---|
| 0:07:07 | uh |
|---|
| 0:07:08 | but it is not at a T for the observed or gram |
|---|
| 0:07:11 | really additive |
|---|
| 0:07:12 | for all what want to estimate that is the parameters and uh |
|---|
| 0:07:16 | it is the only model for which you can get the is uh this |
|---|
| 0:07:19 | i |
|---|
| 0:07:19 | this to be true |
|---|
| 0:07:21 | so i mean the gaussian a assumption and uh you can don't to uh looking at the power spectrogram in |
|---|
| 0:07:26 | it uh it means that actually the power spectrum is uh |
|---|
| 0:07:29 | distributed as an exponential |
|---|
| 0:07:31 | uh |
|---|
| 0:07:32 | with problem itself uh W and H got that we you use the bases dictionary and H uh that time |
|---|
| 0:07:37 | coefficients |
|---|
| 0:07:37 | the time activation |
|---|
| 0:07:39 | and so in my annotation uh H has several role and you want to uh |
|---|
| 0:07:44 | you want to find uh |
|---|
| 0:07:46 | and |
|---|
| 0:07:47 | want to uh |
|---|
| 0:07:48 | you you want to find a a a a a partition |
|---|
| 0:07:51 | of the rows of H in to uh say two groups uh but this may generalized |
|---|
| 0:07:56 | an trial are a number of rules |
|---|
| 0:07:58 | you want to find a partitions of the rows of H in two |
|---|
| 0:08:01 | so here had would be to groups we the the same number of uh the same number of uh of |
|---|
| 0:08:05 | from |
|---|
| 0:08:07 | now coming back to the coming back to the P just lies what is the volume in uh uh what |
|---|
| 0:08:12 | is the the some level of fit shots in a in a model |
|---|
| 0:08:16 | well if you assume that uh the sense of uh each column of uh W |
|---|
| 0:08:21 | sums to one |
|---|
| 0:08:22 | then the some level of one source will be the seven |
|---|
| 0:08:25 | of activation coefficients |
|---|
| 0:08:27 | of uh of a group one which corresponds to force one |
|---|
| 0:08:31 | so what we want mother is |
|---|
| 0:08:33 | uh these |
|---|
| 0:08:34 | these coefficients |
|---|
| 0:08:38 | so |
|---|
| 0:08:39 | inference we propose is to round the grouping at the same time as the factorization |
|---|
| 0:08:45 | uh so this corresponds to uh uh doing a a up |
|---|
| 0:08:49 | if close to to an F and and ink |
|---|
| 0:08:51 | we propose a adding a prior |
|---|
| 0:08:53 | uh that is that sector is in the groups |
|---|
| 0:08:56 | a all the different sources so uh |
|---|
| 0:08:59 | so yes since you have a a nonnegative coefficient this uh this uh and one um is just the the |
|---|
| 0:09:05 | sum of the coefficients of age for uh one schools |
|---|
| 0:09:07 | that is one group at a given time |
|---|
| 0:09:10 | and uh |
|---|
| 0:09:12 | and the i uh here we only assume that is a that it is a a concave function |
|---|
| 0:09:17 | uh |
|---|
| 0:09:20 | and so what this uh what this uh optimization problem tends you is that's you want to have a fit |
|---|
| 0:09:24 | to the data |
|---|
| 0:09:25 | but at the same time of uh you have a prior on that they are uh that uh |
|---|
| 0:09:31 | at a given time there is only a a that they are only a few sources that are active at |
|---|
| 0:09:35 | the same time |
|---|
| 0:09:38 | so uh |
|---|
| 0:09:40 | so in in you know that we have a |
|---|
| 0:09:43 | but that's choice for |
|---|
| 0:09:44 | side |
|---|
| 0:09:45 | but so if you if you if you look at the paper are you |
|---|
| 0:09:48 | you would see that the uh it to that it comes from a a graphical model with |
|---|
| 0:09:51 | we two layers |
|---|
| 0:09:53 | uh i |
|---|
| 0:09:54 | so |
|---|
| 0:09:55 | to much about this |
|---|
| 0:09:57 | um |
|---|
| 0:09:58 | and this corresponds actually uh so to uh |
|---|
| 0:10:02 | maximum like you an france |
|---|
| 0:10:04 | of uh of the problem of a model |
|---|
| 0:10:07 | or |
|---|
| 0:10:07 | even a a out to model of the data |
|---|
| 0:10:09 | and uh |
|---|
| 0:10:11 | a parameter |
|---|
| 0:10:12 | on H |
|---|
| 0:10:14 | um so about the inference of the parameters for the algorithm is uh in uh the the que c'est chance |
|---|
| 0:10:21 | it's uh |
|---|
| 0:10:22 | so it's very hard |
|---|
| 0:10:23 | uh to uh to have a a a and so the related methods to to the parameter on friends we |
|---|
| 0:10:28 | must the results to let's get to the date |
|---|
| 0:10:30 | uh because they go way faster |
|---|
| 0:10:32 | uh |
|---|
| 0:10:32 | here an example of uh a a a a a at the right uh window |
|---|
| 0:10:36 | running the algorithm with you know a a great in reading methods are or multiplicative at that's method and that |
|---|
| 0:10:42 | but you get you of that's them goes away faster and |
|---|
| 0:10:45 | actually are converges to but uh are a but on the pony no |
|---|
| 0:10:50 | so um |
|---|
| 0:10:53 | or or go with and uh a doesn't change significantly from a |
|---|
| 0:10:57 | stand |
|---|
| 0:10:58 | the class i two and F we just add uh |
|---|
| 0:11:00 | terms |
|---|
| 0:11:01 | which correspond to a to our prior |
|---|
| 0:11:03 | and uh since yeah size use a is a concave function |
|---|
| 0:11:08 | uh |
|---|
| 0:11:09 | you have that the |
|---|
| 0:11:11 | site in in upsets a prime uh in is with |
|---|
| 0:11:14 | uh that's some level of source one |
|---|
| 0:11:17 | so what the algorithm than that you is that that each step you are gonna a a bit H so |
|---|
| 0:11:21 | as to get to |
|---|
| 0:11:22 | a better fit of the data uh corresponding to |
|---|
| 0:11:25 | the the class a two and gets you |
|---|
| 0:11:27 | matrix like relation |
|---|
| 0:11:28 | uh |
|---|
| 0:11:30 | and the more source one uh |
|---|
| 0:11:34 | the the the less source one uh will be at a high volume |
|---|
| 0:11:38 | the more you will be uh but then broke coefficient at this time so uh |
|---|
| 0:11:42 | it means that the this uh so this algorithm |
|---|
| 0:11:45 | will push |
|---|
| 0:11:47 | uh a low amplitude sources to zero and keep i i'm should source |
|---|
| 0:11:54 | uh |
|---|
| 0:11:55 | and uh so it's on the fact that uh even if we have a a a a a a a |
|---|
| 0:11:58 | a you prior this doesn't change the speed of uh |
|---|
| 0:12:01 | this doesn't change at of the speed of the algorithm it's are compulsion |
|---|
| 0:12:04 | approximately a thousand iterations durations |
|---|
| 0:12:07 | the time uh the time customs algorithm is |
|---|
| 0:12:10 | read the same as a |
|---|
| 0:12:11 | the classic in |
|---|
| 0:12:13 | now one complicated aspect of uh i having this prior is that the you must to uh selection for the |
|---|
| 0:12:18 | i'd proper all so uh uh i a prime thousand are uh in a mother are and on that uh |
|---|
| 0:12:22 | and uh a a a of the choice of uh of side |
|---|
| 0:12:26 | so |
|---|
| 0:12:27 | even that we have a a given that we actually have a a a a a a graphical model that |
|---|
| 0:12:32 | explains the choice of this prior |
|---|
| 0:12:34 | uh we could result to uh we could is up to uh |
|---|
| 0:12:38 | a a bayesian tools to to estimate the was parameters that |
|---|
| 0:12:41 | uh |
|---|
| 0:12:43 | actually we uh we we devised a statistic it is a a a a uh |
|---|
| 0:12:48 | much more simple |
|---|
| 0:12:49 | uh to uh |
|---|
| 0:12:50 | it you on that so and it's |
|---|
| 0:12:53 | the principle of this to stick is that |
|---|
| 0:12:56 | if you become all the right palm tells then |
|---|
| 0:12:58 | V |
|---|
| 0:12:59 | given this parameter L should be exponentially distributed so |
|---|
| 0:13:03 | uh if you compute now uh |
|---|
| 0:13:06 | sadistic this thing that are we over the estimation of the you H |
|---|
| 0:13:10 | and you have a a a and and you are and you look at is uh at this a random |
|---|
| 0:13:14 | variable then it should be a distributed as an exponential one |
|---|
| 0:13:17 | and you have a a lot of samples of this because you have a a a a a a a |
|---|
| 0:13:20 | as many menu frequencies and it as many a frequency in is uh as many state is the statistics it |
|---|
| 0:13:26 | is you have a a time-frequency bins we have a lot of them |
|---|
| 0:13:28 | and uh then uh run uh computing a chroma graphs none of sadistic becomes a |
|---|
| 0:13:34 | very interesting because it it's a very cheap |
|---|
| 0:13:36 | and you can uh |
|---|
| 0:13:37 | and uh |
|---|
| 0:13:38 | we can just run a whole rid of experiments |
|---|
| 0:13:41 | and look at the parameter values for which you have the lowest a |
|---|
| 0:13:44 | we also have statistic |
|---|
| 0:13:46 | um and so we did that on uh that that to get that check a that a lot |
|---|
| 0:13:51 | or or uh |
|---|
| 0:13:52 | source so we have a |
|---|
| 0:13:53 | and the see that that to generated for from the model |
|---|
| 0:13:57 | and now you can look at uh so i we look at the different number of uh |
|---|
| 0:14:02 | uh a training sample for the mother |
|---|
| 0:14:04 | and you can look at uh at the top |
|---|
| 0:14:06 | yeah value of all set stick it is in blue |
|---|
| 0:14:10 | uh uh in red uh a measure of the mountains to good to my because uh in this uh |
|---|
| 0:14:15 | in this setting uh we have a a we generate it's synthetic that that from a non model so we |
|---|
| 0:14:19 | can uh actually compute |
|---|
| 0:14:21 | you and parameters to is the divergence good to mother |
|---|
| 0:14:24 | and yeah you can see uh uh a a class if you can should scroll also gets which can vacations |
|---|
| 0:14:29 | got is uh if that's a uh uh if a correct source one one source one is a |
|---|
| 0:14:34 | exactly |
|---|
| 0:14:35 | if we cover a a correct is source to and source |
|---|
| 0:14:38 | exactly |
|---|
| 0:14:39 | so uh when they are only a hundred observations |
|---|
| 0:14:42 | can see that the there is |
|---|
| 0:14:44 | with a good the classification accuracy but uh it is difficult to find a minimum of the is T |
|---|
| 0:14:49 | and as you increase uh the the number of points |
|---|
| 0:14:52 | in gets the the the set to get you were uh but on that there are and uh |
|---|
| 0:14:59 | more in see the the minimum of the statistic T |
|---|
| 0:15:01 | and also uh the |
|---|
| 0:15:03 | the development of a model uh says |
|---|
| 0:15:06 | yet but there are you get the rest |
|---|
| 0:15:07 | so so |
|---|
| 0:15:08 | this just means the model that i you have |
|---|
| 0:15:10 | the but are a a are our prior will uh estimate the |
|---|
| 0:15:14 | as as possible |
|---|
| 0:15:17 | this uh this is a based on to that at that time |
|---|
| 0:15:20 | we not want to uh experimental results so uh uh a a first the is to try |
|---|
| 0:15:25 | uh is to trade this in a simple segmentation task or you know that that |
|---|
| 0:15:29 | the |
|---|
| 0:15:30 | it's a given time that is only one still that that is a key |
|---|
| 0:15:34 | and uh uh a good thing is to uh compare are or them with uh just the simple idea of |
|---|
| 0:15:40 | doing a a and then F and then finding the best uh mutation |
|---|
| 0:15:44 | given a a a you given a statistics so he a a a a given a heuristic so he other |
|---|
| 0:15:48 | heuristic is uh |
|---|
| 0:15:49 | compute an and have |
|---|
| 0:15:50 | and uh find the permutation that the minimize is uh this quantity |
|---|
| 0:15:55 | this to give a faq compared |
|---|
| 0:15:57 | so this is a result of a an an F with the this a heuristic grouping so uh uh you |
|---|
| 0:16:02 | can see so the mix was |
|---|
| 0:16:04 | first uh a then speech |
|---|
| 0:16:06 | uh you can see each other sources |
|---|
| 0:16:08 | uh so that's the result of uh and an F with heuristic to groupings we can see that the still |
|---|
| 0:16:12 | a a a lot of uh missing up the |
|---|
| 0:16:15 | the sources are are |
|---|
| 0:16:17 | it's not a lot |
|---|
| 0:16:18 | and uh this is a result with a are them that's long the grouping at the same time as |
|---|
| 0:16:23 | uh the an F |
|---|
| 0:16:25 | so you i can see that uh uh the separation gets a a a a a uh uh |
|---|
| 0:16:30 | that lot uh |
|---|
| 0:16:31 | lot more yeah |
|---|
| 0:16:32 | uh |
|---|
| 0:16:33 | or |
|---|
| 0:16:35 | original result |
|---|
| 0:16:36 | and uh |
|---|
| 0:16:38 | the second experiment that run was on a a a a real of valid signals |
|---|
| 0:16:42 | so uh so |
|---|
| 0:16:43 | we took uh |
|---|
| 0:16:44 | to some from the C sec that the base |
|---|
| 0:16:47 | and we evaluated uh the quality of the separation |
|---|
| 0:16:51 | uh |
|---|
| 0:16:52 | when we vary the degree of overlap between the sources so |
|---|
| 0:16:56 | the them that up the |
|---|
| 0:16:58 | more difficult it becomes to |
|---|
| 0:16:59 | the separation |
|---|
| 0:17:01 | and and i insisting on the fact that we have no they're on the use so uh uh so uh |
|---|
| 0:17:06 | you can talk for perfect separation |
|---|
| 0:17:08 | but uh you would see that the |
|---|
| 0:17:10 | when you varies over a |
|---|
| 0:17:12 | the |
|---|
| 0:17:12 | the less of a lab you have a |
|---|
| 0:17:14 | but they'll the separation so you know you |
|---|
| 0:17:16 | very good separation for a thought T three percent on year |
|---|
| 0:17:18 | or that so that the sources is it is the |
|---|
| 0:17:21 | is is mix |
|---|
| 0:17:23 | uh this is a this |
|---|
| 0:17:24 | as the source |
|---|
| 0:17:26 | as deep dark |
|---|
| 0:17:27 | voice |
|---|
| 0:17:28 | you can see the that of thought people percent of a like to get the |
|---|
| 0:17:31 | very good |
|---|
| 0:17:32 | separation or T |
|---|
| 0:17:34 | terms of as yeah |
|---|
| 0:17:35 | and as the overlapping increases |
|---|
| 0:17:37 | it's |
|---|
| 0:17:38 | uh |
|---|
| 0:17:39 | works and mars |
|---|
| 0:17:40 | so |
|---|
| 0:17:41 | what the prior what they'll prior is that for that is |
|---|
| 0:17:44 | when |
|---|
| 0:17:46 | and uh not all sources are active at the same time it's a people the |
|---|
| 0:17:50 | the where separation so we can not |
|---|
| 0:17:53 | we can listen that uh |
|---|
| 0:17:55 | examples so and it doesn't work |
|---|
| 0:17:58 | that's |
|---|
| 0:17:59 | hasn't work |
|---|
| 0:18:02 | sources |
|---|
| 0:18:16 | i |
|---|
| 0:18:37 | i |
|---|
| 0:18:51 | oh |
|---|
| 0:18:56 | okay so let's six this this is first meets |
|---|
| 0:19:11 | i |
|---|
| 0:19:12 | um |
|---|
| 0:19:13 | uh |
|---|
| 0:19:22 | a |
|---|
| 0:19:24 | i |
|---|
| 0:19:25 | hmmm |
|---|
| 0:19:33 | this is this to don't |
|---|
| 0:19:37 | the |
|---|
| 0:19:37 | source |
|---|
| 0:19:38 | skip directly to the results |
|---|
| 0:19:40 | guess |
|---|
| 0:19:41 | but with source and that's C of would we have an estimate of uh an on is |
|---|
| 0:19:45 | for |
|---|
| 0:19:46 | oh |
|---|
| 0:19:54 | or |
|---|
| 0:19:56 | a |
|---|
| 0:19:57 | oh |
|---|
| 0:19:57 | oh |
|---|
| 0:20:05 | a |
|---|
| 0:20:09 | yeah |
|---|
| 0:20:10 | i |
|---|
| 0:20:12 | do you so we have a ten seconds that |
|---|
| 0:20:14 | a for the computer and so we have a proposed the simple sparsity prior |
|---|
| 0:20:18 | to do a a group uh grouping of the sources and solve the permutation problem in this uh |
|---|
| 0:20:24 | a single channel source separation case |
|---|
| 0:20:26 | uh and we show that the algorithm them but there was of the grouping with the as a a post |
|---|
| 0:20:32 | processing step |
|---|
| 0:20:33 | and if you in future work we will try to incorporate |
|---|
| 0:20:35 | smoothest might prior to uh |
|---|
| 0:20:38 | to understand the time the paul then a mix of H |
|---|
| 0:20:42 | i should |
|---|
| 0:20:50 | we have time for only one with question |
|---|
| 0:21:02 | no |
|---|
| 0:21:03 | so the the most you play |
|---|
| 0:21:05 | they are mostly |
|---|
| 0:21:07 | a love part E |
|---|
| 0:21:08 | component |
|---|
| 0:21:09 | and how mix a very much of like it because they playing |
|---|
| 0:21:12 | according to the egg are so that a single |
|---|
| 0:21:15 | close like it |
|---|
| 0:21:16 | and i am wondering how much the |
|---|
| 0:21:19 | sampling rate and the effect these signs in is you R |
|---|
| 0:21:23 | i mean if you would to |
|---|
| 0:21:25 | hi are and fifty resolution oh |
|---|
| 0:21:27 | what it be different |
|---|
| 0:21:28 | what do you yeah |
|---|
| 0:21:31 | that are separation and that you can see there |
|---|
| 0:21:34 | such a |
|---|
| 0:21:35 | yeah |
|---|
| 0:21:35 | oh O K so we don't talk about this in the article |
|---|
| 0:21:39 | now from uh from my parents this is the |
|---|
| 0:21:42 | does not is sensitive to the simply me that you choose |
|---|
| 0:21:46 | uh |
|---|
| 0:21:48 | for this experiment i chose a sampling rate of uh a twenty two Q do have |
|---|
| 0:21:51 | because uh it just just because of a computing time uh |
|---|
| 0:21:56 | concern |
|---|
| 0:21:57 | and |
|---|
| 0:22:00 | i guess uh for the example or well you have a time voice |
|---|
| 0:22:05 | since uh they play in the |
|---|
| 0:22:07 | approximately in the in the mid and high level of of the spectrogram |
|---|
| 0:22:11 | uh this wouldn't uh uh have too much effect because uh |
|---|
| 0:22:15 | because the |
|---|
| 0:22:16 | so the this i range the frequencies are pretty well separated |
|---|
| 0:22:20 | but uh uh if you have base |
|---|
| 0:22:23 | and another source |
|---|
| 0:22:24 | that purely having a good the resolution uh we will help for uh since we have no model |
|---|
| 0:22:29 | uh since we and number than for the basis of and |
|---|
| 0:22:32 | then a having a would sampling right to i mean have a good something rate will help because you then |
|---|
| 0:22:36 | you can get but the resolution uh |
|---|
| 0:22:38 | you can you can afford |
|---|
| 0:22:39 | a longer time window |
|---|
| 0:22:41 | and |
|---|
| 0:22:41 | but the resolution in the frequency range which is particularly important |
|---|
| 0:22:45 | in the low frequency one |
|---|
| 0:22:48 | a some that i i would say that the results are very robust |
|---|
| 0:22:52 | the problem goes from |
|---|
| 0:22:54 | okay thank you |
|---|