| 0:00:13 | alright right | 
|---|
| 0:00:14 | welcome everybody thanks for coming to my talk uh i'd as already mentioned that's doesn't order performance analysis for distortion | 
|---|
| 0:00:19 | structure chorus | 
|---|
| 0:00:20 | and it's one of me and and uh for some and her from him and a you know | 
|---|
| 0:00:24 | the call | 
|---|
| 0:00:25 | now for brief more i probably don't have to tell you that high resolution is interesting for a number of | 
|---|
| 0:00:30 | applications | 
|---|
| 0:00:31 | uh in particular is pretty type parameter estimation schemes are often used | 
|---|
| 0:00:35 | because the very simple the very flexible they can be a that the very settings setting closed form you don't | 
|---|
| 0:00:39 | need any peak search or something | 
|---|
| 0:00:41 | and they were perform still close to the cramerrao about | 
|---|
| 0:00:44 | um | 
|---|
| 0:00:45 | and the were based this all this pretty have all "'em" the based on the should than write equation which | 
|---|
| 0:00:49 | are a set of were determined equations which are typically solved using just score | 
|---|
| 0:00:53 | speakers it's a | 
|---|
| 0:00:55 | oh have release is actually suboptimal | 
|---|
| 0:00:57 | it is suboptimal because it you know was more is the fact that there are also estimation errors in the | 
|---|
| 0:01:02 | subspace itself | 
|---|
| 0:01:03 | there is a better technique which is called structure these scores which takes this estimation errors into account and also | 
|---|
| 0:01:08 | explicitly exploits the structure of shift write equations | 
|---|
| 0:01:11 | just been proposed in ninety seven | 
|---|
| 0:01:14 | um so it comes of improve the form | 
|---|
| 0:01:16 | a however the problem is that so far to be scored this is pretty was only evaluated uh a using | 
|---|
| 0:01:21 | the simulations so there is no analytical statement for when does it to for better how much so an analytical | 
|---|
| 0:01:27 | performance evaluation is of course pretty is the a desirable | 
|---|
| 0:01:30 | um the goal here | 
|---|
| 0:01:32 | this paper was to apply to | 
|---|
| 0:01:33 | seem frame we used | 
|---|
| 0:01:35 | if only had to analyse a these scores based pretty | 
|---|
| 0:01:38 | which is this paper by a by car gives the | 
|---|
| 0:01:40 | a a or expansion of of the signal subspace | 
|---|
| 0:01:44 | and i'm in order to analyse a structure discourse basis pretty | 
|---|
| 0:01:47 | we use the same frame of before to analyse these based pretty the corresponding graph | 
|---|
| 0:01:51 | solicit down here | 
|---|
| 0:01:53 | a so we analyzed various versions like standard is pretty unitary esprit read or is pretty in and more you | 
|---|
| 0:01:57 | can even do would for non circular esprit in and others | 
|---|
| 0:02:00 | but no those are based on these scores | 
|---|
| 0:02:02 | oh the purpose you was to extend the no to incorporate structured | 
|---|
| 0:02:06 | uh which brings me to outline of their the the big one iteration we had now i'll go to would | 
|---|
| 0:02:11 | be for usual on you again the shift invariance equations | 
|---|
| 0:02:13 | uh and and what is to actually scores mean showing you the concept of the first-order perturbation for the S | 
|---|
| 0:02:18 | P which you might find interesting to use also in other fields | 
|---|
| 0:02:21 | and and then our earlier phones results for these process pretty in and the main part of focus on structure | 
|---|
| 0:02:26 | these scores of to the solution and then you know as | 
|---|
| 0:02:28 | and in some simulation results | 
|---|
| 0:02:30 | so let's start of the review | 
|---|
| 0:02:32 | a what is what is shift invariance is pretty based on the fact that you can divide the rate into | 
|---|
| 0:02:36 | two identical sub arrays to one and J two which of their of the same observations except for phase shift | 
|---|
| 0:02:43 | but which is encoded in this musical all spatial frequency | 
|---|
| 0:02:46 | and a spatial frequencies link to your parameters of interest for instance if you want to do direction of arrival | 
|---|
| 0:02:51 | name to the direction of arrival and T done a simple relation | 
|---|
| 0:02:54 | now these matrices you you you can use selection make J one and J two which operate on your race | 
|---|
| 0:02:59 | during vector a to select the of the first and the second sub array | 
|---|
| 0:03:03 | and shift them very close to you get the same observation except for a face should which contains the problem | 
|---|
| 0:03:07 | of interest | 
|---|
| 0:03:08 | does is for a single source for multiple source | 
|---|
| 0:03:11 | these J one and J two they operate on the race hearing matrix a which contains all the raised and | 
|---|
| 0:03:15 | vectors | 
|---|
| 0:03:16 | and you parameters is of interest are actually five | 
|---|
| 0:03:18 | right right these contains spatial free | 
|---|
| 0:03:21 | no this is of the right equation a matrix equation the problem is the array steering matrix of course and | 
|---|
| 0:03:25 | no | 
|---|
| 0:03:26 | so to get rid of the unknown array steering matrix | 
|---|
| 0:03:29 | what we do is we take our observations let's say X is the is a is a matrix which contains | 
|---|
| 0:03:33 | and sit C but observations of or and and sensors | 
|---|
| 0:03:36 | and we just computed as P D in from the svd P get an estimate for the signal subspace which | 
|---|
| 0:03:41 | are the D don't the left single vector | 
|---|
| 0:03:43 | and then you was the fact that the race during matrix and this made use S and the same column | 
|---|
| 0:03:48 | space approximately because there is noise | 
|---|
| 0:03:50 | so they are related via a transfer matrix T | 
|---|
| 0:03:52 | and this can be used to eliminate right some we are actually have a shift invariance equations which need to | 
|---|
| 0:03:57 | be sold pop i that's the only known | 
|---|
| 0:04:00 | you has we have an estimate for | 
|---|
| 0:04:01 | and i the eigen base so i are are are do once correct range | 
|---|
| 0:04:05 | right this is the basis for is pretty just a very quick review | 
|---|
| 0:04:08 | the main point here is that these of the shift write equations we need to solve the are determined | 
|---|
| 0:04:12 | and we have an estimate for the subspace but it's not accurate | 
|---|
| 0:04:15 | right | 
|---|
| 0:04:15 | so how do we sell | 
|---|
| 0:04:17 | typically people sold just using these were | 
|---|
| 0:04:19 | and these score just mean to minimize the least squares fit between the left and the right hand side of | 
|---|
| 0:04:23 | the equation right only subject to side | 
|---|
| 0:04:25 | uh this gives a very close from solution a very simple one you can use the inverse | 
|---|
| 0:04:29 | but the problem is you we more the fact that you don't know exactly us that is you actually implicitly | 
|---|
| 0:04:34 | that it's perfectly known which is not true we know that there is in there | 
|---|
| 0:04:38 | and that's the idea of structure these chorus | 
|---|
| 0:04:40 | such each score as change this cost function what is to change here | 
|---|
| 0:04:43 | a change to incorporate for each of these occurrences of the subspace bayes us an additional they'll tell us which | 
|---|
| 0:04:48 | explicitly models the fact that we have an estimation error in the subspace that we tried to correct so that | 
|---|
| 0:04:54 | the two sides of the question a line in a better way | 
|---|
| 0:04:56 | and in an in a group regularization term is to make sure that this up would doing a small to | 
|---|
| 0:05:01 | penalise to lot of | 
|---|
| 0:05:03 | this is the cost function for structure these scores | 
|---|
| 0:05:05 | the nice thing is that takes into account errors and subspace | 
|---|
| 0:05:08 | but draw a always done now not only in your course anymore but it's quadratic | 
|---|
| 0:05:12 | so we typically so it iteratively via a local linearization | 
|---|
| 0:05:15 | but it has been shown that only very few iterations are required actually in the high snr regime only one | 
|---|
| 0:05:20 | iteration is required so you can see just as a as a correct | 
|---|
| 0:05:24 | alright right | 
|---|
| 0:05:24 | now to to to come to the performance analysis for this we need to look into this | 
|---|
| 0:05:28 | source of error that actually here the source of average is here is this yeah to yes this error and | 
|---|
| 0:05:33 | the signal subspace we need to analytically grass | 
|---|
| 0:05:36 | and the the frame up we using for this is the one that by but kind of which i just | 
|---|
| 0:05:40 | briefly you want to review was also very simple | 
|---|
| 0:05:42 | you take a look at you and vector or uh observations X not without any noise where you have you | 
|---|
| 0:05:47 | true signal subspace and a to noise subspace if you break again the S | 
|---|
| 0:05:50 | in the presence of noise you only have an estimate so you can say that your estimate signal subspace is | 
|---|
| 0:05:55 | the true one class and error come down | 
|---|
| 0:05:57 | in this trying to find | 
|---|
| 0:05:58 | and for this | 
|---|
| 0:05:59 | for this error and you can always expand into one part which lies in the model bass and one part | 
|---|
| 0:06:04 | which lies in the signal subspace this just because it leads and this N dimensional space so you can always | 
|---|
| 0:06:08 | break it into but two space | 
|---|
| 0:06:10 | and the interpretation of the first component | 
|---|
| 0:06:13 | it's a a of the signal subspace which is in the noise subspace would really model models how much of | 
|---|
| 0:06:17 | the noise leads into the thing as it's how the subspace itself | 
|---|
| 0:06:20 | a raw whereas the second one is our of the signal subspace inside the signal | 
|---|
| 0:06:25 | this one models how the individual singular vectors inside the signal subspace the particular basis we choose how was the | 
|---|
| 0:06:30 | trip right | 
|---|
| 0:06:31 | so obviously the second one plays no role for esprit because the particular basis because the relevant only the first | 
|---|
| 0:06:37 | one | 
|---|
| 0:06:38 | but extensions exist | 
|---|
| 0:06:40 | we only use the first term because the second one's a relevant if you want you can easily incorporate the | 
|---|
| 0:06:44 | second one as well | 
|---|
| 0:06:45 | has been proposed by a car and a you can see it's a very similar expressions of first order expansion | 
|---|
| 0:06:50 | in the noise and which also perturbation | 
|---|
| 0:06:52 | and the second one it's been proposed actually minor college | 
|---|
| 0:06:55 | um but as i said we don't need for this work | 
|---|
| 0:06:58 | um also the colour already ninety three has used this result to analyse standard esprit in is given a first | 
|---|
| 0:07:03 | order expansion for the estimation error indicates spatial frequency using standard esprit which is the simple expression | 
|---|
| 0:07:09 | based on this work with expanded it | 
|---|
| 0:07:11 | um we've be shown that a last year at i cast in in dallas that you can instance perform statistical | 
|---|
| 0:07:16 | expectation of this | 
|---|
| 0:07:17 | assuming white complex gaussian noise | 
|---|
| 0:07:19 | so what it should probably emphasise that this framework | 
|---|
| 0:07:22 | its asymptotic | 
|---|
| 0:07:23 | in the effective snr so you don't need a large number of snapshots of something | 
|---|
| 0:07:28 | uh you you can have a single snapshot if you want as long as the variance is low and it | 
|---|
| 0:07:31 | needs no a particular assumptions about the statistics you don't need gaussian of the noise you don't even need a | 
|---|
| 0:07:37 | gaussian of the symbols you just need the perturbation to be small | 
|---|
| 0:07:40 | but if you as young gas and then you can of course conforms forms that exceed the expectation and you | 
|---|
| 0:07:43 | get needs better | 
|---|
| 0:07:45 | this is lisa and of this read this is nice unitary we and you can do more | 
|---|
| 0:07:48 | these a previous results we shown and based on these now we try to expand them to incorporate structure | 
|---|
| 0:07:54 | so | 
|---|
| 0:07:55 | um what is done here is we first check our extension attention to a special piece of stuff to these | 
|---|
| 0:07:59 | scores which is using a single iteration | 
|---|
| 0:08:01 | and it is not using any regularization | 
|---|
| 0:08:03 | the reason for this is that these assumptions are asymptotically optimal for structured these in the high snr regime | 
|---|
| 0:08:09 | and since the performance and as we do use anyways asymptotic high as are it's it's fine to assume this | 
|---|
| 0:08:14 | right "'cause" it | 
|---|
| 0:08:15 | but it but it asymptotically in you're actually | 
|---|
| 0:08:17 | right | 
|---|
| 0:08:18 | uh i'm of these assumptions you can express the cost function and the solution and in a very simple way | 
|---|
| 0:08:22 | you can say that these solutions sites for structure these is equal to the initial solution get "'em" by this | 
|---|
| 0:08:27 | first plus an update or | 
|---|
| 0:08:29 | and is a a term is the solution of this cost function | 
|---|
| 0:08:32 | which is of course quadratic "'cause" i've said | 
|---|
| 0:08:34 | there is there is a a a a a term image does not depend on a there is a link | 
|---|
| 0:08:37 | at time and then there is a quadratic term right | 
|---|
| 0:08:39 | the quadratic term be black that's the linearization so we are back to a lean this squares problem | 
|---|
| 0:08:44 | and this mean of this problem has of course of very simple solution | 
|---|
| 0:08:47 | for for us uh to be suppressed so this will be the update for me to in this would be | 
|---|
| 0:08:51 | the update for the for the subspace if if you would wanna do the second iteration be actually don't need | 
|---|
| 0:08:56 | the second | 
|---|
| 0:08:56 | since we only going one | 
|---|
| 0:08:58 | so that the main message here is that it actually the up they can be explicitly computed as taking this | 
|---|
| 0:09:02 | vector are are last | 
|---|
| 0:09:04 | which is the vectorized version of the residual matrix after doing least | 
|---|
| 0:09:08 | a be multiplied by that to them as of this matrix F | 
|---|
| 0:09:11 | which is the linear mapping me | 
|---|
| 0:09:12 | and for this we have to find of a first-order expansion | 
|---|
| 0:09:15 | how have we done it we done it by looking at both terms individually | 
|---|
| 0:09:19 | we start with a | 
|---|
| 0:09:19 | first that's of most of that matrix F | 
|---|
| 0:09:22 | what we will the set you can express this matrix a had as equal to and matrix a matrix they'll | 
|---|
| 0:09:28 | have | 
|---|
| 0:09:29 | but the matrix F is constant in the sense that that does not depend on the perturbation itself self right | 
|---|
| 0:09:34 | so if you look at one realisation look at the random of some of perturbation | 
|---|
| 0:09:37 | this some will be constant and this one will be linear right and this gives the matrix F | 
|---|
| 0:09:42 | and therefore if we look at it so inverse since this part is zero mean it's not very hard to | 
|---|
| 0:09:46 | see a that in versus actually equal to the sort inverse of this constant matrix independent of the perturbation | 
|---|
| 0:09:51 | loss a linear term that's a quadratic term plus i wouldn't for | 
|---|
| 0:09:54 | right we don't actually physically need to spend it it's fine to know that | 
|---|
| 0:09:58 | this constant term as we will see the in terms actually not need and this is simplified and is greatly | 
|---|
| 0:10:02 | and for the second term from those are last | 
|---|
| 0:10:05 | uh it's not difficult to see that this can be written as a a a a at the linux pension | 
|---|
| 0:10:09 | for the again | 
|---|
| 0:10:10 | uh the error in the subspace they are U S be multiplied by one matrix mapping the subspace you to | 
|---|
| 0:10:15 | the residual are possible direct term which be more | 
|---|
| 0:10:18 | and uh also this | 
|---|
| 0:10:20 | a a the subspace is the result of a car | 
|---|
| 0:10:22 | has a mean expansion in terms of the actual perturbation them noise | 
|---|
| 0:10:25 | again but forming that a matrix plus a body matrix plus i don't for | 
|---|
| 0:10:30 | now the collect both | 
|---|
| 0:10:31 | if we collect um this and this together we see that the uh a vector all has a lean expansion | 
|---|
| 0:10:36 | and the noise | 
|---|
| 0:10:37 | and now we combine these two results back into the original expression | 
|---|
| 0:10:41 | we can see that if you multiply the um and this one out them be get a linear term which | 
|---|
| 0:10:46 | is this constant term times the leading it | 
|---|
| 0:10:48 | plus a quadratic term just as linear times the the linear to right so the quadratic term again be like | 
|---|
| 0:10:54 | that's | 
|---|
| 0:10:54 | that's first order | 
|---|
| 0:10:55 | so this shows that we don't actually need even the lean expansion of them soon | 
|---|
| 0:10:59 | right | 
|---|
| 0:10:59 | and then we get this very simple result | 
|---|
| 0:11:01 | a which is pretty into it if you start of the noise you have one a mapping matrix from the | 
|---|
| 0:11:05 | noise to the subspace that from the stuff to the residuals and then this matrix here | 
|---|
| 0:11:09 | this is to the vector and selling and us | 
|---|
| 0:11:11 | but as is that are only interested in this up part | 
|---|
| 0:11:13 | so the final result for this upper part is actually this one again | 
|---|
| 0:11:17 | pretty intuitive meaning about mapping | 
|---|
| 0:11:19 | and uh then you can also like in of back to the original expansion of the estimation error you find | 
|---|
| 0:11:23 | a first or expansion of the estimation error of spatial frequencies again | 
|---|
| 0:11:27 | it's a very simple and it has a form just again in its structure very simple to but these press | 
|---|
| 0:11:32 | expansion be shown previously | 
|---|
| 0:11:34 | are we have a me this vector are L as now it's R S L a slightly different but the | 
|---|
| 0:11:37 | form still the same | 
|---|
| 0:11:39 | and again you can if you want a for means mean | 
|---|
| 0:11:42 | um | 
|---|
| 0:11:43 | the the mean square error if you assume as your mean and circle is much white noise you actually don't | 
|---|
| 0:11:47 | you gals in for this | 
|---|
| 0:11:48 | and you get a needs got are again very compact and and simplex spray | 
|---|
| 0:11:51 | now but is all this good for why do we do a why do we go through this analytical result | 
|---|
| 0:11:55 | and what does it show us now that we have the result | 
|---|
| 0:11:57 | the break | 
|---|
| 0:11:58 | think this is good for is if you look at | 
|---|
| 0:12:00 | that's | 
|---|
| 0:12:00 | specific case we can simplify the expression so much that you actually gain inside | 
|---|
| 0:12:05 | in to what is the performance of the schemes of the very specific set | 
|---|
| 0:12:09 | and to to add the this point i brought one example you which is not the paper due to like | 
|---|
| 0:12:13 | of space but i still i think it's interesting | 
|---|
| 0:12:15 | to kind of push for what of it that this we has has applications | 
|---|
| 0:12:18 | and the example is of course the simple one you can think of which is a single source | 
|---|
| 0:12:22 | if you consider a single sauce | 
|---|
| 0:12:24 | and be shown the uh that fully score space is pretty the means error as a break | 
|---|
| 0:12:28 | compact expression if you consider a uniformly a rate of N sensors | 
|---|
| 0:12:32 | um it has to be effective as an front and then it's quadratic and one over one over and basically | 
|---|
| 0:12:37 | then the common or bound also as a very simple expression | 
|---|
| 0:12:40 | which means that you can find the asymptotic efficiency again asymptotic effective snr it still it's about for single snapshot | 
|---|
| 0:12:46 | um is is given by this expression so you have a closed form expression for the asymptotic efficiency only depending | 
|---|
| 0:12:51 | on and | 
|---|
| 0:12:52 | in it's exact except for the fact that it doesn't talk | 
|---|
| 0:12:55 | um so and and what you can see you is that it basically it's start of one for two and | 
|---|
| 0:12:59 | three sensors and then it goes down | 
|---|
| 0:13:01 | so these press based pretty is not efficient for large race on a single song | 
|---|
| 0:13:05 | we did the same number of the structure these scores | 
|---|
| 0:13:07 | and after a number of manipulations we found again a closed-form expression for the mean square | 
|---|
| 0:13:12 | it's a bit more involved but you can do the same thing you can not of like from are bound | 
|---|
| 0:13:15 | i mean square error and you find a close form expression for the asymptotic efficiency which is | 
|---|
| 0:13:20 | a ratio of order only us | 
|---|
| 0:13:22 | interestingly the first three coefficients agree then they start to diff | 
|---|
| 0:13:25 | if you plot is on the same thing as a things you and one it looks like it's almost equal | 
|---|
| 0:13:29 | to one but is actually not if you zoom in a little bit you can see it starts of one | 
|---|
| 0:13:33 | it goes down a little bit and then it goes back out | 
|---|
| 0:13:35 | we don't really have physical explanation for that mathematically you could prove that it is that with simulations you can | 
|---|
| 0:13:40 | verify that like | 
|---|
| 0:13:42 | and you really you have the values for the asymptotic efficiency as as exact number so this is a pretty | 
|---|
| 0:13:46 | value result | 
|---|
| 0:13:47 | it would be interesting to extend this to two sources to see was the performance in terms of separation correlation | 
|---|
| 0:13:52 | and these real or | 
|---|
| 0:13:53 | uh a parameter stuff | 
|---|
| 0:13:55 | alright right | 
|---|
| 0:13:55 | a just a few must simulations and the simulations will be compared is is we compare the empirical error | 
|---|
| 0:14:00 | which you get by actually performing as free on random data and then computing the estimate the spatial frequencies computing | 
|---|
| 0:14:06 | the arrow an averaging it | 
|---|
| 0:14:08 | with the semi analytical results which still depend on the noise realisations of the average these or noise | 
|---|
| 0:14:13 | and a fully analytical results which which for which no simulations actually needed on the one to colour everything is | 
|---|
| 0:14:19 | needed because there are around of that | 
|---|
| 0:14:20 | right | 
|---|
| 0:14:21 | uh a this first example he for uncorrelated sources | 
|---|
| 0:14:24 | you know the the performance of unitary briefly specified | 
|---|
| 0:14:28 | sort | 
|---|
| 0:14:28 | these grass actually scores | 
|---|
| 0:14:30 | is very close so it's of course interesting to see can this really small gap still be reliably predicted | 
|---|
| 0:14:35 | well the S is yes | 
|---|
| 0:14:37 | with our analytical results we become asymptotically optimal yeah the same small yeah | 
|---|
| 0:14:41 | the semi analytical this | 
|---|
| 0:14:42 | the volume | 
|---|
| 0:14:43 | uh another the result this is pretty source which are very strongly correlated there zero point ninety nine core between | 
|---|
| 0:14:48 | any pair of soft | 
|---|
| 0:14:49 | i know we have four as we have to us and standard this spree based on these press actually scores | 
|---|
| 0:14:54 | unitary esprit based on these cost of is discourse | 
|---|
| 0:14:56 | to do this time correlation or is a big gas so you can more clearly distinguished the curve | 
|---|
| 0:15:01 | again this is the i mean uh the results to become | 
|---|
| 0:15:03 | accurate for high snr this | 
|---|
| 0:15:05 | this | 
|---|
| 0:15:05 | the form you | 
|---|
| 0:15:06 | and then use the single source now for single source we have an improvement if you plot of versus the | 
|---|
| 0:15:10 | snr | 
|---|
| 0:15:11 | for for eight sensors between these grass structure course | 
|---|
| 0:15:14 | and again the analytical results | 
|---|
| 0:15:17 | but range to the conclusions | 
|---|
| 0:15:18 | what we present a is a first order | 
|---|
| 0:15:20 | a a perturbation analysis to actually score is pretty | 
|---|
| 0:15:23 | just based on the performance analysis for the S D which is a very nice concept that can be used | 
|---|
| 0:15:27 | also in different field | 
|---|
| 0:15:29 | the nice thing about it its asymptotic and the effective snr so i a small noise variance | 
|---|
| 0:15:33 | well a large and | 
|---|
| 0:15:34 | can be both for whatever you want | 
|---|
| 0:15:36 | and it explicit which means you don't need any assumptions about the statistics you need to noise to be zero | 
|---|
| 0:15:41 | mean but you don't need to source to be gaussian you the noise to be C | 
|---|
| 0:15:44 | you just need to be small | 
|---|
| 0:15:46 | and is also shown means but our assuming zero-mean circular is met guide noise | 
|---|
| 0:15:50 | and now also shown the explicit expressions for single source where you can actually see just talking | 
|---|
| 0:15:55 | this concludes my talk things | 
|---|
| 0:16:03 | we have time for question | 
|---|
| 0:16:14 | yeah | 
|---|
| 0:16:15 | there is a relation but there is also a difference | 
|---|
| 0:16:18 | um | 
|---|
| 0:16:19 | okay here | 
|---|
| 0:16:20 | in an in totally scores you allow for an error | 
|---|
| 0:16:23 | in in in this expression in you mean mapping matrices that's say | 
|---|
| 0:16:27 | but you assume that there as are independent errors on the left and the right hand side of spray | 
|---|
| 0:16:32 | at that's why this is called structure these course totally suppressed would model to independent errors for the left and | 
|---|
| 0:16:37 | the right and say the different | 
|---|
| 0:16:38 | i actually there is a structure in the shift writes equations which tells you that these are actually not and | 
|---|
| 0:16:42 | pain and are almost the same except little selection matrices | 
|---|
| 0:16:45 | and the structure should be incorporated the solution | 
|---|
| 0:16:51 | yeah | 
|---|
| 0:16:52 | yeah | 
|---|
| 0:16:54 | right | 
|---|
| 0:16:55 | this only is yeah | 
|---|
| 0:16:56 | yeah | 
|---|
| 0:17:16 | it doesn't have to be we don't can find that that that you explicitly to be unit | 
|---|
| 0:17:20 | it could be not | 
|---|
| 0:17:25 | well | 
|---|
| 0:17:25 | you for for is pretty you don't need a constraint that does unitary mean | 
|---|
| 0:17:29 | this you can you to any subspace at you one doesn't have to be you know | 
|---|
| 0:17:39 | but you can you can describe a sub using any is any base equivalent typically you the S D because | 
|---|
| 0:17:44 | it gives you know from all the basis and it's nice to work with and it's simple | 
|---|
| 0:17:48 | but you could use any other basis and it would have no inter | 
|---|
| 0:17:51 | form | 
|---|
| 0:17:51 | any base | 
|---|
| 0:17:52 | fine | 
|---|
| 0:17:53 | actually me started rolling taxes pretty the first version | 
|---|
| 0:17:56 | i'll be defined of subspace it was long units are you but then when be corrected it we met every | 
|---|
| 0:18:00 | had no impact on the performance which is | 
|---|
| 0:18:02 | just would be | 
|---|
| 0:18:03 | yeah but just | 
|---|
| 0:18:06 | us us this to be unitary domain | 
|---|
| 0:18:10 | if you had last yeah i | 
|---|
| 0:18:12 | not so mean unit | 
|---|
| 0:18:13 | mean | 
|---|
| 0:18:14 | so can you | 
|---|
| 0:18:15 | the | 
|---|
| 0:18:16 | i you know a | 
|---|
| 0:18:17 | to to issue | 
|---|
| 0:18:18 | you would like to minimize | 
|---|
| 0:18:20 | but which | 
|---|
| 0:18:21 | know | 
|---|
| 0:18:22 | you would still operate | 
|---|
| 0:18:23 | you can men | 
|---|
| 0:18:24 | C | 
|---|
| 0:18:25 | yeah | 
|---|
| 0:18:27 | you could do would but i one | 
|---|
| 0:18:28 | you you what what would be the uh the data | 
|---|
| 0:18:30 | i mean we don't need to same unitary T four | 
|---|
| 0:18:34 | the writing is pretty for using it | 
|---|
| 0:18:37 | but it would be possible | 
|---|
| 0:18:40 | i you you have question | 
|---|
| 0:18:43 | a some thought of a set of goes to infinity so a set of the not so sure | 
|---|
| 0:18:50 | i | 
|---|
| 0:18:50 | i | 
|---|
| 0:18:51 | so what we actually need | 
|---|
| 0:18:52 | let me go to | 
|---|
| 0:18:55 | well | 
|---|
| 0:18:57 | what we need is that this term and frame here that it they can zero we and P D which | 
|---|
| 0:19:01 | is the power of sauce of single and which is the power of your | 
|---|
| 0:19:04 | um and the the noise variance and you and which is the number of snapshots | 
|---|
| 0:19:08 | so if you have a finite snr and you that in going to infinity works just exactly the same way | 
|---|
| 0:19:13 | i i you let and go to infinity or you get the noise variance go zero all | 
|---|
| 0:19:17 | a source power go to | 
|---|
| 0:19:18 | in | 
|---|
| 0:19:27 | yes it single source only for singles | 
|---|
| 0:19:30 | it's not as bad for for multiple | 
|---|
| 0:19:34 | it was a prize one as what for the first time but i can very fight using selection | 
|---|
| 0:19:37 | so | 
|---|
| 0:19:41 | it is surprising because he the low resolution techniques are asymptotically optimal for single source but it's is not | 
|---|
| 0:19:53 | when when that more sources is it it's not as that it's it's very hard to find as expressions explicitly | 
|---|
| 0:19:57 | for most source because the number of terms get | 
|---|
| 0:19:59 | it's very large | 
|---|
| 0:20:00 | very to use we try simplify simplified | 
|---|
| 0:20:02 | we we we actually didn't | 
|---|
| 0:20:03 | get the final result but from from simulations my experience is that | 
|---|
| 0:20:06 | this kind of the the risk case in terms of comparing it with rubber | 
|---|
| 0:20:12 | it also disappear of course if you replace quest squares fast a score | 
|---|
| 0:20:15 | which may serve of the spec | 
|---|
| 0:20:18 | and you just it's just one correction term it's not and it to iterative procedure to apply already of a | 
|---|
| 0:20:23 | single iteration it basically disappear | 
|---|
| 0:20:26 | in it's it's something simple it's holding one set of | 
|---|
| 0:20:29 | when | 
|---|
| 0:20:34 | a | 
|---|
| 0:20:34 | oh yeah | 
|---|
| 0:20:35 | the question | 
|---|
| 0:20:36 | it seems that uh stuff to the squares | 
|---|
| 0:20:39 | but and and these squares | 
|---|
| 0:20:40 | especially when the source is a card | 
|---|
| 0:20:43 | right | 
|---|
| 0:20:44 | so okay translate translated | 
|---|
| 0:20:46 | a | 
|---|
| 0:20:47 | so | 
|---|
| 0:20:48 | estimation use | 
|---|
| 0:20:49 | uh | 
|---|
| 0:20:50 | is words | 
|---|
| 0:20:51 | when the sources are quite | 
|---|
| 0:20:53 | is better one one | 
|---|
| 0:20:55 | because this is main | 
|---|
| 0:20:57 | to my | 
|---|
| 0:20:58 | a problem is that | 
|---|
| 0:21:00 | there is not always a one-to-one mapping between | 
|---|
| 0:21:02 | a better self space and a better performance of this pretty | 
|---|
| 0:21:05 | but but start of a senses as for single source we got a better subspace but the the the the | 
|---|
| 0:21:09 | means but our of the speech that's both exact | 
|---|
| 0:21:12 | sometimes times you get you get a better subspace but it does not help Q in terms of your you | 
|---|
| 0:21:16 | mean square | 
|---|
| 0:21:17 | i would say that there should probably be be something but i don't know if it's the weak link | 
|---|
| 0:21:21 | a strong so this something | 
|---|
| 0:21:22 | still have | 
|---|
| 0:21:26 | which | 
|---|
| 0:21:27 | so let's | 
|---|