Are Linear Maps resistant to Noise?
$begingroup$
Let's assume I have a $m times m$ matrix $M$ with Frobenius norm $1$ and a unit vector $x in S^{m-1}$. I also have a second $m times m$ matrix $M^*$ which is obtained from the first one plus some injected noise $eta$, where every entry $eta_{i,j}$ comes from a Gaussian distribution with mean $0$, such that $||eta||_F = 1/10$.
What can we say about the following expected value $$mathbb{E}[| Mx - M^*x|]?$$
EDIT: How it has been suggested in an answer, $mathbb{E}|eta(x)| <1/10$ since $|eta(x)| < 1/10$ because of the Frobenius norm of the noise, so the equality only holds when vector $x$
is aligned with the highest eigenvector of the noise matrix $eta$
, but this happens with low probability, and this probability should (intuitively) change with respect to $m$
: in a higher dimensional space the probability of two random vectors having a similar direction is lower than in a 2-dimensional one.
Is there a way to improve the bound on this expected value?
linear-algebra probability matrices linear-transformations
$endgroup$
|
show 3 more comments
$begingroup$
Let's assume I have a $m times m$ matrix $M$ with Frobenius norm $1$ and a unit vector $x in S^{m-1}$. I also have a second $m times m$ matrix $M^*$ which is obtained from the first one plus some injected noise $eta$, where every entry $eta_{i,j}$ comes from a Gaussian distribution with mean $0$, such that $||eta||_F = 1/10$.
What can we say about the following expected value $$mathbb{E}[| Mx - M^*x|]?$$
EDIT: How it has been suggested in an answer, $mathbb{E}|eta(x)| <1/10$ since $|eta(x)| < 1/10$ because of the Frobenius norm of the noise, so the equality only holds when vector $x$
is aligned with the highest eigenvector of the noise matrix $eta$
, but this happens with low probability, and this probability should (intuitively) change with respect to $m$
: in a higher dimensional space the probability of two random vectors having a similar direction is lower than in a 2-dimensional one.
Is there a way to improve the bound on this expected value?
linear-algebra probability matrices linear-transformations
$endgroup$
$begingroup$
What exactly is $|x|$? Is it the Euclidean norm?
$endgroup$
– Omnomnomnom
Jan 13 at 21:36
$begingroup$
And what is $|eta|$ in this context?
$endgroup$
– Omnomnomnom
Jan 13 at 21:39
$begingroup$
Sorry, $||eta||$ is the Frobenius norm, while $||x||$ is the euclidean norm. I'll edit. Thanks for the correction
$endgroup$
– Alfred
Jan 13 at 22:55
1
$begingroup$
You’re computing $Bbb E |eta x|$ for what that’s worth
$endgroup$
– Omnomnomnom
Jan 13 at 23:57
1
$begingroup$
We could say $$Bbb E|eta(x)| leq Bbb E |eta|_2 leq Bbb E | eta|_F = 1/10$$ perhaps that helps
$endgroup$
– Omnomnomnom
Jan 14 at 0:00
|
show 3 more comments
$begingroup$
Let's assume I have a $m times m$ matrix $M$ with Frobenius norm $1$ and a unit vector $x in S^{m-1}$. I also have a second $m times m$ matrix $M^*$ which is obtained from the first one plus some injected noise $eta$, where every entry $eta_{i,j}$ comes from a Gaussian distribution with mean $0$, such that $||eta||_F = 1/10$.
What can we say about the following expected value $$mathbb{E}[| Mx - M^*x|]?$$
EDIT: How it has been suggested in an answer, $mathbb{E}|eta(x)| <1/10$ since $|eta(x)| < 1/10$ because of the Frobenius norm of the noise, so the equality only holds when vector $x$
is aligned with the highest eigenvector of the noise matrix $eta$
, but this happens with low probability, and this probability should (intuitively) change with respect to $m$
: in a higher dimensional space the probability of two random vectors having a similar direction is lower than in a 2-dimensional one.
Is there a way to improve the bound on this expected value?
linear-algebra probability matrices linear-transformations
$endgroup$
Let's assume I have a $m times m$ matrix $M$ with Frobenius norm $1$ and a unit vector $x in S^{m-1}$. I also have a second $m times m$ matrix $M^*$ which is obtained from the first one plus some injected noise $eta$, where every entry $eta_{i,j}$ comes from a Gaussian distribution with mean $0$, such that $||eta||_F = 1/10$.
What can we say about the following expected value $$mathbb{E}[| Mx - M^*x|]?$$
EDIT: How it has been suggested in an answer, $mathbb{E}|eta(x)| <1/10$ since $|eta(x)| < 1/10$ because of the Frobenius norm of the noise, so the equality only holds when vector $x$
is aligned with the highest eigenvector of the noise matrix $eta$
, but this happens with low probability, and this probability should (intuitively) change with respect to $m$
: in a higher dimensional space the probability of two random vectors having a similar direction is lower than in a 2-dimensional one.
Is there a way to improve the bound on this expected value?
linear-algebra probability matrices linear-transformations
linear-algebra probability matrices linear-transformations
edited Jan 14 at 15:45
Omnomnomnom
128k790179
128k790179
asked Jan 13 at 20:17
AlfredAlfred
336
336
$begingroup$
What exactly is $|x|$? Is it the Euclidean norm?
$endgroup$
– Omnomnomnom
Jan 13 at 21:36
$begingroup$
And what is $|eta|$ in this context?
$endgroup$
– Omnomnomnom
Jan 13 at 21:39
$begingroup$
Sorry, $||eta||$ is the Frobenius norm, while $||x||$ is the euclidean norm. I'll edit. Thanks for the correction
$endgroup$
– Alfred
Jan 13 at 22:55
1
$begingroup$
You’re computing $Bbb E |eta x|$ for what that’s worth
$endgroup$
– Omnomnomnom
Jan 13 at 23:57
1
$begingroup$
We could say $$Bbb E|eta(x)| leq Bbb E |eta|_2 leq Bbb E | eta|_F = 1/10$$ perhaps that helps
$endgroup$
– Omnomnomnom
Jan 14 at 0:00
|
show 3 more comments
$begingroup$
What exactly is $|x|$? Is it the Euclidean norm?
$endgroup$
– Omnomnomnom
Jan 13 at 21:36
$begingroup$
And what is $|eta|$ in this context?
$endgroup$
– Omnomnomnom
Jan 13 at 21:39
$begingroup$
Sorry, $||eta||$ is the Frobenius norm, while $||x||$ is the euclidean norm. I'll edit. Thanks for the correction
$endgroup$
– Alfred
Jan 13 at 22:55
1
$begingroup$
You’re computing $Bbb E |eta x|$ for what that’s worth
$endgroup$
– Omnomnomnom
Jan 13 at 23:57
1
$begingroup$
We could say $$Bbb E|eta(x)| leq Bbb E |eta|_2 leq Bbb E | eta|_F = 1/10$$ perhaps that helps
$endgroup$
– Omnomnomnom
Jan 14 at 0:00
$begingroup$
What exactly is $|x|$? Is it the Euclidean norm?
$endgroup$
– Omnomnomnom
Jan 13 at 21:36
$begingroup$
What exactly is $|x|$? Is it the Euclidean norm?
$endgroup$
– Omnomnomnom
Jan 13 at 21:36
$begingroup$
And what is $|eta|$ in this context?
$endgroup$
– Omnomnomnom
Jan 13 at 21:39
$begingroup$
And what is $|eta|$ in this context?
$endgroup$
– Omnomnomnom
Jan 13 at 21:39
$begingroup$
Sorry, $||eta||$ is the Frobenius norm, while $||x||$ is the euclidean norm. I'll edit. Thanks for the correction
$endgroup$
– Alfred
Jan 13 at 22:55
$begingroup$
Sorry, $||eta||$ is the Frobenius norm, while $||x||$ is the euclidean norm. I'll edit. Thanks for the correction
$endgroup$
– Alfred
Jan 13 at 22:55
1
1
$begingroup$
You’re computing $Bbb E |eta x|$ for what that’s worth
$endgroup$
– Omnomnomnom
Jan 13 at 23:57
$begingroup$
You’re computing $Bbb E |eta x|$ for what that’s worth
$endgroup$
– Omnomnomnom
Jan 13 at 23:57
1
1
$begingroup$
We could say $$Bbb E|eta(x)| leq Bbb E |eta|_2 leq Bbb E | eta|_F = 1/10$$ perhaps that helps
$endgroup$
– Omnomnomnom
Jan 14 at 0:00
$begingroup$
We could say $$Bbb E|eta(x)| leq Bbb E |eta|_2 leq Bbb E | eta|_F = 1/10$$ perhaps that helps
$endgroup$
– Omnomnomnom
Jan 14 at 0:00
|
show 3 more comments
0
active
oldest
votes
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3072466%2fare-linear-maps-resistant-to-noise%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
0
active
oldest
votes
0
active
oldest
votes
active
oldest
votes
active
oldest
votes
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3072466%2fare-linear-maps-resistant-to-noise%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
What exactly is $|x|$? Is it the Euclidean norm?
$endgroup$
– Omnomnomnom
Jan 13 at 21:36
$begingroup$
And what is $|eta|$ in this context?
$endgroup$
– Omnomnomnom
Jan 13 at 21:39
$begingroup$
Sorry, $||eta||$ is the Frobenius norm, while $||x||$ is the euclidean norm. I'll edit. Thanks for the correction
$endgroup$
– Alfred
Jan 13 at 22:55
1
$begingroup$
You’re computing $Bbb E |eta x|$ for what that’s worth
$endgroup$
– Omnomnomnom
Jan 13 at 23:57
1
$begingroup$
We could say $$Bbb E|eta(x)| leq Bbb E |eta|_2 leq Bbb E | eta|_F = 1/10$$ perhaps that helps
$endgroup$
– Omnomnomnom
Jan 14 at 0:00