Smoothness means that the gradient is Lipschitz continuous
Clash Royale CLAN TAG#URR8PPP
up vote
1
down vote
favorite
Let $f:mathbbR^drightarrow mathbbR$ be a convex and differentiable function. We are saying that $f$ is smooth (with parameter $L$) if $forall mathbfx, mathbfy in mathbbR^d: f(mathbfy) leq f(mathbfx) + nabla f(mathbfx)^T (mathbfy - mathbfx) + fracL2|mathbfx-mathbfy|^2$
I am trying to prove the following Lemma:
Let $f:mathbbR^d rightarrow mathbbR$ be convex and differentiable. The following two statements are equivalent:
(i) f is smooth with parameter $L$
(ii) $|nabla f(mathbfx) - nabla f(mathbfy)| leq L|mathbfx-mathbfy| forall mathbfx, mathbfy in mathbbR^d$
I've managed to do (ii) $rightarrow$ (i) by using the first order characterization of convexity, subtracting and adding the term $nabla f(mathbfx)^T(mathbfy-mathbfx)$ and then by using the Cauchy-Schwarz inequality.
However I am stuck in (i) $rightarrow$ (ii). I am taking the smoothness condition one time for $mathbfx, mathbfy$ and one time for $mathbfy, mathbfx$ and I am adding the inequalities produced leading to
$(nabla f(mathbfx) - nabla f(mathbfy))^T (mathbfx - mathbfy) leq L |mathbfx - mathbfy |^2 $
but this doesn't seem to lead somewhere. Any ideas on how to prove this?
optimization convex-analysis lipschitz-functions
add a comment |Â
up vote
1
down vote
favorite
Let $f:mathbbR^drightarrow mathbbR$ be a convex and differentiable function. We are saying that $f$ is smooth (with parameter $L$) if $forall mathbfx, mathbfy in mathbbR^d: f(mathbfy) leq f(mathbfx) + nabla f(mathbfx)^T (mathbfy - mathbfx) + fracL2|mathbfx-mathbfy|^2$
I am trying to prove the following Lemma:
Let $f:mathbbR^d rightarrow mathbbR$ be convex and differentiable. The following two statements are equivalent:
(i) f is smooth with parameter $L$
(ii) $|nabla f(mathbfx) - nabla f(mathbfy)| leq L|mathbfx-mathbfy| forall mathbfx, mathbfy in mathbbR^d$
I've managed to do (ii) $rightarrow$ (i) by using the first order characterization of convexity, subtracting and adding the term $nabla f(mathbfx)^T(mathbfy-mathbfx)$ and then by using the Cauchy-Schwarz inequality.
However I am stuck in (i) $rightarrow$ (ii). I am taking the smoothness condition one time for $mathbfx, mathbfy$ and one time for $mathbfy, mathbfx$ and I am adding the inequalities produced leading to
$(nabla f(mathbfx) - nabla f(mathbfy))^T (mathbfx - mathbfy) leq L |mathbfx - mathbfy |^2 $
but this doesn't seem to lead somewhere. Any ideas on how to prove this?
optimization convex-analysis lipschitz-functions
This is explained in the first lecture entitled "Gradient method" in Vandenberghe's UCLA 236c notes.
– littleO
Jul 23 at 7:20
Thanks! Didn't have the notion of co-coercivity in our lecture notes!
– dimkou
Jul 23 at 7:23
add a comment |Â
up vote
1
down vote
favorite
up vote
1
down vote
favorite
Let $f:mathbbR^drightarrow mathbbR$ be a convex and differentiable function. We are saying that $f$ is smooth (with parameter $L$) if $forall mathbfx, mathbfy in mathbbR^d: f(mathbfy) leq f(mathbfx) + nabla f(mathbfx)^T (mathbfy - mathbfx) + fracL2|mathbfx-mathbfy|^2$
I am trying to prove the following Lemma:
Let $f:mathbbR^d rightarrow mathbbR$ be convex and differentiable. The following two statements are equivalent:
(i) f is smooth with parameter $L$
(ii) $|nabla f(mathbfx) - nabla f(mathbfy)| leq L|mathbfx-mathbfy| forall mathbfx, mathbfy in mathbbR^d$
I've managed to do (ii) $rightarrow$ (i) by using the first order characterization of convexity, subtracting and adding the term $nabla f(mathbfx)^T(mathbfy-mathbfx)$ and then by using the Cauchy-Schwarz inequality.
However I am stuck in (i) $rightarrow$ (ii). I am taking the smoothness condition one time for $mathbfx, mathbfy$ and one time for $mathbfy, mathbfx$ and I am adding the inequalities produced leading to
$(nabla f(mathbfx) - nabla f(mathbfy))^T (mathbfx - mathbfy) leq L |mathbfx - mathbfy |^2 $
but this doesn't seem to lead somewhere. Any ideas on how to prove this?
optimization convex-analysis lipschitz-functions
Let $f:mathbbR^drightarrow mathbbR$ be a convex and differentiable function. We are saying that $f$ is smooth (with parameter $L$) if $forall mathbfx, mathbfy in mathbbR^d: f(mathbfy) leq f(mathbfx) + nabla f(mathbfx)^T (mathbfy - mathbfx) + fracL2|mathbfx-mathbfy|^2$
I am trying to prove the following Lemma:
Let $f:mathbbR^d rightarrow mathbbR$ be convex and differentiable. The following two statements are equivalent:
(i) f is smooth with parameter $L$
(ii) $|nabla f(mathbfx) - nabla f(mathbfy)| leq L|mathbfx-mathbfy| forall mathbfx, mathbfy in mathbbR^d$
I've managed to do (ii) $rightarrow$ (i) by using the first order characterization of convexity, subtracting and adding the term $nabla f(mathbfx)^T(mathbfy-mathbfx)$ and then by using the Cauchy-Schwarz inequality.
However I am stuck in (i) $rightarrow$ (ii). I am taking the smoothness condition one time for $mathbfx, mathbfy$ and one time for $mathbfy, mathbfx$ and I am adding the inequalities produced leading to
$(nabla f(mathbfx) - nabla f(mathbfy))^T (mathbfx - mathbfy) leq L |mathbfx - mathbfy |^2 $
but this doesn't seem to lead somewhere. Any ideas on how to prove this?
optimization convex-analysis lipschitz-functions
asked Jul 23 at 7:01
dimkou
535
535
This is explained in the first lecture entitled "Gradient method" in Vandenberghe's UCLA 236c notes.
– littleO
Jul 23 at 7:20
Thanks! Didn't have the notion of co-coercivity in our lecture notes!
– dimkou
Jul 23 at 7:23
add a comment |Â
This is explained in the first lecture entitled "Gradient method" in Vandenberghe's UCLA 236c notes.
– littleO
Jul 23 at 7:20
Thanks! Didn't have the notion of co-coercivity in our lecture notes!
– dimkou
Jul 23 at 7:23
This is explained in the first lecture entitled "Gradient method" in Vandenberghe's UCLA 236c notes.
– littleO
Jul 23 at 7:20
This is explained in the first lecture entitled "Gradient method" in Vandenberghe's UCLA 236c notes.
– littleO
Jul 23 at 7:20
Thanks! Didn't have the notion of co-coercivity in our lecture notes!
– dimkou
Jul 23 at 7:23
Thanks! Didn't have the notion of co-coercivity in our lecture notes!
– dimkou
Jul 23 at 7:23
add a comment |Â
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
active
oldest
votes
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f2860081%2fsmoothness-means-that-the-gradient-is-lipschitz-continuous%23new-answer', 'question_page');
);
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
This is explained in the first lecture entitled "Gradient method" in Vandenberghe's UCLA 236c notes.
– littleO
Jul 23 at 7:20
Thanks! Didn't have the notion of co-coercivity in our lecture notes!
– dimkou
Jul 23 at 7:23