Why a continuous but not uniformly continuous function $f$ in $mathbb{R}^n$ becomes uniformly continuous when...
$begingroup$
I've seen the proof of the title and roughly follow every steps of the proof. But I cannot see the intuition of this statement. Could someone show me a concrete example using $delta-epsilon$? Thanks!
real-analysis
$endgroup$
add a comment |
$begingroup$
I've seen the proof of the title and roughly follow every steps of the proof. But I cannot see the intuition of this statement. Could someone show me a concrete example using $delta-epsilon$? Thanks!
real-analysis
$endgroup$
add a comment |
$begingroup$
I've seen the proof of the title and roughly follow every steps of the proof. But I cannot see the intuition of this statement. Could someone show me a concrete example using $delta-epsilon$? Thanks!
real-analysis
$endgroup$
I've seen the proof of the title and roughly follow every steps of the proof. But I cannot see the intuition of this statement. Could someone show me a concrete example using $delta-epsilon$? Thanks!
real-analysis
real-analysis
asked Dec 20 '18 at 4:43
Cathy Cathy
30529
30529
add a comment |
add a comment |
4 Answers
4
active
oldest
votes
$begingroup$
Let me give you some examples that may help to form the intuition about compact sets. Although these examples will be functions on $Bbb R$, the idea for $Bbb R^n$ is not that different.
By the Heine-Borel theorem, compact sets in $Bbb R^n$ are precisely those that are closed and bounded. Let's see what happens when these $2$ conditions are not satisfied.
Closed: Strong oscillation is one reason that prevents a continuous function from being uniformly continuous. Consider the continuous function $f:(0,1]toBbb R$ defined by
$$
f(x)=sinleft(frac 1x right).
$$
This function is not uniformly continuous because the oscillation becomes infinitely strong near $x=0$. This happens because $x=0$ is on the boundary of $(0,1]$ but $f(0)$ need not be defined, hence the behaviour of $f$ can be very wild near the boundary points. On the other hand, if our domain is compact, says $D=[0,1]$ then $f(0)$ must be defined so the oscillation near $x=0$, if there is any, must "dies out" fast enough so that $lim_{yto 0}f(y)=f(0)$.
Bounded: Steepness is another reason that a function can fail to be uniformly continuous. Consider $f:[0,infty)to Bbb R$ befined by
$$
f(x)=x^2.
$$
Suppose that $a<b$. By the mean value theorem, we have $|f(a)-f(b)|=|f'(xi)||a-b|$ for some $xiin(a,b)$. We know that $f'(x)=2x$ and that our domain is not bounded, thus $f'(xi)$ is unbounded too. This implies that $f$ is not uniformly continuous. If, however, our domain is bounded, says $D=[0,M]$, then $f'(x)$ is bounded and our function must be uniformly continuous.
In general, $f$ need not be differentiable so the above reasoning doesn't apply literally. Still, I hope that gives some intuition that might help. The situation is also related to the extreme value theorem, i.e. that a continuous function cannot be unbounded on a compact set.
$endgroup$
add a comment |
$begingroup$
A concrete example can be as follows. Take e.g. $f(x) = x^2$. That is a continuous function on $mathbb{R}$, but it is not uniformly continuous because $|f(y) - f(x)| = |y^2 - x^2|$ cannot be kept smaller that a given $varepsilon>0$ for every $x,y$ whose difference $|y-x|$ is less that a given $delta_{varepsilon}$ depending only on $varepsilon$ but not on $x,y$. In other words, there is no $delta_{varepsilon}>0$ such that $|y-x| < delta_{varepsilon}$ implies $|f(y) - f(x)| = |y^2 - x^2| < delta_{varepsilon}$, because $|y^2 - x^2| = |x+y||x-y|$, and it does not matter how small $|x-y|$ is, the other factor $|x+y|$ can be made large enough (by picking $x,y$ large) so that $|y^2 - x^2| geq delta_{varepsilon}$. However, a compact set of $mathbb{R}$ such as the interval $[0,1]$ is bounded, and that makes the factor $|x+y|$ also bounded, more specifically $|x+y| leq 2$ on $[0,1]$. So now we can find a $delta_{epsilon} = frac{varepsilon}{2}$ such that $|y-x| < delta_{varepsilon}$ implies
$$|y^2 - x^2| = |x+y||x-y| < 2 delta_{varepsilon} = 2 cdot frac{varepsilon}{2} = varepsilon$$
Edit: By the time I finished writing my answer I found that BigbearZzz had already posted a more complete answer. I'll post mine anyway as a simpler version, and using $varepsilon-delta$ as requested by the OP.
$endgroup$
add a comment |
$begingroup$
If for say $epsilon=1$ there's no $delta$ that works in the entire compact set, then let $x_n$ be such that $delta=1/n$ doesn't work. Then an accumulation point of the $x_n$ is a point where $f$ is not continuous at all.
$endgroup$
add a comment |
$begingroup$
Consider $f(x)=frac{1}{x}$ which is continuous on $(0,infty)$ but not uniformly so.
Why can't we find one $deltagt 0$ that works for all $xin (0,infty)$? As $x$ gets closer to $0$, $ f$ gets steeper so that small changes in $x$ lead to larger changes in $f$.
Suppose we are given $epsilongt 0$ and $ain(0,infty)$. We want to find $deltagt 0$ such that $|x-a|ltdeltaimplies|frac{1}{a}-frac{1}{x}|ltepsilon$.
$$|frac{1}{a}-frac{1}{x}|ltepsilonimplies -epsilonltfrac{1}{a}-frac{1}{x}ltepsilonimplies -frac{1}{a}-epsilonlt -frac{1}{x}ltepsilon-frac{1}{a}impliesfrac{1}{a}+epsilongtfrac{1}{x}gtfrac{1}{a}-epsilon$$
The right hand side is equivalent to $frac{1+aepsilon}{a}gtfrac{1}{x}gtfrac{1-aepsilon}{a}$.
If $epsilon$ is large enough that $1-aepsilonlt 0$, we can restrict $epsilon$ to be smaller so that $1-aepsilongt 0$. We can do this because if the absolute value is less than the smaller $epsilon$ then it must also be less than the larger $epsilon$.
Restricting $epsilon$ and taking reciprocals gives us $frac{a}{1+aepsilon}lt xltfrac{a}{1-aepsilon}$.
For small $epsilon$, $ frac{a}{1+aepsilon}$ is a number slightly smaller than $a$ and $frac{a}{1-aepsilon}$ is a number slightly larger than $a$. Therefore, as $a$ gets closer to $0$, we require a smaller $delta$ to make $|frac{1}{a}-frac{1}{x}|ltepsilon$.
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3047165%2fwhy-a-continuous-but-not-uniformly-continuous-function-f-in-mathbbrn-bec%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Let me give you some examples that may help to form the intuition about compact sets. Although these examples will be functions on $Bbb R$, the idea for $Bbb R^n$ is not that different.
By the Heine-Borel theorem, compact sets in $Bbb R^n$ are precisely those that are closed and bounded. Let's see what happens when these $2$ conditions are not satisfied.
Closed: Strong oscillation is one reason that prevents a continuous function from being uniformly continuous. Consider the continuous function $f:(0,1]toBbb R$ defined by
$$
f(x)=sinleft(frac 1x right).
$$
This function is not uniformly continuous because the oscillation becomes infinitely strong near $x=0$. This happens because $x=0$ is on the boundary of $(0,1]$ but $f(0)$ need not be defined, hence the behaviour of $f$ can be very wild near the boundary points. On the other hand, if our domain is compact, says $D=[0,1]$ then $f(0)$ must be defined so the oscillation near $x=0$, if there is any, must "dies out" fast enough so that $lim_{yto 0}f(y)=f(0)$.
Bounded: Steepness is another reason that a function can fail to be uniformly continuous. Consider $f:[0,infty)to Bbb R$ befined by
$$
f(x)=x^2.
$$
Suppose that $a<b$. By the mean value theorem, we have $|f(a)-f(b)|=|f'(xi)||a-b|$ for some $xiin(a,b)$. We know that $f'(x)=2x$ and that our domain is not bounded, thus $f'(xi)$ is unbounded too. This implies that $f$ is not uniformly continuous. If, however, our domain is bounded, says $D=[0,M]$, then $f'(x)$ is bounded and our function must be uniformly continuous.
In general, $f$ need not be differentiable so the above reasoning doesn't apply literally. Still, I hope that gives some intuition that might help. The situation is also related to the extreme value theorem, i.e. that a continuous function cannot be unbounded on a compact set.
$endgroup$
add a comment |
$begingroup$
Let me give you some examples that may help to form the intuition about compact sets. Although these examples will be functions on $Bbb R$, the idea for $Bbb R^n$ is not that different.
By the Heine-Borel theorem, compact sets in $Bbb R^n$ are precisely those that are closed and bounded. Let's see what happens when these $2$ conditions are not satisfied.
Closed: Strong oscillation is one reason that prevents a continuous function from being uniformly continuous. Consider the continuous function $f:(0,1]toBbb R$ defined by
$$
f(x)=sinleft(frac 1x right).
$$
This function is not uniformly continuous because the oscillation becomes infinitely strong near $x=0$. This happens because $x=0$ is on the boundary of $(0,1]$ but $f(0)$ need not be defined, hence the behaviour of $f$ can be very wild near the boundary points. On the other hand, if our domain is compact, says $D=[0,1]$ then $f(0)$ must be defined so the oscillation near $x=0$, if there is any, must "dies out" fast enough so that $lim_{yto 0}f(y)=f(0)$.
Bounded: Steepness is another reason that a function can fail to be uniformly continuous. Consider $f:[0,infty)to Bbb R$ befined by
$$
f(x)=x^2.
$$
Suppose that $a<b$. By the mean value theorem, we have $|f(a)-f(b)|=|f'(xi)||a-b|$ for some $xiin(a,b)$. We know that $f'(x)=2x$ and that our domain is not bounded, thus $f'(xi)$ is unbounded too. This implies that $f$ is not uniformly continuous. If, however, our domain is bounded, says $D=[0,M]$, then $f'(x)$ is bounded and our function must be uniformly continuous.
In general, $f$ need not be differentiable so the above reasoning doesn't apply literally. Still, I hope that gives some intuition that might help. The situation is also related to the extreme value theorem, i.e. that a continuous function cannot be unbounded on a compact set.
$endgroup$
add a comment |
$begingroup$
Let me give you some examples that may help to form the intuition about compact sets. Although these examples will be functions on $Bbb R$, the idea for $Bbb R^n$ is not that different.
By the Heine-Borel theorem, compact sets in $Bbb R^n$ are precisely those that are closed and bounded. Let's see what happens when these $2$ conditions are not satisfied.
Closed: Strong oscillation is one reason that prevents a continuous function from being uniformly continuous. Consider the continuous function $f:(0,1]toBbb R$ defined by
$$
f(x)=sinleft(frac 1x right).
$$
This function is not uniformly continuous because the oscillation becomes infinitely strong near $x=0$. This happens because $x=0$ is on the boundary of $(0,1]$ but $f(0)$ need not be defined, hence the behaviour of $f$ can be very wild near the boundary points. On the other hand, if our domain is compact, says $D=[0,1]$ then $f(0)$ must be defined so the oscillation near $x=0$, if there is any, must "dies out" fast enough so that $lim_{yto 0}f(y)=f(0)$.
Bounded: Steepness is another reason that a function can fail to be uniformly continuous. Consider $f:[0,infty)to Bbb R$ befined by
$$
f(x)=x^2.
$$
Suppose that $a<b$. By the mean value theorem, we have $|f(a)-f(b)|=|f'(xi)||a-b|$ for some $xiin(a,b)$. We know that $f'(x)=2x$ and that our domain is not bounded, thus $f'(xi)$ is unbounded too. This implies that $f$ is not uniformly continuous. If, however, our domain is bounded, says $D=[0,M]$, then $f'(x)$ is bounded and our function must be uniformly continuous.
In general, $f$ need not be differentiable so the above reasoning doesn't apply literally. Still, I hope that gives some intuition that might help. The situation is also related to the extreme value theorem, i.e. that a continuous function cannot be unbounded on a compact set.
$endgroup$
Let me give you some examples that may help to form the intuition about compact sets. Although these examples will be functions on $Bbb R$, the idea for $Bbb R^n$ is not that different.
By the Heine-Borel theorem, compact sets in $Bbb R^n$ are precisely those that are closed and bounded. Let's see what happens when these $2$ conditions are not satisfied.
Closed: Strong oscillation is one reason that prevents a continuous function from being uniformly continuous. Consider the continuous function $f:(0,1]toBbb R$ defined by
$$
f(x)=sinleft(frac 1x right).
$$
This function is not uniformly continuous because the oscillation becomes infinitely strong near $x=0$. This happens because $x=0$ is on the boundary of $(0,1]$ but $f(0)$ need not be defined, hence the behaviour of $f$ can be very wild near the boundary points. On the other hand, if our domain is compact, says $D=[0,1]$ then $f(0)$ must be defined so the oscillation near $x=0$, if there is any, must "dies out" fast enough so that $lim_{yto 0}f(y)=f(0)$.
Bounded: Steepness is another reason that a function can fail to be uniformly continuous. Consider $f:[0,infty)to Bbb R$ befined by
$$
f(x)=x^2.
$$
Suppose that $a<b$. By the mean value theorem, we have $|f(a)-f(b)|=|f'(xi)||a-b|$ for some $xiin(a,b)$. We know that $f'(x)=2x$ and that our domain is not bounded, thus $f'(xi)$ is unbounded too. This implies that $f$ is not uniformly continuous. If, however, our domain is bounded, says $D=[0,M]$, then $f'(x)$ is bounded and our function must be uniformly continuous.
In general, $f$ need not be differentiable so the above reasoning doesn't apply literally. Still, I hope that gives some intuition that might help. The situation is also related to the extreme value theorem, i.e. that a continuous function cannot be unbounded on a compact set.
answered Dec 20 '18 at 5:15
BigbearZzzBigbearZzz
9,01021652
9,01021652
add a comment |
add a comment |
$begingroup$
A concrete example can be as follows. Take e.g. $f(x) = x^2$. That is a continuous function on $mathbb{R}$, but it is not uniformly continuous because $|f(y) - f(x)| = |y^2 - x^2|$ cannot be kept smaller that a given $varepsilon>0$ for every $x,y$ whose difference $|y-x|$ is less that a given $delta_{varepsilon}$ depending only on $varepsilon$ but not on $x,y$. In other words, there is no $delta_{varepsilon}>0$ such that $|y-x| < delta_{varepsilon}$ implies $|f(y) - f(x)| = |y^2 - x^2| < delta_{varepsilon}$, because $|y^2 - x^2| = |x+y||x-y|$, and it does not matter how small $|x-y|$ is, the other factor $|x+y|$ can be made large enough (by picking $x,y$ large) so that $|y^2 - x^2| geq delta_{varepsilon}$. However, a compact set of $mathbb{R}$ such as the interval $[0,1]$ is bounded, and that makes the factor $|x+y|$ also bounded, more specifically $|x+y| leq 2$ on $[0,1]$. So now we can find a $delta_{epsilon} = frac{varepsilon}{2}$ such that $|y-x| < delta_{varepsilon}$ implies
$$|y^2 - x^2| = |x+y||x-y| < 2 delta_{varepsilon} = 2 cdot frac{varepsilon}{2} = varepsilon$$
Edit: By the time I finished writing my answer I found that BigbearZzz had already posted a more complete answer. I'll post mine anyway as a simpler version, and using $varepsilon-delta$ as requested by the OP.
$endgroup$
add a comment |
$begingroup$
A concrete example can be as follows. Take e.g. $f(x) = x^2$. That is a continuous function on $mathbb{R}$, but it is not uniformly continuous because $|f(y) - f(x)| = |y^2 - x^2|$ cannot be kept smaller that a given $varepsilon>0$ for every $x,y$ whose difference $|y-x|$ is less that a given $delta_{varepsilon}$ depending only on $varepsilon$ but not on $x,y$. In other words, there is no $delta_{varepsilon}>0$ such that $|y-x| < delta_{varepsilon}$ implies $|f(y) - f(x)| = |y^2 - x^2| < delta_{varepsilon}$, because $|y^2 - x^2| = |x+y||x-y|$, and it does not matter how small $|x-y|$ is, the other factor $|x+y|$ can be made large enough (by picking $x,y$ large) so that $|y^2 - x^2| geq delta_{varepsilon}$. However, a compact set of $mathbb{R}$ such as the interval $[0,1]$ is bounded, and that makes the factor $|x+y|$ also bounded, more specifically $|x+y| leq 2$ on $[0,1]$. So now we can find a $delta_{epsilon} = frac{varepsilon}{2}$ such that $|y-x| < delta_{varepsilon}$ implies
$$|y^2 - x^2| = |x+y||x-y| < 2 delta_{varepsilon} = 2 cdot frac{varepsilon}{2} = varepsilon$$
Edit: By the time I finished writing my answer I found that BigbearZzz had already posted a more complete answer. I'll post mine anyway as a simpler version, and using $varepsilon-delta$ as requested by the OP.
$endgroup$
add a comment |
$begingroup$
A concrete example can be as follows. Take e.g. $f(x) = x^2$. That is a continuous function on $mathbb{R}$, but it is not uniformly continuous because $|f(y) - f(x)| = |y^2 - x^2|$ cannot be kept smaller that a given $varepsilon>0$ for every $x,y$ whose difference $|y-x|$ is less that a given $delta_{varepsilon}$ depending only on $varepsilon$ but not on $x,y$. In other words, there is no $delta_{varepsilon}>0$ such that $|y-x| < delta_{varepsilon}$ implies $|f(y) - f(x)| = |y^2 - x^2| < delta_{varepsilon}$, because $|y^2 - x^2| = |x+y||x-y|$, and it does not matter how small $|x-y|$ is, the other factor $|x+y|$ can be made large enough (by picking $x,y$ large) so that $|y^2 - x^2| geq delta_{varepsilon}$. However, a compact set of $mathbb{R}$ such as the interval $[0,1]$ is bounded, and that makes the factor $|x+y|$ also bounded, more specifically $|x+y| leq 2$ on $[0,1]$. So now we can find a $delta_{epsilon} = frac{varepsilon}{2}$ such that $|y-x| < delta_{varepsilon}$ implies
$$|y^2 - x^2| = |x+y||x-y| < 2 delta_{varepsilon} = 2 cdot frac{varepsilon}{2} = varepsilon$$
Edit: By the time I finished writing my answer I found that BigbearZzz had already posted a more complete answer. I'll post mine anyway as a simpler version, and using $varepsilon-delta$ as requested by the OP.
$endgroup$
A concrete example can be as follows. Take e.g. $f(x) = x^2$. That is a continuous function on $mathbb{R}$, but it is not uniformly continuous because $|f(y) - f(x)| = |y^2 - x^2|$ cannot be kept smaller that a given $varepsilon>0$ for every $x,y$ whose difference $|y-x|$ is less that a given $delta_{varepsilon}$ depending only on $varepsilon$ but not on $x,y$. In other words, there is no $delta_{varepsilon}>0$ such that $|y-x| < delta_{varepsilon}$ implies $|f(y) - f(x)| = |y^2 - x^2| < delta_{varepsilon}$, because $|y^2 - x^2| = |x+y||x-y|$, and it does not matter how small $|x-y|$ is, the other factor $|x+y|$ can be made large enough (by picking $x,y$ large) so that $|y^2 - x^2| geq delta_{varepsilon}$. However, a compact set of $mathbb{R}$ such as the interval $[0,1]$ is bounded, and that makes the factor $|x+y|$ also bounded, more specifically $|x+y| leq 2$ on $[0,1]$. So now we can find a $delta_{epsilon} = frac{varepsilon}{2}$ such that $|y-x| < delta_{varepsilon}$ implies
$$|y^2 - x^2| = |x+y||x-y| < 2 delta_{varepsilon} = 2 cdot frac{varepsilon}{2} = varepsilon$$
Edit: By the time I finished writing my answer I found that BigbearZzz had already posted a more complete answer. I'll post mine anyway as a simpler version, and using $varepsilon-delta$ as requested by the OP.
answered Dec 20 '18 at 5:34
mlerma54mlerma54
1,167148
1,167148
add a comment |
add a comment |
$begingroup$
If for say $epsilon=1$ there's no $delta$ that works in the entire compact set, then let $x_n$ be such that $delta=1/n$ doesn't work. Then an accumulation point of the $x_n$ is a point where $f$ is not continuous at all.
$endgroup$
add a comment |
$begingroup$
If for say $epsilon=1$ there's no $delta$ that works in the entire compact set, then let $x_n$ be such that $delta=1/n$ doesn't work. Then an accumulation point of the $x_n$ is a point where $f$ is not continuous at all.
$endgroup$
add a comment |
$begingroup$
If for say $epsilon=1$ there's no $delta$ that works in the entire compact set, then let $x_n$ be such that $delta=1/n$ doesn't work. Then an accumulation point of the $x_n$ is a point where $f$ is not continuous at all.
$endgroup$
If for say $epsilon=1$ there's no $delta$ that works in the entire compact set, then let $x_n$ be such that $delta=1/n$ doesn't work. Then an accumulation point of the $x_n$ is a point where $f$ is not continuous at all.
answered Dec 20 '18 at 5:04
Bjørn Kjos-HanssenBjørn Kjos-Hanssen
2,107918
2,107918
add a comment |
add a comment |
$begingroup$
Consider $f(x)=frac{1}{x}$ which is continuous on $(0,infty)$ but not uniformly so.
Why can't we find one $deltagt 0$ that works for all $xin (0,infty)$? As $x$ gets closer to $0$, $ f$ gets steeper so that small changes in $x$ lead to larger changes in $f$.
Suppose we are given $epsilongt 0$ and $ain(0,infty)$. We want to find $deltagt 0$ such that $|x-a|ltdeltaimplies|frac{1}{a}-frac{1}{x}|ltepsilon$.
$$|frac{1}{a}-frac{1}{x}|ltepsilonimplies -epsilonltfrac{1}{a}-frac{1}{x}ltepsilonimplies -frac{1}{a}-epsilonlt -frac{1}{x}ltepsilon-frac{1}{a}impliesfrac{1}{a}+epsilongtfrac{1}{x}gtfrac{1}{a}-epsilon$$
The right hand side is equivalent to $frac{1+aepsilon}{a}gtfrac{1}{x}gtfrac{1-aepsilon}{a}$.
If $epsilon$ is large enough that $1-aepsilonlt 0$, we can restrict $epsilon$ to be smaller so that $1-aepsilongt 0$. We can do this because if the absolute value is less than the smaller $epsilon$ then it must also be less than the larger $epsilon$.
Restricting $epsilon$ and taking reciprocals gives us $frac{a}{1+aepsilon}lt xltfrac{a}{1-aepsilon}$.
For small $epsilon$, $ frac{a}{1+aepsilon}$ is a number slightly smaller than $a$ and $frac{a}{1-aepsilon}$ is a number slightly larger than $a$. Therefore, as $a$ gets closer to $0$, we require a smaller $delta$ to make $|frac{1}{a}-frac{1}{x}|ltepsilon$.
$endgroup$
add a comment |
$begingroup$
Consider $f(x)=frac{1}{x}$ which is continuous on $(0,infty)$ but not uniformly so.
Why can't we find one $deltagt 0$ that works for all $xin (0,infty)$? As $x$ gets closer to $0$, $ f$ gets steeper so that small changes in $x$ lead to larger changes in $f$.
Suppose we are given $epsilongt 0$ and $ain(0,infty)$. We want to find $deltagt 0$ such that $|x-a|ltdeltaimplies|frac{1}{a}-frac{1}{x}|ltepsilon$.
$$|frac{1}{a}-frac{1}{x}|ltepsilonimplies -epsilonltfrac{1}{a}-frac{1}{x}ltepsilonimplies -frac{1}{a}-epsilonlt -frac{1}{x}ltepsilon-frac{1}{a}impliesfrac{1}{a}+epsilongtfrac{1}{x}gtfrac{1}{a}-epsilon$$
The right hand side is equivalent to $frac{1+aepsilon}{a}gtfrac{1}{x}gtfrac{1-aepsilon}{a}$.
If $epsilon$ is large enough that $1-aepsilonlt 0$, we can restrict $epsilon$ to be smaller so that $1-aepsilongt 0$. We can do this because if the absolute value is less than the smaller $epsilon$ then it must also be less than the larger $epsilon$.
Restricting $epsilon$ and taking reciprocals gives us $frac{a}{1+aepsilon}lt xltfrac{a}{1-aepsilon}$.
For small $epsilon$, $ frac{a}{1+aepsilon}$ is a number slightly smaller than $a$ and $frac{a}{1-aepsilon}$ is a number slightly larger than $a$. Therefore, as $a$ gets closer to $0$, we require a smaller $delta$ to make $|frac{1}{a}-frac{1}{x}|ltepsilon$.
$endgroup$
add a comment |
$begingroup$
Consider $f(x)=frac{1}{x}$ which is continuous on $(0,infty)$ but not uniformly so.
Why can't we find one $deltagt 0$ that works for all $xin (0,infty)$? As $x$ gets closer to $0$, $ f$ gets steeper so that small changes in $x$ lead to larger changes in $f$.
Suppose we are given $epsilongt 0$ and $ain(0,infty)$. We want to find $deltagt 0$ such that $|x-a|ltdeltaimplies|frac{1}{a}-frac{1}{x}|ltepsilon$.
$$|frac{1}{a}-frac{1}{x}|ltepsilonimplies -epsilonltfrac{1}{a}-frac{1}{x}ltepsilonimplies -frac{1}{a}-epsilonlt -frac{1}{x}ltepsilon-frac{1}{a}impliesfrac{1}{a}+epsilongtfrac{1}{x}gtfrac{1}{a}-epsilon$$
The right hand side is equivalent to $frac{1+aepsilon}{a}gtfrac{1}{x}gtfrac{1-aepsilon}{a}$.
If $epsilon$ is large enough that $1-aepsilonlt 0$, we can restrict $epsilon$ to be smaller so that $1-aepsilongt 0$. We can do this because if the absolute value is less than the smaller $epsilon$ then it must also be less than the larger $epsilon$.
Restricting $epsilon$ and taking reciprocals gives us $frac{a}{1+aepsilon}lt xltfrac{a}{1-aepsilon}$.
For small $epsilon$, $ frac{a}{1+aepsilon}$ is a number slightly smaller than $a$ and $frac{a}{1-aepsilon}$ is a number slightly larger than $a$. Therefore, as $a$ gets closer to $0$, we require a smaller $delta$ to make $|frac{1}{a}-frac{1}{x}|ltepsilon$.
$endgroup$
Consider $f(x)=frac{1}{x}$ which is continuous on $(0,infty)$ but not uniformly so.
Why can't we find one $deltagt 0$ that works for all $xin (0,infty)$? As $x$ gets closer to $0$, $ f$ gets steeper so that small changes in $x$ lead to larger changes in $f$.
Suppose we are given $epsilongt 0$ and $ain(0,infty)$. We want to find $deltagt 0$ such that $|x-a|ltdeltaimplies|frac{1}{a}-frac{1}{x}|ltepsilon$.
$$|frac{1}{a}-frac{1}{x}|ltepsilonimplies -epsilonltfrac{1}{a}-frac{1}{x}ltepsilonimplies -frac{1}{a}-epsilonlt -frac{1}{x}ltepsilon-frac{1}{a}impliesfrac{1}{a}+epsilongtfrac{1}{x}gtfrac{1}{a}-epsilon$$
The right hand side is equivalent to $frac{1+aepsilon}{a}gtfrac{1}{x}gtfrac{1-aepsilon}{a}$.
If $epsilon$ is large enough that $1-aepsilonlt 0$, we can restrict $epsilon$ to be smaller so that $1-aepsilongt 0$. We can do this because if the absolute value is less than the smaller $epsilon$ then it must also be less than the larger $epsilon$.
Restricting $epsilon$ and taking reciprocals gives us $frac{a}{1+aepsilon}lt xltfrac{a}{1-aepsilon}$.
For small $epsilon$, $ frac{a}{1+aepsilon}$ is a number slightly smaller than $a$ and $frac{a}{1-aepsilon}$ is a number slightly larger than $a$. Therefore, as $a$ gets closer to $0$, we require a smaller $delta$ to make $|frac{1}{a}-frac{1}{x}|ltepsilon$.
answered Dec 20 '18 at 5:38
John DoumaJohn Douma
5,83521520
5,83521520
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f3047165%2fwhy-a-continuous-but-not-uniformly-continuous-function-f-in-mathbbrn-bec%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown