ATAR Notes: Forum
Uni Stuff => Science => Faculties => Mathematics => Topic started by: /0 on July 24, 2010, 12:29:44 pm
-
In the complex derivative,
 = \lim_{h \to 0}\frac{f(z+h)-f(z)}{h})
Can 'h' be anything?
This looks a bit like the directional derivative formula in
:
 = \lim_{h \to 0} \frac{f(\vec{x}+h\vec{u})-f(\vec{x})}{h})
But yet there is a 'unique'
? I would have thought
should vary depending on the direction you approach.
(This question of course leads to how the Cauchy-Riemann equations came about)
Thanks
-
Remember the definitions: a function is differentiable only if that limit exists.
If you get 2 different values depending on which direction you approach, then the limit does not exist and hence the function is not differentiable. if it -is- unique, then we say it is differentiable.
for example f(z)=Re(z)+2Im(z)= is not complex differentiable at 0 because the limit doesn't exist.
-
But if you think of the function as a 'contour map' on the x-y plane (
)... if every point in a domain has derivatives from every side equal, then surely the function must be constant?
-
It is true that if the directional derivatives of f:R^2->R at each point are the same for every direction, then f is constant. (because if directional derivative is the same in all directions it implies the gradient is 0 at that point. if the gradient at 0 at every point then f is constant)
However you can't think of a complex function C->C as a contour map, remember the range of the function the complex numbers, so the graph is a surface in C^2, or R^4.
just try a simple example, check that if f(z)=z^2 then
-
Ah, ok thanks zzdfa =)
Another Q, if you approach a point along the imaginary axis then:
 = \lim_{k \to 0} \frac{f(z+ik)-f(z)}{ik}=-i\frac{\partial f}{\partial y})
Is the
from the
in the limit?
Since -f(x,y)}{h})
It seems that if you set
, then it essentially becomes the first limit above:
Since
(
is the same as
right?)
So is it really ok to take the
out?
-
It seems that if you set
, then it essentially becomes the first limit above:
Since
(
is the same as
right?)
Yes the -i is from the 1/i, your argument doesn't work because
is actually equal to -f(x,y)}{ik}=i \frac{\partial f}{\partial x} )
because ik->0 is the same as k->0,
and
=f(x+i(y+ik))=f(x-k+iy)=f(x-k,y))
I think where you might be getting confused is the association between the function f:C->C and the function f*:R^2->C.
given a function with complex domain, f(z), the associated function f*:R^2->C is defined by f*(x,y)=f(x+yi).
of course in the above we used f to denote both f and f* because you can just look at the number of arguments to figure out which is which.
-
Thanks zzdfa xD
When people talk of a function being differentiable, do they mean differentiable in a neighbourhood, or differentiable everywhere?
Usually people only test differentiability at (0,0), is there any reason why?
The lecture notes give examples or functions which are non-differentiable:
,
,
,
and )
Does this mean differentiable nowhere, or does it mean not differentiable at a selection of points?
Also, the tute says that
is not analytic at
. But  = |x+iy|^2 = x^2+y^2 = u(x,y)+iv(x,y))
So
,
,
,
.
So at
, the Cauchy Riemann equations are satisfied, so why isn't the function analytic at z = 0?
Thanks
-
You might see your lecturer use (0,0) alot in examples, when computing derivatives, because it's simplifies the calculations a bit. But of course there is no special significance since you can translate any function by f(z-p).
Generally in complex analysis when we say differentiable we mean differentiable on some open set U.
You can say that your function f is differentiable at (0,0) with derivative 0, but you can't say it's analytic because analyticity at p means 'differentiable in an open set containing p'. so to show that f is analytic around p you have to show that the CR equations are satisfied in an open set around p, which is isn't in your case.
-
You might see your lecturer use (0,0) alot in examples, when computing derivatives, because it's simplifies the calculations a bit. But of course there is no special significance since you can translate any function by f(z-p).
Generally in complex analysis when we say differentiable we mean differentiable on some open set U.
You can say that your function f is differentiable at (0,0) with derivative 0, but you can't say it's analytic because analyticity at p means 'differentiable in an open set containing p'. so to show that f is analytic around p you have to show that the CR equations are satisfied in an open set around p, which is isn't in your case.
No, when you say differentiable, you just mean differentiable at a point. When it is holomorphic/analytic at a point, then it must be differentiable in a neighbourhood of that point.
-
ah yeah, I even contradicted myself in the next sentence :-\
-
Thanks zzdfa xD
When people talk of a function being differentiable, do they mean differentiable in a neighbourhood, or differentiable everywhere?
Usually people only test differentiability at (0,0), is there any reason why?
The lecture notes give examples or functions which are non-differentiable:
,
,
,
and )
Does this mean differentiable nowhere, or does it mean not differentiable at a selection of points?
Also, the tute says that
is not analytic at
. But  = |x+iy|^2 = x^2+y^2 = u(x,y)+iv(x,y))
So
,
,
,
.
So at
, the Cauchy Riemann equations are satisfied, so why isn't the function analytic at z = 0?
Thanks
Who's your tutor? As was mentioned, it's differentiable at zero, but analytic (well I prefer holomorphic) means it has to be differentiable in a neighbourhood of zero, which isn't the case here.
-
My tutor's Davidson Ng, but we haven't had a tutorial yet.
I understand what you mean, that it must be differentiable in a neighbourhood... but the theorem that I used is as follows:
Let
be given by
, for
. Assume that all partial derivatives
,
,
,
exist, are continuous, and satisfy the Cauchy-Riemann equations at
. Then
is holomorphic at
.
It never says anything about being differentiable in a neighbourhood, only about the continuity of the partial derivatives
-
Where did you see that?
satisfies CR equations at z = 0 and nowhere else, the partials are continuous but f is nowhere analytic.
-
I believe it's called the Looman-Menchoff Theorem
The version I quoted is what's in the lecture notes. Is there a difference?
-
Yeah. The result you should remember is that if CR equations are satisfied at a point and the partials are continuous at the point then the function is differentiable at the point. Then being holomorphic/analytic in some open set just means differentiable everywhere in the open set, so CR equations are satisfied + continuous partials everywhere in the open set. Analytic at a single point would mean differentiable in some open set containing the point.
-
Ooohhhh, I see now, thanks =D
-
Suppose you need to integrate a real function such as
so you introduce a complex function which has poles at the blue crosses in the picture (and some others in the lower half plane that aren't important). Aren't both of the possible contours just as valid as each other? The only difference is that one of them encloses 4 poles while the other only encloses 2, but once you shrink the small semi-cirlces to zero radius they should have zero contribution anyway..
-
Of course. But in either case the little semicircles may have a non-neglible contribution even as you shrink them smaller and smaller.
-
Ah thanks humph, I skimmed over that lemma before
-
Bit of a dumb question here,
If
, then why doesn't
have a branch cut on the negative real axis (assuming the principal logarithm)?
-
The branch cut from the negative logarithm is because when you approach it from one side you get a different limit than from the other side. But note that
, so it doesn't make any difference when you take an integer power.
-
o right, thanks
-
Just a few last minute questions xD
1. If you have a function like
.
Is it ok if for
and
you choose different intervals of existence for
and
?
i.e. can you rotate each of the branch cuts however you like?
And would this mean you have to deal with each logarithm as a separate entity? (i.e.
anymore?)
2. Also, for
, we take
. The solutions take the left contour, but I'm wonder, why couldn't we just take the right contour and just ignore the pole at
?
Thanks
-
From the PDEs exam:
"
a) Obtain the general solution of the first order PDE:

b) If we prescribe
on the upper portion of the hyperbola
,
, show that no solution exists unless
is of a special form. Find this form and show that in such a case there are infinitely many solutions."
I tried solving this the way it's done in the lecture notes but it didn't work =/

,
, 
I can't solve any of these straight away so the usual method is already looking shaky
But I notice that,

And,
, which means
, 
Hence,
, but
is a function of
so
.
Still, this isn't the solution... and I couldn't do the next part of the question
How can you solve it?
-
Hah, this was one of my assignment questions a couple of years ago. I'll just post it up in full.
Let us parametrise the characteristic curve in the plane of this partial differential equation by
, and let
. As the chain rule implies that

it follows that we can write
e^{x(s)} + y(s)u(s) , \qquad \frac{dx}{ds} = y(s) , \qquad \frac{dy}{ds} = x(s).<br />)
Writing
, we see that the latter two ordinary differential equations are equivalent to the vector-valued ordinary differential equation
.<br />)
The matrix above has eigenvalues
with corresponding eigenvectors
, and so the solution to this ordinary differential equation is given by
 = c_1 \begin{pmatrix} 1 \\ 1 \end{pmatrix} e^{s} + c_2 \begin{pmatrix} 1 \\ -1 \end{pmatrix} e^{-s} ,<br />)
where
. Thus we have that
and
. Substituting these values into the earlier expression for
, we have that
e^{c_1 e^{s} + c_2 e^{-s}} + (c_1 e^{s} - c_2 e^{-s})u(s),<br />)
which, by multiplying through by the integrating factor
, can be rearranged to read
 = c_1 e^{s} + c_2 e^{-s},<br />)
and so we can integrate to find that
 = c_1 e^{s} - c_2 e^{-s} + u_0,<br />)
or equivalently, by substituting back in
,
, and then multiplying through by
,
 = e^{x}\left(y + u_0\right),<br />)
where
depends on the chosen characteristic curve. These characteristic curves are the solutions to

which, by separating variables, we can solve to find that

where
. Thus we have that
for any arbitrary function
, and so
 = e^{x}\left(y + f\left(y^2 - x^2\right)\right).<br />)
With this solution of the partial differential equation, we therefore have that on the upper portion of the hyperbola
,
,
 = e^{x}\left(y + f(1)\right).<br />)
Thus if
is not of the form
for some
, then we cannot have that
on the upper portion of the hyperbola
,
. If, on the other hand,
is of this form, then there are infinitely many solutions of the partial differential equation satisfying this prescribed condition, as there are infinitely many functions
satisfying
; for example, we could have
for any
.
-
lol wow that's pretty full-on. I don't know why it didn't occur to me to use the theory of DE systems from 2405... i mean, that stuff's made for these PDEs. But even if it did the double exponentials would probably have put me off.
But yeah that definitely makes sense, thanks a bunch humph :)